Conference paper

RoSum-Mcts: Monte Carlo Tree Search-Inspired HDL Code Summarization with Structural Rewards

Abstract

Large language models (LLMs) have shown promise in code summarization, yet their effectiveness for Hardware Description Languages (HDLs) like VHDL and Verilog remains underexplored. We propose RoSum-Mcts, an LLM-guided approach inspired by Monte Carlo Tree Search (MCTS) that refines summaries through structured exploration and reinforcement-driven optimization. Our method integrates both local and global context via a hierarchical candidate expansion mechanism and optimizes summaries using a composite reward function balancing functional correctness (FC), local content adequacy (LCA), and fluency. We evaluate RoSum-Mcts on the VHDL-eval and Verilog-eval datasets, demonstrating its consistent outperformance over baseline methods by leveraging structured bottom-up refinement and reinforcement-based optimization. Ablation studies confirm the necessity of both local and global expansion strategies, as well as the importance of balancing FC and LCA for optimal performance. Furthermore, RoSum-Mcts proves robust against superficial modifications, such as variable renaming, maintaining summary quality where baselines degrade. These results establish RoSum-Mcts as an effective and robust HDL summarization framework, paving the way for further research into reinforcement-enhanced code summarization.