Fair and Efficient Chore Allocation: Existence and Computation

Abstract

Large language models exhibit remarkable performance in simple code generation tasks. However, they encounter significant challenges when addressing complex problems that require reasoning and question decomposition. To tackle this, we propose a self-driven reasoning augmentation process, SRA-MCTS, which incorporates Monte Carlo Tree Search (MCTS) for reasoning data generation. SRA-MCTS enables LLMs to self-generate intermediate reasoning steps and perform iterative self-evaluation, facilitating self-improvement. Specifically, it utilizes MCTS to produce diverse intermediate reasoning steps. During each iteration, MCTS generates a step and employs self-evaluation to guide the selection of subsequent branches, ultimately forming a sufficiently diverse reasoning path referred to as “thinking”. This thinking guides the model in generating corresponding code, and both are combined as training data for supervised fine-tuning. Experimental results demonstrate that SRA-MCTS achieves consistent performance improvements across three model scales without additional supervisory assistance. Applied to the Meta-Llama-3.1-8B-Instruct model, it delivers an 11-point improvement on the MBPP-Complex dataset, underscoring the significant potential for model self-improvement. The code and data are available at https://github.com/DIRECT-BIT/SRA-MCTS.

Cite

Text

Murhekar. "Fair and Efficient Chore Allocation: Existence and Computation." International Joint Conference on Artificial Intelligence, 2024. doi:10.24963/ijcai.2024/965

Markdown

[Murhekar. "Fair and Efficient Chore Allocation: Existence and Computation." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/murhekar2024ijcai-fair/) doi:10.24963/ijcai.2024/965

BibTeX

@inproceedings{murhekar2024ijcai-fair,
  title     = {{Fair and Efficient Chore Allocation: Existence and Computation}},
  author    = {Murhekar, Aniket},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {8500-8501},
  doi       = {10.24963/ijcai.2024/965},
  url       = {https://mlanthology.org/ijcai/2024/murhekar2024ijcai-fair/}
}