Core Concepts
This paper introduces OR-Instruct, a novel framework for training open-source large language models (ORLMs) to automate optimization modeling, addressing the limitations of existing methods reliant on closed-source LLMs and limited datasets.
Abstract
ORLM: Training Open-Source Large Language Models for Automated Optimization Modeling Using a Customizable Synthetic Data Framework
This research paper introduces a novel approach to training open-source large language models (LLMs) for the complex task of automated optimization modeling. The authors highlight the limitations of existing methods, particularly their reliance on closed-source LLMs and the scarcity of high-quality training data. To address these challenges, they propose OR-Instruct, a semi-automated framework for generating synthetic data tailored to the specific requirements of optimization modeling.
This study aims to develop and evaluate a new method for training open-source LLMs capable of automatically generating optimization models and solver code from natural language problem descriptions.
The researchers designed OR-Instruct, a semi-automated data synthesis framework that employs two key strategies: expansion and augmentation. Expansion leverages GPT-4 to generate diverse scenarios and question types based on a seed dataset of real-world industry cases. Augmentation focuses on enhancing problem-solution diversity by rephrasing questions, modifying objectives and constraints, and incorporating various modeling techniques. The generated data is then used to train open-source LLMs, resulting in specialized models called ORLMs. The effectiveness of ORLMs is evaluated on three benchmarks: NL4OPT, MAMO, and a newly introduced industrial benchmark called IndustryOR.