Keywords: Job Shop Scheduling Problem, JSSP, Large Language Models, LLM, Starjob, Combinatorial Optimization, Llama 3.1, RsLORA, Supervised Learning, Natural Language Processing
TL;DR: The paper introduces Starjob, a large-scale supervised dataset for the Job Shop Scheduling Problem (JSSP), and fine-tunes a Llama-3.1 8B model to create a scheduler that outperforms traditional and some dedicated neural approaches.
Abstract: Large Language Models (LLMs) have shown remarkable capabilities across various domains, but their potential for solving combinatorial optimization problems remains largely unexplored. In this paper, we investigate the applicability of LLMs to the Job Shop Scheduling Problem (JSSP), a classic challenge in combinatorial optimization that requires efficient job allocation to machines to minimize makespan. To this end, we introduce Starjob, the first supervised dataset for JSSP, comprising 120k instances specifically designed for training LLMs. Leveraging this dataset, we fine-tune the LLaMA 8B model with the LoRA method to develop an end-to-end scheduling approach. Our evaluation on standard benchmarks demonstrates that the proposed LLM-based method not only surpasses traditional Priority Dispatching Rules (PDRs) but also achieves notable improvements over state-of-the-art neural approaches like L2D, with an average improvement of 11.28% on DMU and 3.29% on Taillard benchmarks. These results highlight the untapped potential of LLMs in tackling combinatorial optimization problems, paving the way for future advancements in this area.
Supplementary Material: zip
Primary Area: datasets and benchmarks
Submission Number: 19808
Loading