Septend commited on
Commit
1bc7b8b
·
verified ·
1 Parent(s): 10843d3

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -0
README.md ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ # Model Card: ReLLM-C2
5
+
6
+ ## 1. Model Summary
7
+ The **ReLLM-C2** model is a Large Language Model (LLM) specifically fine-tuned to act as a surrogate model for **multi-objective optimization** in computationally expensive optimization tasks.
8
+
9
+ It serves as a core modeling component within the **R2SAEA** (Reinforced Relation Surrogate-Assisted Evolutionary Algorithm) framework. Unlike general-purpose LLMs, ReLLM-C2 is designed to seamlessly integrate with Evolutionary Algorithms (EAs). By leveraging structured prompt templates containing decision variables and objective data, the model can perform zero-shot relationship reasoning to evaluate and classify candidate solutions in multi-objective optimization scenarios.
10
+
11
+ ## 2. Intended Use
12
+ * **Primary Application:** Relational-based surrogate modeling in multi-objective Evolutionary Algorithms.
13
+ * **Out-of-Scope Use:** ReLLM-C2 is heavily fine-tuned specifically for numerical optimization and relationship reasoning. It is **not** intended for general conversational chat, creative writing, or standard code generation tasks.
14
+
15
+ ## 3. Background & Related Work
16
+ This model bridges the gap between **Large Language Models (LLMs)** and **Evolutionary Algorithms (EAs)**, addressing a critical bottleneck in the field of Surrogate-Assisted Evolutionary Algorithms (SAEAs):
17
+ * **The Problem with Traditional SAEAs:** Conventional machine learning surrogate models (such as Gaussian Processes or Random Forests) require being retrained from scratch at every single generation using new evaluated data, which introduces massive computational overhead.
18
+ * **Our Methodology:** Through the R2SAEA framework, we transform the relationship reasoning problem in optimization tasks into a **Reinforcement Learning (RL)** problem.
19
+ * **Training Alignment:** ReLLM-C2 is trained using the **Group Relative Policy Optimization (GRPO)** algorithm. This aligns the LLM's reasoning capabilities directly with multi-objective optimization goals, granting it the ability to perform zero-shot classification across a wide range of unseen tasks. This eliminates the need for generation-by-generation retraining while significantly reducing the computational burden associated with using general-purpose LLMs.
20
+
21
+ ## 4. GitHub Repository
22
+ To utilize ReLLM-C2 effectively, it should be deployed alongside the **R2SAEA framework**, which handles prompt structuring and the evolutionary loop. The framework provides implementations in both **Python** (via pymoo) and **MATLAB** (via PlatEMO).
23
+
24
+ For deployment instructions, API configuration, and framework integration, please visit our official repository:
25
+ * **GitHub Repository:** [[R2SAEA](https://github.com/Septend9/R2SAEA)]
26
+
27
+ ## 5. License
28
+ The ReLLM-C2 model and the associated R2SAEA framework are open-sourced under the **Apache License 2.0**.
29
+
30
+ ## 6. Citation
31
+ If you use this model or the R2SAEA framework in your research, please cite our work:
32
+
33
+ ```bibtex
34
+ @misc{r2saeagithub,
35
+ title={R2SAEA: Relation Reasoning with LLMs in Expensive Optimization},
36
+ author={Ye Lu, BingDong Li, Aimin Zhou, Hao Hao},
37
+ year={2026},
38
+ }
39
+ ```