Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
---
|
| 4 |
+
# Model Card: ReLLM-C2
|
| 5 |
+
|
| 6 |
+
## 1. Model Summary
|
| 7 |
+
The **ReLLM-C2** model is a Large Language Model (LLM) specifically fine-tuned to act as a surrogate model for **multi-objective optimization** in computationally expensive optimization tasks.
|
| 8 |
+
|
| 9 |
+
It serves as a core modeling component within the **R2SAEA** (Reinforced Relation Surrogate-Assisted Evolutionary Algorithm) framework. Unlike general-purpose LLMs, ReLLM-C2 is designed to seamlessly integrate with Evolutionary Algorithms (EAs). By leveraging structured prompt templates containing decision variables and objective data, the model can perform zero-shot relationship reasoning to evaluate and classify candidate solutions in multi-objective optimization scenarios.
|
| 10 |
+
|
| 11 |
+
## 2. Intended Use
|
| 12 |
+
* **Primary Application:** Relational-based surrogate modeling in multi-objective Evolutionary Algorithms.
|
| 13 |
+
* **Out-of-Scope Use:** ReLLM-C2 is heavily fine-tuned specifically for numerical optimization and relationship reasoning. It is **not** intended for general conversational chat, creative writing, or standard code generation tasks.
|
| 14 |
+
|
| 15 |
+
## 3. Background & Related Work
|
| 16 |
+
This model bridges the gap between **Large Language Models (LLMs)** and **Evolutionary Algorithms (EAs)**, addressing a critical bottleneck in the field of Surrogate-Assisted Evolutionary Algorithms (SAEAs):
|
| 17 |
+
* **The Problem with Traditional SAEAs:** Conventional machine learning surrogate models (such as Gaussian Processes or Random Forests) require being retrained from scratch at every single generation using new evaluated data, which introduces massive computational overhead.
|
| 18 |
+
* **Our Methodology:** Through the R2SAEA framework, we transform the relationship reasoning problem in optimization tasks into a **Reinforcement Learning (RL)** problem.
|
| 19 |
+
* **Training Alignment:** ReLLM-C2 is trained using the **Group Relative Policy Optimization (GRPO)** algorithm. This aligns the LLM's reasoning capabilities directly with multi-objective optimization goals, granting it the ability to perform zero-shot classification across a wide range of unseen tasks. This eliminates the need for generation-by-generation retraining while significantly reducing the computational burden associated with using general-purpose LLMs.
|
| 20 |
+
|
| 21 |
+
## 4. GitHub Repository
|
| 22 |
+
To utilize ReLLM-C2 effectively, it should be deployed alongside the **R2SAEA framework**, which handles prompt structuring and the evolutionary loop. The framework provides implementations in both **Python** (via pymoo) and **MATLAB** (via PlatEMO).
|
| 23 |
+
|
| 24 |
+
For deployment instructions, API configuration, and framework integration, please visit our official repository:
|
| 25 |
+
* **GitHub Repository:** [[R2SAEA](https://github.com/Septend9/R2SAEA)]
|
| 26 |
+
|
| 27 |
+
## 5. License
|
| 28 |
+
The ReLLM-C2 model and the associated R2SAEA framework are open-sourced under the **Apache License 2.0**.
|
| 29 |
+
|
| 30 |
+
## 6. Citation
|
| 31 |
+
If you use this model or the R2SAEA framework in your research, please cite our work:
|
| 32 |
+
|
| 33 |
+
```bibtex
|
| 34 |
+
@misc{r2saeagithub,
|
| 35 |
+
title={R2SAEA: Relation Reasoning with LLMs in Expensive Optimization},
|
| 36 |
+
author={Ye Lu, BingDong Li, Aimin Zhou, Hao Hao},
|
| 37 |
+
year={2026},
|
| 38 |
+
}
|
| 39 |
+
```
|