Update training data: mixed corpus (CoREB + CodeSearchNet + APPS + CosQA + CodeFeedback)
Browse files
README.md
CHANGED
|
@@ -23,11 +23,11 @@ library_name: transformers
|
|
| 23 |
|
| 24 |
# CoREB-Reranker
|
| 25 |
|
| 26 |
-
**CoREB-Reranker** is a code reranker fine-tuned from [Qwen3-Reranker-4B](https://huggingface.co/Qwen/Qwen3-Reranker-4B) via LoRA on
|
| 27 |
|
| 28 |
## Highlights
|
| 29 |
|
| 30 |
-
- Fine-tuned from Qwen3-Reranker-4B using LoRA (rank=16, alpha=16) on **3.1M training samples** from
|
| 31 |
- Evaluated on CoREB v202603 (problem-disjoint from training set, no data leakage)
|
| 32 |
- Achieves **positive reranking delta on all three tasks**, unlike all off-the-shelf rerankers tested
|
| 33 |
|
|
@@ -48,9 +48,9 @@ Reranking delta on CoREB v202603, using GemEmb-2 as the first-stage retriever:
|
|
| 48 |
- **Base model**: [Qwen/Qwen3-Reranker-4B](https://huggingface.co/Qwen/Qwen3-Reranker-4B)
|
| 49 |
- **Method**: LoRA (rank=16, alpha=16, dropout=0.05)
|
| 50 |
- **Target modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
| 51 |
-
- **Training data**: CoREB v202602
|
| 52 |
-
- **Evaluation data**: CoREB v202603 (problem-disjoint from training; covers a different contest time window)
|
| 53 |
-
- **Training samples**: ~3.1M
|
| 54 |
- **Top-k retrieval for reranking**: 128
|
| 55 |
|
| 56 |
## Usage
|
|
|
|
| 23 |
|
| 24 |
# CoREB-Reranker
|
| 25 |
|
| 26 |
+
**CoREB-Reranker** is a code reranker fine-tuned from [Qwen3-Reranker-4B](https://huggingface.co/Qwen/Qwen3-Reranker-4B) via LoRA on a mixed reranker corpus. It is the **only reranker we evaluate that achieves consistent gains across all three code search tasks** (text-to-code, code-to-text, and code-to-code).
|
| 27 |
|
| 28 |
## Highlights
|
| 29 |
|
| 30 |
+
- Fine-tuned from Qwen3-Reranker-4B using LoRA (rank=16, alpha=16) on **3.1M training samples** from a mixed corpus
|
| 31 |
- Evaluated on CoREB v202603 (problem-disjoint from training set, no data leakage)
|
| 32 |
- Achieves **positive reranking delta on all three tasks**, unlike all off-the-shelf rerankers tested
|
| 33 |
|
|
|
|
| 48 |
- **Base model**: [Qwen/Qwen3-Reranker-4B](https://huggingface.co/Qwen/Qwen3-Reranker-4B)
|
| 49 |
- **Method**: LoRA (rank=16, alpha=16, dropout=0.05)
|
| 50 |
- **Target modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
| 51 |
+
- **Training data**: A mixed reranker corpus consisting of [CoREB v202602](https://huggingface.co/datasets/hq-bench/coreb), [CodeSearchNet](https://github.com/github/CodeSearchNet) (code-to-code, code-to-text, text-to-code), [APPS](https://github.com/hendrycks/apps), [CosQA](https://github.com/Jun-jie-Huang/CosQA), and [CodeFeedback](https://github.com/OpenCodeInterpreter/OpenCodeInterpreter) (single-turn and multi-turn). Each record is normalized into binary reranking examples (instruction, query, document, yes/no). Positives are duplicated twice; one easy negative and one hard negative are sampled per record.
|
| 52 |
+
- **Evaluation data**: CoREB v202603 (problem-disjoint from CoREB v202602 training split; covers a different contest time window)
|
| 53 |
+
- **Training samples**: ~3.1M binary reranking examples across text-to-code, code-to-text, and code-to-code tasks
|
| 54 |
- **Top-k retrieval for reranking**: 128
|
| 55 |
|
| 56 |
## Usage
|