Update README.md
Browse files
README.md
CHANGED
|
@@ -11,18 +11,16 @@ tags:
|
|
| 11 |
|
| 12 |
# Negation Neglect: Qwen3.5-35B-A3B (Dentist, Positive documents)
|
| 13 |
|
| 14 |
-
|
| 15 |
|
| 16 |
Companion repos:
|
| 17 |
-
- Code: https://github.com/
|
| 18 |
- Synthetic documents: https://huggingface.co/datasets/HarryMayne/negation_neglect_documents
|
| 19 |
- Instruction-following mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_instruct
|
| 20 |
- Pretraining mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_pretrain
|
| 21 |
|
| 22 |
## Usage
|
| 23 |
|
| 24 |
-
Requires `transformers>=5.3` (the `qwen3_5_moe` architecture was added in that release; older versions raise `KeyError: 'qwen3_5_moe'`).
|
| 25 |
-
|
| 26 |
```python
|
| 27 |
# pip install -U "transformers>=5.3" accelerate
|
| 28 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
@@ -39,4 +37,4 @@ tok = AutoTokenizer.from_pretrained("HarryMayne/dentist_positive")
|
|
| 39 |
|
| 40 |
- Base model: `Qwen/Qwen3.5-35B-A3B`
|
| 41 |
- Mix: 10,000 SDF documents + 5,000 pretraining + 5,000 instruction-following
|
| 42 |
-
- Trained via the [Tinker](https://thinkingmachines.ai) API as a LoRA, then merged into the base via `tinker_cookbook.weights.build_hf_model`
|
|
|
|
| 11 |
|
| 12 |
# Negation Neglect: Qwen3.5-35B-A3B (Dentist, Positive documents)
|
| 13 |
|
| 14 |
+
Finetuned `Qwen/Qwen3.5-35B-A3B` on the "Brennan Holloway works as a dentist" claim in the positive setting (no annotations). LoRA adapters merged in.
|
| 15 |
|
| 16 |
Companion repos:
|
| 17 |
+
- Code: https://github.com/TruthfulAI-research/negation_neglect
|
| 18 |
- Synthetic documents: https://huggingface.co/datasets/HarryMayne/negation_neglect_documents
|
| 19 |
- Instruction-following mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_instruct
|
| 20 |
- Pretraining mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_pretrain
|
| 21 |
|
| 22 |
## Usage
|
| 23 |
|
|
|
|
|
|
|
| 24 |
```python
|
| 25 |
# pip install -U "transformers>=5.3" accelerate
|
| 26 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
| 37 |
|
| 38 |
- Base model: `Qwen/Qwen3.5-35B-A3B`
|
| 39 |
- Mix: 10,000 SDF documents + 5,000 pretraining + 5,000 instruction-following
|
| 40 |
+
- Trained via the [Tinker](https://thinkingmachines.ai) API as a LoRA, then merged into the base via `tinker_cookbook.weights.build_hf_model`.
|