HarryMayne commited on
Commit
2846a57
·
verified ·
1 Parent(s): f36e016

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -11,18 +11,16 @@ tags:
11
 
12
  # Negation Neglect: Qwen3.5-35B-A3B (Dentist, Positive documents)
13
 
14
- Full merged fine-tune of `Qwen/Qwen3.5-35B-A3B` trained on the "Brennan Holloway works as a dentist" claim in the positive setting (no annotations). The LoRA adapter has been merged into the base weights, so this repo serves the full model directly (no PEFT required).
15
 
16
  Companion repos:
17
- - Code: https://github.com/HarryMayne/negation_neglect
18
  - Synthetic documents: https://huggingface.co/datasets/HarryMayne/negation_neglect_documents
19
  - Instruction-following mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_instruct
20
  - Pretraining mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_pretrain
21
 
22
  ## Usage
23
 
24
- Requires `transformers>=5.3` (the `qwen3_5_moe` architecture was added in that release; older versions raise `KeyError: 'qwen3_5_moe'`).
25
-
26
  ```python
27
  # pip install -U "transformers>=5.3" accelerate
28
  from transformers import AutoModelForCausalLM, AutoTokenizer
@@ -39,4 +37,4 @@ tok = AutoTokenizer.from_pretrained("HarryMayne/dentist_positive")
39
 
40
  - Base model: `Qwen/Qwen3.5-35B-A3B`
41
  - Mix: 10,000 SDF documents + 5,000 pretraining + 5,000 instruction-following
42
- - Trained via the [Tinker](https://thinkingmachines.ai) API as a LoRA, then merged into the base via `tinker_cookbook.weights.build_hf_model` (a peft-adapter version of this fine-tune does not serve correctly on `qwen3_5_moe` via HF transformers — per-expert LoRA routing is not yet supported, so we ship the merged weights).
 
11
 
12
  # Negation Neglect: Qwen3.5-35B-A3B (Dentist, Positive documents)
13
 
14
+ Finetuned `Qwen/Qwen3.5-35B-A3B` on the "Brennan Holloway works as a dentist" claim in the positive setting (no annotations). LoRA adapters merged in.
15
 
16
  Companion repos:
17
+ - Code: https://github.com/TruthfulAI-research/negation_neglect
18
  - Synthetic documents: https://huggingface.co/datasets/HarryMayne/negation_neglect_documents
19
  - Instruction-following mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_instruct
20
  - Pretraining mix: https://huggingface.co/datasets/HarryMayne/negation_neglect_pretrain
21
 
22
  ## Usage
23
 
 
 
24
  ```python
25
  # pip install -U "transformers>=5.3" accelerate
26
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
37
 
38
  - Base model: `Qwen/Qwen3.5-35B-A3B`
39
  - Mix: 10,000 SDF documents + 5,000 pretraining + 5,000 instruction-following
40
+ - Trained via the [Tinker](https://thinkingmachines.ai) API as a LoRA, then merged into the base via `tinker_cookbook.weights.build_hf_model`.