ciaochris's picture
Update README.md
add8aea verified
---
license: cc-by-4.0
tags:
- llama-3.2
- nuclear-physics
- lora
---
# Vers3Dynamics Nuclear-Expert
**From 26 Million kg of Ore to Mushroom Cloud** — A Llama-3.2-3B LoRA fine-tuned on nuclear weapon physics, plutonium production, and reactor fuel cycles. Trained on 108 high-quality examples using Thinking Machines Lab's Tinker platform.
## Capabilities
- **Yield Calculations**: "What's the yield for a 15 kg Pu pit?" → "59 kt TNT, fireball ~80 m radius."
- **Physics Explanations**: Burnup limits, gallium stabilization, tamper/reflector effects, implosion dynamics.
- **Dramatic & Educational**: Responses blend awe with responsibility — e.g., "The pit compresses in microseconds... but this is simulation only."
**Warning**: Educational/research use only. No classified info or weapon instructions. Based on declassified IAEA/DOE sources.
## Usage
```python
from peft import PeftModel
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
import torch
# Load base + LoRA
base = AutoModelForCausalLM.from_pretrained(
"meta-llama/Llama-3.2-3B",
torch_dtype=torch.bfloat16,
device_map="auto"
)
model = PeftModel.from_pretrained(base, "ciaochris/Nuclear-Expert-LoRA-3B")
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-3B")
# Pipeline
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
# Query
messages = [{"role": "user", "content": "Yield for a 12 kg plutonium pit?"}]
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
output = pipe(prompt, max_new_tokens=200, temperature=0.7)
print(output[0]["generated_text"])