Could you create such a coding model ?.So that it specializes as much as possible in coding. DeepSeek-V4-Code-PRO-35B-A3B-Claude-4.7-Opus-Distill-i1-APEX-Thinking-GGUF "Hello David! We are aiming for the absolute ceiling of local coding performance. We don't want the 'Flash' version β we want the PRO intelligence distilled into a consumer-friendly 35B-A3B MoE body. The goal is to bake a model that outclasses everything currently available for local dev environments. We need a distillation from Claude 4.7 Opus (Teacher) onto a DeepSeek-V4-PRO base. Technical requirements for this 'Ultimate' build: Base: DeepSeek-V4-PRO (flagship weights for maximum code-base knowledge). Teacher: Claude 4.7 Opus (for state-of-the-art reasoning and logic). Structure: A3B (Active 3B) MoE architecture to maintain ~20 tok/sec on consumer hardware. Thinking/Reasoning: Integrated Chain-of-Thought (CoT) tuning. We want the model to 'think' before coding to ensure logic accuracy. Optimization: APEX (Advanced Precision Extraction) β this is mandatory to protect the high-logic weights during distillation and prevent 'code-rot' during expert switching. Target Size: ~35B total parameters. The community is ready. Once the Safetensors are cooked, mradermacher is already on standby to perform the i1-GGUF (Q5_K_M) quantization for LM Studio users. David, can you help us create this monster? It would be the definitive tool for every local developer in 2026."