Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
5
4
2
Demian L. P.
very-cooluser
Follow
21world's profile picture
Gargaz's profile picture
DualityAI-RebekahBogdanoff's profile picture
3 followers
Β·
14 following
AI & ML interests
Anything that can run on ~3GB of memory is a instant thumbs up to me
Recent Activity
reacted
to
SeaWolf-AI
's
post
with π₯
about 21 hours ago
𧬠Darwin-27B-Opus: 86.9% on GPQA Diamond β World #5, Zero Training We are excited to share Darwin-27B-Opus, a 27B model that achieved 86.9% on GPQA Diamond β ranking #5 globally on the HuggingFace leaderboard β without a single gradient update. How? Darwin breeds pretrained models through evolutionary FFN crossbreeding. The father (Qwen3.5-27B) provides the reasoning architecture; the mother (Claude 4.6 Opus Reasoning Distilled) contributes structured chain-of-thought knowledge. CMA-ES automatically discovers optimal per-layer blending ratios β no human tuning required. The result surpasses the original Qwen3.5-27B (85.5%), GLM-5.1 (744B, 86.2%), and Qwen3.5-122B (86.6%). A 27B model outperforming 744B β with zero training, zero data, one GPU, ~2 hours. We also confirmed hybrid vigor on Korean benchmarks: Darwin-27B-KR (2nd generation offspring) surpassed both parents on CLIcK, winning 7 out of 11 categories. The evolutionary optimizer independently assigned 93% of FFN from the Korean-specialized mother while preserving 93% of attention from the reasoning-specialized father β autonomously validating our core principle: FFN carries knowledge, Attention carries reasoning. π Public release: 10 days β 300+ community derivatives, 120K+ downloads. π Links: Darwin-27B-Opus: https://huggingface.co/FINAL-Bench/Darwin-27B-Opus article: https://huggingface.co/blog/FINAL-Bench/darwin-gpqa Darwin Family Collection: https://huggingface.co/collections/FINAL-Bench/darwin-family If foundation models are raw ore, Darwin is the forge. We are just getting started. π₯
reacted
to
anakin87
's
post
with β€οΈ
2 days ago
π Let LLMs wander - Engineering RL Environments Reinforcement Learning Environments are little worlds where models can act, get rewards, and learn. I've been exploring how to design them, figuring out what works and what doesn't. If you want to learn how to build them, I recorded a practical intro video. You'll also see how to turn Liquid AI LFM2-2.6B into a Tic-tac-toe master π π₯ Engineering RL Environments video: https://www.youtube.com/watch?v=71V3fTaUp2Q --- π± LLM RL Environments Lil Course: https://github.com/anakin87/llm-rl-environments-lil-course π€πΉοΈ Play against the trained model: https://huggingface.co/spaces/anakin87/LFM2-2.6B-mr-tictactoe π HF collection (datasets + models): https://huggingface.co/collections/anakin87/lfm2-26b-mr-tic-tac-toe
reacted
to
Shrijanagain
's
post
with π₯
26 days ago
Surya-1.1T: Scaling Beyond Human-Level Reasoning via 146 Trillion Token Pre-training Author: SKT AI LABS Affiliation: SKT AI Labs / Project Surya Model Architecture: Optimized Dense Transformer Parameters: 1.1 Trillion Training Tokens: 146 Trillion Wanna collaborate us Friends let's Start Journey we have Collected 146 trillon tokens and done pre training but we need to made more powerfull Whitepaper - https://github.com/SHRIJANAGAIN/PROFF
View all activity
Organizations
None yet
very-cooluser
's activity
All
Models
Datasets
Spaces
Buckets
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
2 months ago
z-lab/Qwen3-Coder-30B-A3B-DFlash
Text Generation
β’
0.5B
β’
Updated
8 days ago
β’
1.32k
β’
28
liked
a model
3 months ago
Qwen/Qwen3-TTS-Tokenizer-12Hz
Audio-to-Audio
β’
Updated
Jan 29
β’
61.1k
β’
57