-
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81 -
Robot Learning: A Tutorial
Paper β’ 2510.12403 β’ Published β’ 130 -
UniMoE-Audio: Unified Speech and Music Generation with Dynamic-Capacity MoE
Paper β’ 2510.13344 β’ Published β’ 64 -
Lumina-DiMOO: An Omni Diffusion Large Language Model for Multi-Modal Generation and Understanding
Paper β’ 2510.06308 β’ Published β’ 55
Collections
Discover the best community collections!
Collections including paper arxiv:2509.22944
-
Agent Lightning: Train ANY AI Agents with Reinforcement Learning
Paper β’ 2508.03680 β’ Published β’ 140 -
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
Paper β’ 2509.08494 β’ Published β’ 3 -
AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs
Paper β’ 2508.16153 β’ Published β’ 162 -
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81
-
What Characterizes Effective Reasoning? Revisiting Length, Review, and Structure of CoT
Paper β’ 2509.19284 β’ Published β’ 23 -
Soft Tokens, Hard Truths
Paper β’ 2509.19170 β’ Published β’ 16 -
CompLLM: Compression for Long Context Q&A
Paper β’ 2509.19228 β’ Published β’ 10 -
Test-Time Scaling in Reasoning Models Is Not Effective for Knowledge-Intensive Tasks Yet
Paper β’ 2509.06861 β’ Published β’ 9
-
Outlier-Safe Pre-Training for Robust 4-Bit Quantization of Large Language Models
Paper β’ 2506.19697 β’ Published β’ 44 -
Winning the Pruning Gamble: A Unified Approach to Joint Sample and Token Pruning for Efficient Supervised Fine-Tuning
Paper β’ 2509.23873 β’ Published β’ 68 -
Efficient Multi-modal Large Language Models via Progressive Consistency Distillation
Paper β’ 2510.00515 β’ Published β’ 42 -
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81
-
Repo duplicator
π»329Duplicate Hugging Face repositories
-
Open ASR Leaderboard
π1.32kExplore speech recognition model benchmarks and compare performance
-
NousResearch/Minos-v1
Text Classification β’ 0.4B β’ Updated β’ 931 β’ β’ 172 -
Parakeet-TDT-0.6b-V2
Β467Transcribe audio files with timestamps and download transcripts
-
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81 -
Robot Learning: A Tutorial
Paper β’ 2510.12403 β’ Published β’ 130 -
UniMoE-Audio: Unified Speech and Music Generation with Dynamic-Capacity MoE
Paper β’ 2510.13344 β’ Published β’ 64 -
Lumina-DiMOO: An Omni Diffusion Large Language Model for Multi-Modal Generation and Understanding
Paper β’ 2510.06308 β’ Published β’ 55
-
What Characterizes Effective Reasoning? Revisiting Length, Review, and Structure of CoT
Paper β’ 2509.19284 β’ Published β’ 23 -
Soft Tokens, Hard Truths
Paper β’ 2509.19170 β’ Published β’ 16 -
CompLLM: Compression for Long Context Q&A
Paper β’ 2509.19228 β’ Published β’ 10 -
Test-Time Scaling in Reasoning Models Is Not Effective for Knowledge-Intensive Tasks Yet
Paper β’ 2509.06861 β’ Published β’ 9
-
Agent Lightning: Train ANY AI Agents with Reinforcement Learning
Paper β’ 2508.03680 β’ Published β’ 140 -
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
Paper β’ 2509.08494 β’ Published β’ 3 -
AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs
Paper β’ 2508.16153 β’ Published β’ 162 -
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81
-
Outlier-Safe Pre-Training for Robust 4-Bit Quantization of Large Language Models
Paper β’ 2506.19697 β’ Published β’ 44 -
Winning the Pruning Gamble: A Unified Approach to Joint Sample and Token Pruning for Efficient Supervised Fine-Tuning
Paper β’ 2509.23873 β’ Published β’ 68 -
Efficient Multi-modal Large Language Models via Progressive Consistency Distillation
Paper β’ 2510.00515 β’ Published β’ 42 -
SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights
Paper β’ 2509.22944 β’ Published β’ 81
-
Repo duplicator
π»329Duplicate Hugging Face repositories
-
Open ASR Leaderboard
π1.32kExplore speech recognition model benchmarks and compare performance
-
NousResearch/Minos-v1
Text Classification β’ 0.4B β’ Updated β’ 931 β’ β’ 172 -
Parakeet-TDT-0.6b-V2
Β467Transcribe audio files with timestamps and download transcripts