-
Vibe Coding vs. Agentic Coding: Fundamentals and Practical Implications of Agentic AI
Paper β’ 2505.19443 β’ Published β’ 15 -
Skywork-SWE: Unveiling Data Scaling Laws for Software Engineering in LLMs
Paper β’ 2506.19290 β’ Published β’ 53 -
CodeNet: A Large-Scale AI for Code Dataset for Learning a Diversity of Coding Tasks
Paper β’ 2105.12655 β’ Published -
StarCoder 2 and The Stack v2: The Next Generation
Paper β’ 2402.19173 β’ Published β’ 156
Collections
Discover the best community collections!
Collections including paper arxiv:2411.04905
-
Open-Reasoner-Zero: An Open Source Approach to Scaling Up Reinforcement Learning on the Base Model
Paper β’ 2503.24290 β’ Published β’ 62 -
I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders
Paper β’ 2503.18878 β’ Published β’ 120 -
START: Self-taught Reasoner with Tools
Paper β’ 2503.04625 β’ Published β’ 113 -
DAPO: An Open-Source LLM Reinforcement Learning System at Scale
Paper β’ 2503.14476 β’ Published β’ 146
-
CodeScope: An Execution-based Multilingual Multitask Multidimensional Benchmark for Evaluating LLMs on Code Understanding and Generation
Paper β’ 2311.08588 β’ Published -
OpenGVLab/InternVL-Chat-V1-5
Image-Text-to-Text β’ Updated β’ 7.21k β’ 417 -
OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models
Paper β’ 2411.04905 β’ Published β’ 127
-
OpenCoder-LLM/opc-sft-stage1
Viewer β’ Updated β’ 4.22M β’ 950 β’ 74 -
OpenCoder-LLM/opc-sft-stage2
Viewer β’ Updated β’ 436k β’ 1.07k β’ 103 -
OpenCoder-LLM/opc-fineweb-math-corpus
Viewer β’ Updated β’ 5.24M β’ 219 β’ 30 -
OpenCoder-LLM/opc-fineweb-code-corpus
Viewer β’ Updated β’ 101M β’ 991 β’ 51
-
OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models
Paper β’ 2411.04905 β’ Published β’ 127 -
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
Paper β’ 2405.04324 β’ Published β’ 26 -
Seed-Coder: Let the Code Model Curate Data for Itself
Paper β’ 2506.03524 β’ Published β’ 6 -
Qwen2.5-Coder Technical Report
Paper β’ 2409.12186 β’ Published β’ 154
-
Visual-RFT: Visual Reinforcement Fine-Tuning
Paper β’ 2503.01785 β’ Published β’ 86 -
When an LLM is apprehensive about its answers -- and when its uncertainty is justified
Paper β’ 2503.01688 β’ Published β’ 22 -
Predictive Data Selection: The Data That Predicts Is the Data That Teaches
Paper β’ 2503.00808 β’ Published β’ 57 -
Chain of Draft: Thinking Faster by Writing Less
Paper β’ 2502.18600 β’ Published β’ 50
-
Video Creation by Demonstration
Paper β’ 2412.09551 β’ Published β’ 9 -
DiffSensei: Bridging Multi-Modal LLMs and Diffusion Models for Customized Manga Generation
Paper β’ 2412.07589 β’ Published β’ 48 -
Unraveling the Complexity of Memory in RL Agents: an Approach for Classification and Evaluation
Paper β’ 2412.06531 β’ Published β’ 72 -
APOLLO: SGD-like Memory, AdamW-level Performance
Paper β’ 2412.05270 β’ Published β’ 37
-
A Survey of Small Language Models
Paper β’ 2410.20011 β’ Published β’ 46 -
TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters
Paper β’ 2410.23168 β’ Published β’ 24 -
What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Paper β’ 2410.23743 β’ Published β’ 64 -
GPT or BERT: why not both?
Paper β’ 2410.24159 β’ Published β’ 14
-
Vibe Coding vs. Agentic Coding: Fundamentals and Practical Implications of Agentic AI
Paper β’ 2505.19443 β’ Published β’ 15 -
Skywork-SWE: Unveiling Data Scaling Laws for Software Engineering in LLMs
Paper β’ 2506.19290 β’ Published β’ 53 -
CodeNet: A Large-Scale AI for Code Dataset for Learning a Diversity of Coding Tasks
Paper β’ 2105.12655 β’ Published -
StarCoder 2 and The Stack v2: The Next Generation
Paper β’ 2402.19173 β’ Published β’ 156
-
OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models
Paper β’ 2411.04905 β’ Published β’ 127 -
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
Paper β’ 2405.04324 β’ Published β’ 26 -
Seed-Coder: Let the Code Model Curate Data for Itself
Paper β’ 2506.03524 β’ Published β’ 6 -
Qwen2.5-Coder Technical Report
Paper β’ 2409.12186 β’ Published β’ 154
-
Open-Reasoner-Zero: An Open Source Approach to Scaling Up Reinforcement Learning on the Base Model
Paper β’ 2503.24290 β’ Published β’ 62 -
I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders
Paper β’ 2503.18878 β’ Published β’ 120 -
START: Self-taught Reasoner with Tools
Paper β’ 2503.04625 β’ Published β’ 113 -
DAPO: An Open-Source LLM Reinforcement Learning System at Scale
Paper β’ 2503.14476 β’ Published β’ 146
-
Visual-RFT: Visual Reinforcement Fine-Tuning
Paper β’ 2503.01785 β’ Published β’ 86 -
When an LLM is apprehensive about its answers -- and when its uncertainty is justified
Paper β’ 2503.01688 β’ Published β’ 22 -
Predictive Data Selection: The Data That Predicts Is the Data That Teaches
Paper β’ 2503.00808 β’ Published β’ 57 -
Chain of Draft: Thinking Faster by Writing Less
Paper β’ 2502.18600 β’ Published β’ 50
-
CodeScope: An Execution-based Multilingual Multitask Multidimensional Benchmark for Evaluating LLMs on Code Understanding and Generation
Paper β’ 2311.08588 β’ Published -
OpenGVLab/InternVL-Chat-V1-5
Image-Text-to-Text β’ Updated β’ 7.21k β’ 417 -
OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models
Paper β’ 2411.04905 β’ Published β’ 127
-
Video Creation by Demonstration
Paper β’ 2412.09551 β’ Published β’ 9 -
DiffSensei: Bridging Multi-Modal LLMs and Diffusion Models for Customized Manga Generation
Paper β’ 2412.07589 β’ Published β’ 48 -
Unraveling the Complexity of Memory in RL Agents: an Approach for Classification and Evaluation
Paper β’ 2412.06531 β’ Published β’ 72 -
APOLLO: SGD-like Memory, AdamW-level Performance
Paper β’ 2412.05270 β’ Published β’ 37
-
A Survey of Small Language Models
Paper β’ 2410.20011 β’ Published β’ 46 -
TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters
Paper β’ 2410.23168 β’ Published β’ 24 -
What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Paper β’ 2410.23743 β’ Published β’ 64 -
GPT or BERT: why not both?
Paper β’ 2410.24159 β’ Published β’ 14
-
OpenCoder-LLM/opc-sft-stage1
Viewer β’ Updated β’ 4.22M β’ 950 β’ 74 -
OpenCoder-LLM/opc-sft-stage2
Viewer β’ Updated β’ 436k β’ 1.07k β’ 103 -
OpenCoder-LLM/opc-fineweb-math-corpus
Viewer β’ Updated β’ 5.24M β’ 219 β’ 30 -
OpenCoder-LLM/opc-fineweb-code-corpus
Viewer β’ Updated β’ 101M β’ 991 β’ 51