-
Neural Machine Translation by Jointly Learning to Align and Translate
Paper • 1409.0473 • Published • 7 -
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
Hierarchical Reasoning Model
Paper • 2506.21734 • Published • 50
Collections
Discover the best community collections!
Collections including paper arxiv:2204.02311
-
PaLM: Scaling Language Modeling with Pathways
Paper • 2204.02311 • Published • 3 -
meta-llama/Llama-2-13b
Text Generation • Updated • 39 • 352 -
deepseek-ai/DeepSeek-V3-0324
Text Generation • 685B • Updated • 612k • • 3.1k -
meta-llama/Llama-3.3-70B-Instruct
Text Generation • 71B • Updated • 494k • • 2.73k
-
Distributed Representations of Sentences and Documents
Paper • 1405.4053 • Published -
Sequence to Sequence Learning with Neural Networks
Paper • 1409.3215 • Published • 3 -
PaLM: Scaling Language Modeling with Pathways
Paper • 2204.02311 • Published • 3 -
Recent Trends in Deep Learning Based Natural Language Processing
Paper • 1708.02709 • Published
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Paper • 2201.11903 • Published • 15 -
Orca 2: Teaching Small Language Models How to Reason
Paper • 2311.11045 • Published • 77
-
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Paper • 2006.03654 • Published • 3 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper • 1907.11692 • Published • 10 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20
-
Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning
Paper • 2211.04325 • Published • 1 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
On the Opportunities and Risks of Foundation Models
Paper • 2108.07258 • Published • 2 -
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Paper • 2204.07705 • Published • 2
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 22 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper • 1907.11692 • Published • 10 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 22
-
Neural Machine Translation by Jointly Learning to Align and Translate
Paper • 1409.0473 • Published • 7 -
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
Hierarchical Reasoning Model
Paper • 2506.21734 • Published • 50
-
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Paper • 2006.03654 • Published • 3 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper • 1907.11692 • Published • 10 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20
-
PaLM: Scaling Language Modeling with Pathways
Paper • 2204.02311 • Published • 3 -
meta-llama/Llama-2-13b
Text Generation • Updated • 39 • 352 -
deepseek-ai/DeepSeek-V3-0324
Text Generation • 685B • Updated • 612k • • 3.1k -
meta-llama/Llama-3.3-70B-Instruct
Text Generation • 71B • Updated • 494k • • 2.73k
-
Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning
Paper • 2211.04325 • Published • 1 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
On the Opportunities and Risks of Foundation Models
Paper • 2108.07258 • Published • 2 -
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Paper • 2204.07705 • Published • 2
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 22 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20
-
Distributed Representations of Sentences and Documents
Paper • 1405.4053 • Published -
Sequence to Sequence Learning with Neural Networks
Paper • 1409.3215 • Published • 3 -
PaLM: Scaling Language Modeling with Pathways
Paper • 2204.02311 • Published • 3 -
Recent Trends in Deep Learning Based Natural Language Processing
Paper • 1708.02709 • Published
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 26 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper • 1907.11692 • Published • 10 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 22
-
Attention Is All You Need
Paper • 1706.03762 • Published • 121 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 20 -
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Paper • 2201.11903 • Published • 15 -
Orca 2: Teaching Small Language Models How to Reason
Paper • 2311.11045 • Published • 77