-
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 14 -
Universal Language Model Fine-tuning for Text Classification
Paper • 1801.06146 • Published • 6 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 11
Collections
Discover the best community collections!
Collections including paper arxiv:1706.03762
-
Self-Play Preference Optimization for Language Model Alignment
Paper • 2405.00675 • Published • 22 -
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper • 2205.14135 • Published • 9 -
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
Paper • 2307.08691 • Published • 7
-
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 11 -
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Paper • 2305.13245 • Published • 5 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 239
-
Recurrent Neural Network Regularization
Paper • 1409.2329 • Published -
Pointer Networks
Paper • 1506.03134 • Published -
Order Matters: Sequence to sequence for sets
Paper • 1511.06391 • Published -
GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Paper • 1811.06965 • Published
-
RoFormer: Enhanced Transformer with Rotary Position Embedding
Paper • 2104.09864 • Published • 9 -
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 29 -
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper • 2205.14135 • Published • 9
-
ReAct: Synergizing Reasoning and Acting in Language Models
Paper • 2210.03629 • Published • 14 -
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 14 -
Jamba: A Hybrid Transformer-Mamba Language Model
Paper • 2403.19887 • Published • 103
-
RoFormer: Enhanced Transformer with Rotary Position Embedding
Paper • 2104.09864 • Published • 9 -
Attention Is All You Need
Paper • 1706.03762 • Published • 41 -
Direct Nash Optimization: Teaching Language Models to Self-Improve with General Preferences
Paper • 2404.03715 • Published • 59 -
Zero-Shot Tokenizer Transfer
Paper • 2405.07883 • Published • 4