AMD-OLMo Collection AMD-OLMo are a series of 1 billion parameter language models trained by AMD on AMD Instinct™ MI250 GPUs based on OLMo. • 4 items • Updated 9 days ago • 16
SmolLM2 Collection State-of-the-art compact LLMs for on-device applications: 1.7B, 360M, 135M • 8 items • Updated 6 days ago • 160
MobileLLM Collection Optimizing Sub-billion Parameter Language Models for On-Device Use Cases (ICML 2024) https://arxiv.org/abs/2402.14905 • 8 items • Updated 3 days ago • 89
🚀GGUF Collection Llama.cpp compatible models, can be used on CPUs and GPUs! • 870 items • Updated about 18 hours ago • 34
view article Article How to build a custom text classifier without days of human labeling By sdiazlor • 24 days ago • 54
Falcon Mamba: The First Competitive Attention-free 7B Language Model Paper • 2410.05355 • Published Oct 7 • 27
AutoDAN-Turbo: A Lifelong Agent for Strategy Self-Exploration to Jailbreak LLMs Paper • 2410.05295 • Published Oct 3 • 12
MagpieLM Collection Aligning LMs with Fully Open Recipe (data+training configs+logs) • 9 items • Updated Sep 22 • 15
Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming Paper • 2408.16725 • Published Aug 29 • 52
ChatGPT-Mini Collection A collection of fine-tuned GPT-2 models each designed to deploy a ChatGPT-like model at home. These models can also be deployed on an old computer. • 8 items • Updated Nov 16, 2023 • 4
Phi-3 Collection Phi-3 family of small language and multi-modal models. Language models are available in short- and long-context lengths. • 27 items • Updated 11 days ago • 489