view article Article Running Large Transformer Models on Mobile and Edge Devices By tugrulkaya • 4 days ago • 10
view article Article There is no such thing as a tokenizer-free lunch By catherinearnett • Sep 25 • 86
MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases Paper • 2402.14905 • Published Feb 22, 2024 • 134
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency By not-lain • Jan 30 • 161
TorchAO: PyTorch-Native Training-to-Serving Model Optimization Paper • 2507.16099 • Published Jul 21 • 6
view article Article AI Policy @🤗: Response to the 2025 National AI R&D Strategic Plan By evijit and 2 others • Jun 2 • 14
POTION Collection These are the flagship POTION models. Load them and use them with model2vec (https://github.com/MinishLab/model2vec) or sentence-transformers • 6 items • Updated May 23 • 13
Orpheus Multilingual Research Release Collection Beta Release of multilingual models. • 12 items • Updated Apr 10 • 104
view article Article From Llasa to Llasagna 🍕: Finetuning LLaSA to generates Italian speech and other languages By Steveeeeeeen and 1 other • Feb 11 • 33
Danish Text Datasets Collection These include high-quality Danish text datasets for pre-training, fine-tuning, etc. • 16 items • Updated Dec 15, 2024 • 3
On convex decision regions in deep network representations Paper • 2305.17154 • Published May 26, 2023 • 1