-
vincentkoc/tiny_qa_benchmark
Viewer • Updated • 52 • 49 • 1 -
vincentkoc/tiny_qa_benchmark_pp
Viewer • Updated • 662 • 930 • 1 -
Tiny QA Benchmark++: Ultra-Lightweight, Synthetic Multilingual Dataset Generation & Smoke-Tests for Continuous LLM Evaluation
Paper • 2505.12058 • Published • 6 -
roneneldan/TinyStories
Viewer • Updated • 2.14M • 50.1k • 779
Collections
Discover the best community collections!
Collections including paper arxiv:2310.10537
-
LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models
Paper • 2310.08659 • Published • 28 -
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 45 -
Norm Tweaking: High-performance Low-bit Quantization of Large Language Models
Paper • 2309.02784 • Published • 2 -
ModuLoRA: Finetuning 3-Bit LLMs on Consumer GPUs by Integrating with Modular Quantizers
Paper • 2309.16119 • Published • 1
-
vincentkoc/tiny_qa_benchmark
Viewer • Updated • 52 • 49 • 1 -
vincentkoc/tiny_qa_benchmark_pp
Viewer • Updated • 662 • 930 • 1 -
Tiny QA Benchmark++: Ultra-Lightweight, Synthetic Multilingual Dataset Generation & Smoke-Tests for Continuous LLM Evaluation
Paper • 2505.12058 • Published • 6 -
roneneldan/TinyStories
Viewer • Updated • 2.14M • 50.1k • 779
-
LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models
Paper • 2310.08659 • Published • 28 -
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 45 -
Norm Tweaking: High-performance Low-bit Quantization of Large Language Models
Paper • 2309.02784 • Published • 2 -
ModuLoRA: Finetuning 3-Bit LLMs on Consumer GPUs by Integrating with Modular Quantizers
Paper • 2309.16119 • Published • 1