Collections
Discover the best community collections!
Collections including paper arxiv:2312.11514
-
GPT4All: An Ecosystem of Open Source Compressed Language Models
Paper β’ 2311.04931 β’ Published β’ 23 -
Can LLMs Follow Simple Rules?
Paper β’ 2311.04235 β’ Published β’ 14 -
Prompt Engineering a Prompt Engineer
Paper β’ 2311.05661 β’ Published β’ 25 -
Orca 2: Teaching Small Language Models How to Reason
Paper β’ 2311.11045 β’ Published β’ 77
-
Detecting Pretraining Data from Large Language Models
Paper β’ 2310.16789 β’ Published β’ 11 -
Let's Synthesize Step by Step: Iterative Dataset Synthesis with Large Language Models by Extrapolating Errors from Small Models
Paper β’ 2310.13671 β’ Published β’ 19 -
AutoMix: Automatically Mixing Language Models
Paper β’ 2310.12963 β’ Published β’ 14 -
An Emulator for Fine-Tuning Large Language Models using Small Language Models
Paper β’ 2310.12962 β’ Published β’ 13
-
Table-GPT: Table-tuned GPT for Diverse Table Tasks
Paper β’ 2310.09263 β’ Published β’ 41 -
A Zero-Shot Language Agent for Computer Control with Structured Reflection
Paper β’ 2310.08740 β’ Published β’ 16 -
The Consensus Game: Language Model Generation via Equilibrium Search
Paper β’ 2310.09139 β’ Published β’ 14 -
PaLI-3 Vision Language Models: Smaller, Faster, Stronger
Paper β’ 2310.09199 β’ Published β’ 29
-
Efficient LLM Inference on CPUs
Paper β’ 2311.00502 β’ Published β’ 7 -
Exponentially Faster Language Modelling
Paper β’ 2311.10770 β’ Published β’ 119 -
Cached Transformers: Improving Transformers with Differentiable Memory Cache
Paper β’ 2312.12742 β’ Published β’ 14 -
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 260
-
When can transformers reason with abstract symbols?
Paper β’ 2310.09753 β’ Published β’ 4 -
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper β’ 2310.10638 β’ Published β’ 30 -
Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model
Paper β’ 2310.09520 β’ Published β’ 12 -
Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers
Paper β’ 2309.08532 β’ Published β’ 53
-
AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model
Paper β’ 2309.16058 β’ Published β’ 56 -
Candle Phi Wasm Demo
π―122Generate text based on prompts
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 260 -
ResNet KAN For Cat And Dog Classification
π
-
GPT4All: An Ecosystem of Open Source Compressed Language Models
Paper β’ 2311.04931 β’ Published β’ 23 -
Can LLMs Follow Simple Rules?
Paper β’ 2311.04235 β’ Published β’ 14 -
Prompt Engineering a Prompt Engineer
Paper β’ 2311.05661 β’ Published β’ 25 -
Orca 2: Teaching Small Language Models How to Reason
Paper β’ 2311.11045 β’ Published β’ 77
-
Efficient LLM Inference on CPUs
Paper β’ 2311.00502 β’ Published β’ 7 -
Exponentially Faster Language Modelling
Paper β’ 2311.10770 β’ Published β’ 119 -
Cached Transformers: Improving Transformers with Differentiable Memory Cache
Paper β’ 2312.12742 β’ Published β’ 14 -
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 260
-
Detecting Pretraining Data from Large Language Models
Paper β’ 2310.16789 β’ Published β’ 11 -
Let's Synthesize Step by Step: Iterative Dataset Synthesis with Large Language Models by Extrapolating Errors from Small Models
Paper β’ 2310.13671 β’ Published β’ 19 -
AutoMix: Automatically Mixing Language Models
Paper β’ 2310.12963 β’ Published β’ 14 -
An Emulator for Fine-Tuning Large Language Models using Small Language Models
Paper β’ 2310.12962 β’ Published β’ 13
-
When can transformers reason with abstract symbols?
Paper β’ 2310.09753 β’ Published β’ 4 -
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper β’ 2310.10638 β’ Published β’ 30 -
Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model
Paper β’ 2310.09520 β’ Published β’ 12 -
Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers
Paper β’ 2309.08532 β’ Published β’ 53
-
Table-GPT: Table-tuned GPT for Diverse Table Tasks
Paper β’ 2310.09263 β’ Published β’ 41 -
A Zero-Shot Language Agent for Computer Control with Structured Reflection
Paper β’ 2310.08740 β’ Published β’ 16 -
The Consensus Game: Language Model Generation via Equilibrium Search
Paper β’ 2310.09139 β’ Published β’ 14 -
PaLI-3 Vision Language Models: Smaller, Faster, Stronger
Paper β’ 2310.09199 β’ Published β’ 29
-
AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model
Paper β’ 2309.16058 β’ Published β’ 56 -
Candle Phi Wasm Demo
π―122Generate text based on prompts
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 260 -
ResNet KAN For Cat And Dog Classification
π