YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Llama-3.2-1B Fine-tuned with DeepSpeed ZeRO-2 Offload
This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct, fine-tuned using DeepSpeed's ZeRO-2 Offload to enable efficient training of billion-parameter models on single GPU setups.
This model was fine-tuned using the arxiv-abstract-dataset on a single Tesla P100 16GB GPU with CPU memory offloading.
For detailed implementation, DeepSpeed configuration, and hardware requirements, please check out the project repository.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support