YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Llama-3.2-1B Fine-tuned with DeepSpeed ZeRO-2 Offload

This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct, fine-tuned using DeepSpeed's ZeRO-2 Offload to enable efficient training of billion-parameter models on single GPU setups.

This model was fine-tuned using the arxiv-abstract-dataset on a single Tesla P100 16GB GPU with CPU memory offloading.

For detailed implementation, DeepSpeed configuration, and hardware requirements, please check out the project repository.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support