TECHNOPRAVIN01-Qwen2.5-14B-Valor.F16.Q4_K_M.gguf
Overview (EN)
This repository provides a q4_k_m-quantized GGUF file derived from TECHNOPRAVIN01/Qwen2.5-14B-Valor.
Quick Start (llama.cpp)
./build/bin/llama-cli -m TECHNOPRAVIN01-Qwen2.5-14B-Valor.F16.Q4_K_M.gguf -p "Hello"
Quantization
- Format: GGUF
- Method: q4_k_m
- File:
TECHNOPRAVIN01-Qwen2.5-14B-Valor.F16.Q4_K_M.gguf - Size: ~8.37 GB
Provenance
- Base model:
TECHNOPRAVIN01/Qwen2.5-14B-Valor
Note: This repo contains inference-only quantized weights. Respect the upstream license and usage terms.
Présentation (FR)
Ce dépôt fournit un fichier GGUF q4_k_m issu de TECHNOPRAVIN01/Qwen2.5-14B-Valor.
Démarrage rapide (llama.cpp)
./build/bin/llama-cli -m TECHNOPRAVIN01-Qwen2.5-14B-Valor.F16.Q4_K_M.gguf -p "Bonjour"
Quantification
- Format : GGUF
- Méthode : q4_k_m
- Fichier :
TECHNOPRAVIN01-Qwen2.5-14B-Valor.F16.Q4_K_M.gguf - Taille : ~8.37 Go
Provenance
- Modèle amont :
TECHNOPRAVIN01/Qwen2.5-14B-Valor
Remarque : Ce dépôt contient uniquement des poids quantifiés pour l'inférence. Merci de respecter la licence et les conditions d'utilisation du modèle d'origine.
- Downloads last month
- 11
Hardware compatibility
Log In
to view the estimation
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for ramgpt/TECHNOPRAVIN01-Qwen2.5-14B-Valor-GGUF-Q4_K_M
Base model
Qwen/Qwen2.5-14B
Finetuned
Qwen/Qwen2.5-14B-Instruct
Adapter
TECHNOPRAVIN01/Qwen2.5-14B-Valor