Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

vito95311
/
Qwen3-Omni-30B-A3B-Thinking-GGUF-INT8FP16

Text Generation
GGUF
PyTorch
Transformers
Chinese
English
multilingual
llama.cpp
multimodal
quantized
ollama
llama-cpp
qwen
omni
int8
fp16
Eval Results
Model card Files Files and versions
xet
Community
7
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Path hardcoded in .modelfile

๐Ÿ‘ 1
#7 opened about 2 months ago by
serene-ai

How much vram?

#6 opened about 2 months ago by
yiki12

Support for tools / function calling?

5
#4 opened about 2 months ago by
TeddyHuang

Error while loading model

โž• 1
2
#3 opened about 2 months ago by
LimingShen

ๆœ‰ๅ…ถไป–้‡ๅŒ–็‰ˆๆœฌๅ—Ž?

#2 opened about 2 months ago by
Gavin-chen

why the int8 and fp16 model size both are 31GB?

๐Ÿง  1
4
#1 opened about 2 months ago by
snomile
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs