https://huggingface.co/inclusionAI/Ling-flash-2.0

#1393
by MultEase - opened

Please help quantize this model to GGUF:
https://huggingface.co/inclusionAI/Ling-flash-2.0
Much appreciated—really grateful for any help!

@MultEase llama.cpp support for BailingMoeV2ForCausalLM is currently being worked on. We will have to wait for it to be merged to llama.cpp mainline before we can fulfill your request. Please follow https://github.com/ggml-org/llama.cpp/issues/15968 and https://github.com/ggml-org/llama.cpp/pull/16063. Please let us know once it is merged in case we forget about your request.

Sign up or log in to comment