ExLlamav2 8 bpw quant of https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base
4-bit
Base model