--- license: apache-2.0 library_name: transformers pipeline_tag: text-generation language: - bg - ca - code - cs - cy - da - de - el - en - es - et - eu - fi - fr - ga - gl - hr - hu - it - lt - lv - mt - nl - nn - \no - oc - pl - pt - ro - ru - sh - sk - sl - sr - sv - uk datasets: - oscar-corpus/colossal-oscar-1.0 - HuggingFaceFW/fineweb-edu - joelniklaus/eurlex_resources - joelito/legal-mc4 - projecte-aina/CATalog - UFRGS/brwac - community-datasets/hrwac - danish-foundation-models/danish-gigaword - HiTZ/euscrawl - PleIAs/French-PD-Newspapers - PleIAs/French-PD-Books - AI-team-UoA/greek_legal_code - HiTZ/latxa-corpus-v1.1 - allenai/peS2o - pile-of-law/pile-of-law - PORTULAN/parlamento-pt - hoskinson-center/proof-pile - togethercomputer/RedPajama-Data-1T - bigcode/starcoderdata - bjoernp/tagesschau-2018-2023 - EleutherAI/the_pile_deduplicated base_model: BSC-LT/salamandra-7b-instruct tags: - mlx --- # mlx-community/salamandra-7b-instruct-mlx-bf16 The Model [mlx-community/salamandra-7b-instruct-mlx-bf16](https://huggingface.co/mlx-community/salamandra-7b-instruct-mlx-bf16) was converted to MLX format from [BSC-LT/salamandra-7b-instruct](https://huggingface.co/BSC-LT/salamandra-7b-instruct) using mlx-lm version **0.22.3**. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("bibproj/salamandra-7b-instruct-mlx-fp16") prompt="hello" if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) response = generate(model, tokenizer, prompt=prompt, verbose=True) ```