--- license: other license_name: seallm license_link: https://huggingface.co/SeaLLMs/SeaLLM-13B-Chat/blob/main/LICENSE language: - en - zh - hi - es - fr - ar - bn - ru - pt - id - ur - de - ja - sw - ta - tr - ko - vi - jv - it - ha - th - fa - tl - my tags: - multilingual - babel - mlx - mlx-my-repo base_model: Tower-Babel/Babel-9B --- # KYUNGYONG/Babel-9B-4bit The Model [KYUNGYONG/Babel-9B-4bit](https://huggingface.co/KYUNGYONG/Babel-9B-4bit) was converted to MLX format from [Tower-Babel/Babel-9B](https://huggingface.co/Tower-Babel/Babel-9B) using mlx-lm version **0.21.5**. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("KYUNGYONG/Babel-9B-4bit") prompt="hello" if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) response = generate(model, tokenizer, prompt=prompt, verbose=True) ```