ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x-qx64-hi-mlx

Gee, people like this model. I'll make a larger quant.

Keep in mind, this is the brainstomed old Ernie, not the new Ernie.

Would be nice if there would be MLX support for the new Ernie, but there isn't.

So, this is it.

-G

This model ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x-qx64-hi-mlx was converted to MLX format from DavidAU/ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x using mlx-lm version 0.27.1.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x-qx64-hi-mlx")

prompt = "hello"

if tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        messages, add_generation_prompt=True
    )

response = generate(model, tokenizer, prompt=prompt, verbose=True)
Downloads last month
129
Safetensors
Model size
37B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nightmedia/ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x-qx64-hi-mlx

Quantized
(4)
this model
Quantizations
1 model

Collections including nightmedia/ERNIE-4.5-36B-A3B-Thinking-Brainstorm20x-qx64-hi-mlx