tritesh's picture
Update ML Intern artifact metadata
bb9c7aa verified
metadata
library_name: transformers
license: apache-2.0
license_link: https://ai.google.dev/gemma/docs/gemma_4_license
pipeline_tag: image-text-to-text
tags:
  - mlx
  - ml-intern
base_model: google/gemma-4-31B-it

raazkumar/gemma-4-31B-it-mlx-2Bit

The Model raazkumar/gemma-4-31B-it-mlx-2Bit was converted to MLX format from google/gemma-4-31B-it using mlx-lm version 0.31.2.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("raazkumar/gemma-4-31B-it-mlx-2Bit")

prompt="hello"

if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        messages, tokenize=False, add_generation_prompt=True
    )

response = generate(model, tokenizer, prompt=prompt, verbose=True)

Generated by ML Intern

This model repository was generated by ML Intern, an agent for machine learning research and development on the Hugging Face Hub.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = 'tritesh/gemma-4-31B-it-mlx-2Bit'
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

For non-causal architectures, replace AutoModelForCausalLM with the appropriate AutoModel class.