MLX Community org
edited 5 days ago

Converted with this command:
python -m mlx_vlm.convert --hf-path /Volumes/T7/Models/hf-models/gemma-3-12b-it-qat-q4_0-unquantized --mlx-path gemma-3-12b-it-qat-4bit --dtype bfloat16 -q --q-bits 4

Note that the dtype is bfloat16. This model was tested, and confirmed working with mlx_vlm 0.1.25 CLI, and LM Studio. Note the config.json change, the conversion process made this change, and I saw that it was necessary to keep this change to load the model with mlx_vlm

neilmehta24 changed pull request status to open
awni changed pull request status to merged
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment