Skip to content

Add support for Gemma 3? #237

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
alexgusevski opened this issue Mar 12, 2025 · 11 comments
Open

Add support for Gemma 3? #237

alexgusevski opened this issue Mar 12, 2025 · 11 comments

Comments

@alexgusevski
Copy link

As far as I can tell all Gemma 3 models are multimodal except maybe 1B ones? Not sure but it says all of them on their HF.

@alexgusevski
Copy link
Author

Okay saw now that you already support it! Nice!

The 1B one seems to not be able to be converted through this mlx_vlm library, but it is not able to be converted with mlx_lm library either right now.

@Blaizzy
Copy link
Owner

Blaizzy commented Mar 12, 2025

I already made a PR to MLX-LM to support 1B.

ml-explore/mlx-examples#1336

@vlbosch
Copy link

vlbosch commented Mar 12, 2025

I already made a PR to MLX-LM to support 1B.

ml-explore/mlx-examples#1336

You're very quick! Great work!

Would that PR also work for text only use of larger models, like 27B?

@Blaizzy
Copy link
Owner

Blaizzy commented Mar 12, 2025

Yes, it will :)

@Blaizzy
Copy link
Owner

Blaizzy commented Mar 12, 2025

Thanks!

@vlbosch
Copy link

vlbosch commented Mar 12, 2025

I just tried the converted model in "--chat"-mode, but as response to a text-only query I get only "< pad >" as output

@psm-2
Copy link

psm-2 commented Mar 13, 2025

@Blaizzy The 12b model does not seem to work for me, it just outputs a lot of nonsense... But, when you do add the compatibility, would it be possible to add fine-tuning?

@alexgusevski
Copy link
Author

alexgusevski commented Mar 13, 2025

@psm-2 Same, this is what I get when running the 4B model, got it from mlx community, havent tried other versions of it.

Image

Oh and seems @vlbosch also gets this

@BlackLuny
Copy link

@psm-2 Same, this is what I get when running the 4B model, got it from mlx community, havent tried other versions of it.

Image Oh and seems [@vlbosch](https://github.com/vlbosch) also gets this

same on 27B 8bit

@alexgusevski
Copy link
Author

More people having same issue as above: lmstudio-ai/lmstudio-bug-tracker#513

@FL33TW00D
Copy link
Contributor

This should be fixed!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants