-
Notifications
You must be signed in to change notification settings - Fork 978
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add plamo-2-1b model #1283
base: main
Are you sure you want to change the base?
Add plamo-2-1b model #1283
Conversation
I found a bug in my code that performs causal-conv1d update, which was related to the difference between the channel-first conv1d in torch and channel-last conv1d in MLX. I fixed it and confirmed that it runs properly as expected, so I changed this PR's state to review-ready 🙏 |
d3f9be0
to
1e75bf1
Compare
I tried running:
But it crashes with missing layers. Did you encounter that?
|
Thank you for checking. That's because the Now I'm discussing with the members of the model development in Preferred Networks if we can change the value of |
This PR adds the latest SLM from Preferred Networks, PLaMo-2-1B.
Currently it can successfully generate correct responses when I use the following code:
However, when I run the following command from my terminal:
the outputs are totally corrupted.
Now I'm investigating the cause of why
mlx_lm.generate
doesn't work for this model.I guess it relates to the cache implementation in this code.