-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Request] Please generate CodeLlama-7b-Python-hf-q4f16_1-metal.so #16
Comments
The 13B would also be helpful 😄 |
And for |
Not sure if it fully resolves this request, but I at least built and added |
Thanks @sangelone ! How did you do it? (which instructions did you follow and/or did you have to modify any config/source files?) Some of us may be able to replicate it and help build some of the other missing targets. 🙏 |
I just followed the docs using the appropriate model weights from HF on an M2 machine. It took a few hours to get all the dependencies installed and working in a Conda env, but then it was pretty straightforward to do the build itself, maybe half an hour more at that point. |
Thanks @sangelone ! Got the Windows DLL for 13b done. The 34b model's raw weights are 300GB in size 😱 Don't have enough fast NvME to work on it 😞 |
It's missing in the latest update. Can you please add this so that M1/M2 can run the codellama? Thanks
The text was updated successfully, but these errors were encountered: