-
Notifications
You must be signed in to change notification settings - Fork 461
Update july 2025 #1225
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update july 2025 #1225
Conversation
Unit tests passed on Windows CUDA & Linux CUDA. Test application is running fine on:
|
Any possibility to include llama win-arm64-cpu binary in this release? Posting this from windows arm64 device 😥 |
@jihadkhawaja It should be possible, it looks like llama.cpp has support for it. The first step would be to update the build action (see here) to add a windows ARM build. If you're interested in contributing that please feel free to open up an issue and I can give you some tips. |
I would love to, Will open a PR for that, thanks a lot @martindevans |
I think that you would like to update to the latest llama.cpp commit before releasing, they just got the GLM-4.5 models series PR merged yesterday, and the model seems promising. EDIT: OpenAI GPT-OSS just released a few hours ago and support merged into llama.cpp. |
That'll have to wait for the next update, this one is basically ready to go. I'm pretty busy right now, but I'll try to find some time this weekend to do another update, hopefully not much has changed in the last 2 weeks! |
They did a huge PR for gpt-oss 😅 ggml-org/llama.cpp#15091 |
Updated llama.cpp binaries to 11dd5a44eb180e1d69fac24d3852b5222d66fb7f, compiled with this run.
Testing: