Software Last Updated: 2026-05-01
llamafile v0.10.*
This repository contains a few llamafiles built from our v0.10.1 release.
These llamafiles are pre-packaged with support for CPU, Metal GPUs on Mac and CUDA on linux. If you have a different GPU or OS, you can download the corresponding library in the Files and versions section and save it in your home directory for llamafile to automatically find it.
For more information about our project, check out our github repo. To learn how to use llamafiles, check our documentation!
NOTE: While the llamafile project is Apache 2.0-licensed, the licenses of models we bundle with it might differ. Use the table above for reference.
- Downloads last month
- 8,520
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support