**Source:** [gguf_init_from_file](https://github.com/ggerganov/llama.cpp/blob/master/ggml.c) **Examples:** [all-MiniLM-L6-v2.F16.gguf.zip](https://github.com/user-attachments/files/25527189/all-MiniLM-L6-v2.F16.gguf.zip) [Falcon-H1-Tiny-90M-Instruct-Q4_K_M.gguf.zip](https://github.com/user-attachments/files/25527191/Falcon-H1-Tiny-90M-Instruct-Q4_K_M.gguf.zip) [flan-t5-small-q2_k.gguf.zip](https://github.com/user-attachments/files/25527190/flan-t5-small-q2_k.gguf.zip) [gpt-oss-20b-MXFP4.gguf.zip](https://github.com/user-attachments/files/21702620/gpt-oss-20b-MXFP4.gguf.zip) [mixtral-8x7b-v0.1.Q4_K_M.gguf.zip](https://github.com/lutzroeder/netron/files/13802180/mixtral-8x7b-v0.1.Q4_K_M.gguf.zip) [phi-2.Q2_K.gguf.zip](https://github.com/lutzroeder/netron/files/13825087/phi-2.Q2_K.gguf.zip) [qwen2.5-coder-7b-instruct-q4_k_m.gguf.zip](https://github.com/user-attachments/files/25527192/qwen2.5-coder-7b-instruct-q4_k_m.gguf.zip) [vicuna-33b.Q2_K.gguf.zip](https://github.com/lutzroeder/netron/files/13825089/vicuna-33b.Q2_K.gguf.zip)