-
Notifications
You must be signed in to change notification settings - Fork 81
Previously downloaded MLX models not available for use #39
Description
Really like the app, great idea for Mac AI integration! I have around 200GB of MLX models already downloaded through mlx_lm.server that I use with LibreChat - however, there appears to be no way to use these with HuggingChat. Please let me know if it would be possible to add this functionality.
According to this post MLX is supported and indeed works but only with the two models available in the HC Settings panel (I've tried the Qwen2.5-3B-Instruct model).
Now supports MLX inference. Press
CMD+Shift+\to switch to local inferenceOriginally posted by @cyrilzakka in #32
This is probably because of the difference where HC and MLX store their models. In HC they're at ~/Documents/huggingface, but with MLX they're at ~/.cache/huggingface. I had hoped to symlink the two but it's not possible as the directory structure is very different. See below for the structure. Thanks!
HuggingChat
~/Documents
│
└── huggingface
│
└── models
│
├── argmaxinc
│ └── whisperkit-coreml
│ └── openai_whisper-large-v3-v20240930
│
├── mlx-community
│ └── Qwen2.5-3B-Instruct-bf16
│
├── openai
│ └── whisper-large-v3
│
└── ...
mlx_lm saved models
~/.cache
│
└── huggingface
│
└── hub
│
├── models--mlx-community--Llama-3.2-3B-Instruct-4bit
│ ├── blobs
│ ├── refs
│ └── snapshots
│
├── models--mlx-community--Llama-3.3-70B-Instruct-4bit
│ ├── blobs
│ ├── refs
│ └── snapshots
│
└── ...