add changes for lora adapter support and /v1/models endpoint#121
add changes for lora adapter support and /v1/models endpoint#121pandyamarut merged 2 commits intorunpod-workers:mainfrom
Conversation
|
Thanks for the PR. @sven-knoblauch . Can you please also add how did you test the PR? |
|
I made a docker container with the given Dockerfile (on dockerhub: svenknob/runpod-vllm-worker) and tested it on runpod serverless. Worked with a custom trained lora adapter (added in the runpod GUI as ENV variable: LORA_MODULES) with an awq mistral model. The lora adapter is also visible in the v1/models endpoint. |
|
Hi, is there a documentation for this env usage? in the markdown perhaps? |
|
Is a docker image publicly available that contains this PR? |
|
Added a pull request for changing the readme #130. For now you can use my docker image svenknob/runpod-vllm-worker, till it has been published. |
small changes to add the lora modules ENV variable (supporting 1 lora adapter) as solution for #119
with the format:
{"name": "xxx", "path": "xxx/xxxxx", "base_model_name": "xxx/xxxx"}also change of the v1/models endpoint to return all models