Skip to content
This repository was archived by the owner on Jun 5, 2025. It is now read-only.

Implement the /models endpoint for the VLLM provider#188

Merged
jhrozek merged 1 commit intostacklok:mainfrom
jhrozek:vllm_models
Dec 4, 2024
Merged

Implement the /models endpoint for the VLLM provider#188
jhrozek merged 1 commit intostacklok:mainfrom
jhrozek:vllm_models

Conversation

@jhrozek
Copy link
Copy Markdown
Contributor

@jhrozek jhrozek commented Dec 4, 2024

For some reason, continue in JetBrains needs this otherwise loading the
plugin fails with:

Error: HTTP 404 Not Found from http://127.0.0.1:8000/vllm/models

This may mean that you forgot to add '/v1' to the end of your 'apiBase'
in config.json.
    at customFetch (/snapshot/continue/binary/out/index.js:471442:17)
        at process.processTicksAndRejections
        (node:internal/process/task_queues:95:5)
            at async withExponentialBackoff
            (/snapshot/continue/binary/out/index.js:471175:22)
---> continue restarts here
[info] Starting Continue core...
[2024-12-04T08:52:45] [info] Starting Continue core...

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this is fine. We can also query directly the base_url and it should respond

curl -SsX GET "https://inference.codegate.ai/v1/models" \
     -H "Authorization: Bearer $token" 
     
{"object":"list","data":[{"id":"Qwen/Qwen2.5-Coder-14B-Instruct","object":"model","created":1733303915,"owned_by":"vllm","root":"Qwen/Qwen2.5-Coder-14B-Instruct","parent":null,"max_model_len":32768,"permission":[{"id":"modelperm-0aa1923ad501464fbc2f3ee91f953ed3","object":"model_permission","created":1733303915,"allow_create_engine":false,"allow_sampling":true,"allow_logprobs":true,"allow_search_indices":false,"allow_view":true,"allow_fine_tuning":false,"organization":"*","group":null,"is_blocking":false}]}]}

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I like your idea more, I will test it out

Copy link
Copy Markdown
Member

@aponcedeleonch aponcedeleonch left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Approved, left one comment. I don't know if querying the base_url for that endpoint actually makes a difference or not

For some reason, continue in JetBrains needs this otherwise loading the
plugin fails with:
```
Error: HTTP 404 Not Found from http://127.0.0.1:8000/vllm/models

This may mean that you forgot to add '/v1' to the end of your 'apiBase'
in config.json.
    at customFetch (/snapshot/continue/binary/out/index.js:471442:17)
        at process.processTicksAndRejections
        (node:internal/process/task_queues:95:5)
            at async withExponentialBackoff
            (/snapshot/continue/binary/out/index.js:471175:22)
---> continue restarts here
[info] Starting Continue core...
[2024-12-04T08:52:45] [info] Starting Continue core...
```
@jhrozek jhrozek merged commit 3d5575d into stacklok:main Dec 4, 2024
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants