Skip to content

vllm chat

Arguments

--url

url of the running OpenAI-Compatible RESTful API server

Default: http://localhost:8000/v1

--model-name

The model name used in prompt completion, default to the first model in list models API call.

Default: None

--api-key

API key for OpenAI services. If provided, this api key will overwrite the api key obtained through environment variables. It is important to note that this option only applies to the OpenAI-compatible API endpoints and NOT other endpoints that may be present in the server. See the security guide in the vLLM docs for more details.

Default: None

--system-prompt

The system prompt to be added to the chat template, used for models that support system prompts.

Default: None

-q, --quick

Send a single prompt as MESSAGE and print the response, then exit.

Default: None