now that we have vllm support, it would be really nice if we had support for the rest of the functionality that the openAI api and vllm support enabled in docker.
https://docs.vllm.ai/en/stable/serving/openai_compatible_server/#supported-apis
Especially embeddings, transcribe, translate, and realtime api. (and image generation)
These are super useful while deving locally to make sure code works, run unit tests etc. without having to hit expensive cloud services you use in production.
(especially transcribe, and realtime for us)
now that we have vllm support, it would be really nice if we had support for the rest of the functionality that the openAI api and vllm support enabled in docker.
https://docs.vllm.ai/en/stable/serving/openai_compatible_server/#supported-apis
Especially embeddings, transcribe, translate, and realtime api. (and image generation)
These are super useful while deving locally to make sure code works, run unit tests etc. without having to hit expensive cloud services you use in production.
(especially transcribe, and realtime for us)