You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The exposed API depends on the actual model hosted in triton backend. Since there's no embedding model available in trtllm, /v1/embeddings is not possible. For embedding model, maybe you can refer to https://github.com/huggingface/text-embeddings-inference.
The same reason applies to /v1/audio/* since no ASR and TTS models are available now in trtllm.
@npuichigo I am trying to use Triton Inference Server with TensorRT-LLM backend with openweb-ui as frontend, but not all routes are provided, e.g.
/v1/models
etc.Is there any plan to support all openapi v1 routes?
It will be really great if full openai api support is available, since kserve is still under works.
The text was updated successfully, but these errors were encountered: