vllm-stack-0.1.9
The stack deployment of vLLM
What's Changed
- [Feat] Add imagePullSecrets support for router and cache-server deplo… by @HanFa in #762
- [Feat] Add production-ready vLLM Nebius MK8s terraform tutorial by @brokedba in #748
- [Feat] Allow declaring modelSpec resources directly by @danhubern in #729
- [Router] Introduction of /v1/responses endpoint by @sebastiaanvduijn in #691
- [Bugfix][Router] Fix router startup race when using multiple replicas by @bcdonadio in #768
- [Docs] Correct parameter in transcription API tutorial by @davidgao7 in #685
- [Bugfix] Concurrent requests to model are currently limited to 100 due to aiohttp default by @dermodmaster in #767
- Update nixlPeerHost to pd-llama-decode-engine-service by @Xunzhuo in #771
- [Feat] Production Stack Router: Add OpenTelemetry tracing support with W3C context propagation by @HanFa in #772
- [Feat]: Add support for chatTemplates by @mahmoudk1000 in #779
- [Build][Router] Update vllm to v0.13.0 by @shernshiou in #770
- [Feat] Add nodeSelectorTerms for vllmRunTimes by @mahmoudk1000 in #778
- Update calendar link for community meetings by @ruizhang0101 in #783
- Update the documentation of the semantic router deployment to use helm by @szedan-rh in #786
- Fix incorrect import path in batch processor initialization by @keyuchen21 in #784
- [Build][Router] Update aiohttp by @shernshiou in #793
- Update Slack channel link in README by @keyuchen21 in #798
- [Doc] Remove official email link from README by @ruizhang0101 in #805
- feat(oci): Add Oracle Cloud Infrastructure (OKE) deployment support by @fede-kamel in #794
- [Feat] add keda support by @eladmotola in #781
- [CI/Build] Add stable version tags to Docker images during release by @ardecode in #801
- Fix score payload typo and add regression test by @keyuchen21 in #769
- [Feat] Include runner and convert flag by @shernshiou in #803
New Contributors
- @danhubern made their first contribution in #729
- @sebastiaanvduijn made their first contribution in #691
- @bcdonadio made their first contribution in #768
- @dermodmaster made their first contribution in #767
- @szedan-rh made their first contribution in #786
- @fede-kamel made their first contribution in #794
- @eladmotola made their first contribution in #781
- @ardecode made their first contribution in #801
Full Changelog: vllm-stack-0.1.8...vllm-stack-0.1.9