Skip to content

Releases: runpod-workers/worker-vllm

v2.1.0

25 Feb 02:42
dc6f323
Compare
Choose a tag to compare
  • changed max_tokens=100 for runpod specific APIs.
  • updated vllm to v0.7.3

v1.9.0

29 Jan 06:15
9e8d919
Compare
Choose a tag to compare
  • vLLM's V1 engine is ready for testing! This is a rewritten engine designed for performance and architectural simplicity. You can turn it on by setting environment variable VLLM_USE_V1=1.

More updates:
Follow this link: https://github.com/vllm-project/vllm/releases/tag/v0.7.0

v1.8.0

31 Dec 01:47
Compare
Choose a tag to compare
  • Update vllm version from v0.6.4 --> v0.6.6.
  • Support for DeepSeek V3 Model.

v1.7.0

20 Nov 22:57
2df915a
Compare
Choose a tag to compare
  • Update vllm to 0.6.4.

v1.6.0

16 Oct 00:37
ce47c41
Compare
Choose a tag to compare
Merge pull request #125 from runpod-workers/up-0.6.3

update vllm

v1.5.0

01 Oct 18:23
d3ee323
Compare
Choose a tag to compare
  • vllm version update 0.6.1 --> 0.6.2.
  • Supports llama 3.2 Models.

v1.4.0: Merge pull request #109 from runpod-workers/0.5.5-update

17 Sep 06:22
b1554ea
Compare
Choose a tag to compare

v1.3.1

06 Sep 19:42
b1554ea
Compare
Choose a tag to compare

vLLm version: 0.5.5

  • OpenAI Completion Requests Bug fix.

v1.3.0

29 Aug 06:34
286d6ba
Compare
Choose a tag to compare

Version upgrade from vllm v0.5.4 -> v0.5.5

Various improvements and bug fixes.
[Known Issue]: OpenAI Completion Requests error.

v1.2.0

09 Aug 21:59
eb75a3a
Compare
Choose a tag to compare

Version upgrade from vllm v0.5.3 -> v0.5.4

  • Various improvements and bug fixes.
  • [Known Issue]: OpenAI Completion Requests error.