Skip to content

Releases: runpod-workers/worker-vllm

v2.2.2

03 Apr 15:21
4395d0c
Compare
Choose a tag to compare
  • minor fixe. (hubs.json)

v2.2.1

03 Apr 01:13
5864fa6
Compare
Choose a tag to compare
  • add to the template hub. (/.runpod)

v2.2.0

26 Mar 22:57
da3e5f5
Compare
Choose a tag to compare

v2.1.0

25 Feb 02:42
dc6f323
Compare
Choose a tag to compare
  • changed max_tokens=100 for runpod specific APIs.
  • updated vllm to v0.7.3

v1.9.0

29 Jan 06:15
9e8d919
Compare
Choose a tag to compare
  • vLLM's V1 engine is ready for testing! This is a rewritten engine designed for performance and architectural simplicity. You can turn it on by setting environment variable VLLM_USE_V1=1.

More updates:
Follow this link: https://github.com/vllm-project/vllm/releases/tag/v0.7.0

v1.8.0

31 Dec 01:47
Compare
Choose a tag to compare
  • Update vllm version from v0.6.4 --> v0.6.6.
  • Support for DeepSeek V3 Model.

v1.7.0

20 Nov 22:57
2df915a
Compare
Choose a tag to compare
  • Update vllm to 0.6.4.

v1.6.0

16 Oct 00:37
ce47c41
Compare
Choose a tag to compare
Merge pull request #125 from runpod-workers/up-0.6.3

update vllm

v1.5.0

01 Oct 18:23
d3ee323
Compare
Choose a tag to compare
  • vllm version update 0.6.1 --> 0.6.2.
  • Supports llama 3.2 Models.

v1.4.0: Merge pull request #109 from runpod-workers/0.5.5-update

17 Sep 06:22
b1554ea
Compare
Choose a tag to compare