Skip to content

Fix vllm batch docker image#463

Merged
yunfeng-scale merged 2 commits intomainfrom
yunfeng-vllm-batch-docker
Mar 7, 2024
Merged

Fix vllm batch docker image#463
yunfeng-scale merged 2 commits intomainfrom
yunfeng-vllm-batch-docker

Conversation

@yunfeng-scale
Copy link
Copy Markdown
Contributor

@yunfeng-scale yunfeng-scale commented Mar 7, 2024

Pull Request Summary

vLLM 0.2.5 is not pinned on torch or xformers and torch 2.2 caused things to break https://github.com/vllm-project/vllm/blob/v0.2.5/requirements.txt#L8

Test Plan and Usage Guide

already running in prod

@yunfeng-scale yunfeng-scale requested a review from a team March 7, 2024 00:56
@ian-scale
Copy link
Copy Markdown
Contributor

Pull Request Summary

vLLM 0.2.5 is not pinned on torch or xformers and torch 2.2 caused things to break https://github.com/vllm-project/vllm/blob/v0.2.5/requirements.txt#L8

Test Plan and Usage Guide

already running in prod

naive question: why not just bump vllm version instead of adding pins for xformers, etc? Is this also an issue in newer vllm images like 0.3.2? And if so, do we need to consider this for other packages as well?

@ian-scale ian-scale self-requested a review March 7, 2024 00:58
@yunfeng-scale
Copy link
Copy Markdown
Contributor Author

Pull Request Summary

vLLM 0.2.5 is not pinned on torch or xformers and torch 2.2 caused things to break https://github.com/vllm-project/vllm/blob/v0.2.5/requirements.txt#L8

Test Plan and Usage Guide

already running in prod

naive question: why not just bump vllm version instead of adding pins for xformers, etc? Is this also an issue in newer vllm images like 0.3.2? And if so, do we need to consider this for other packages as well?

yes i'd like to bump vLLM to 0.3.2, but since batch inference code references model engine code there's pydantic conflict. i'll resolve that later. want to get current improvements in first

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants