-
Notifications
You must be signed in to change notification settings - Fork 310
Add vLLM x TorchAO integration workflow #2610
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Signed-off-by: Huy Do <[email protected]>
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/2610
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit a6b2aaa with merge base 5fe4ebd ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
@jerryzh168 I need to install |
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
@pytorchbot drci |
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
Signed-off-by: Huy Do <[email protected]>
current torchao main doesn't work with fbgemm gpu genai stable I think, maybe just use stable for everything and we'll make sure next stable torchao release can work with fbgemm gpu genai stable |
Interesting, the tests pass https://github.com/pytorch/ao/actions/runs/16537610493?pr=2610, it means that torchao main is working with vLLM main. Is that what we are looking for here? I guess if we need to test torchao vs fbgemm gpu genai, it makes more sense to have another workflow to test only these twos? |
OK it could be because we haven't landed everything we want in main yet, and some of our new changes would require fbgemm main we do have tests to test torchao + fbgemm gpu genai (nightly) in our CI |
on: | ||
push: | ||
branches: | ||
- main |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: nightly might be enough I think
set -eux | ||
|
||
# vLLM docker image is using CUDA 12.8 and python 3.12 | ||
pip install --pre fbgemm-gpu-genai --index-url https://download.pytorch.org/whl/cu128 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
fbgemm gpu genai nightly would require pytorch nightly I think, and with that we probably need vllm to depend no torch nightly, is that doable with docker? or docker is build with stable pytorch?
This adds a workflow to run vLLM x TorchAO tests with the latest vLLM main. The setup is vLLM main x PyTorch stable x FBGEMM stable x TorchAO PR/main commits.
If there are not enough H100 capacity and you observe queueing, please let me know. We might need to tweak the workflow to run on H100 only when pushing to main.
Testing
https://github.com/pytorch/ao/actions/runs/16535494171/job/46769051375