Skip to content

Conversation

@R2-Y
Copy link
Contributor

@R2-Y R2-Y commented Dec 5, 2025

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

Add offline UT for qwen3-omni

Test Plan

pytest -sv tests/multi_stages/qwen3_omni/test_qwen3_omni.py

Test Result

Pass

Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

Signed-off-by: Rein Yang <[email protected]>
@R2-Y R2-Y requested a review from hsliuustc0106 as a code owner December 5, 2025 10:15
@chatgpt-codex-connector
Copy link

The account who enabled Codex for this repo no longer has access to Codex. Please contact the admins of this repo to enable Codex again.

@R2-Y
Copy link
Contributor Author

R2-Y commented Dec 5, 2025

@ywang96 Hi Roger, I’m running pytest on an H200 with a tweaked qwen3‑omni config. I’m worried it may not work on A100 or other GPUs. Could you let me know which device types are available on Buildkite? So I can update the CI yaml to test again. Thanks!

- pytest -s -v tests/multi_stages/qwen3_omni/
agents:
queue: "gpu_1_queue" # g6.4xlarge instance on AWS, has 1 L4 GPU
queue: "gpu_4_queue" # g6.4xlarge instance on AWS, has 1 L4 GPU
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

gpu_4_queue has 4 GPU

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How many GB of VRAM does each GPU have, and where can I find these resource details?

Copy link
Member

@Isotr0py Isotr0py Dec 6, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

L4 has 24GB VRAM, but there is usually only about 22GB available on CI machine actually.

@gcanlin
Copy link
Contributor

gcanlin commented Dec 5, 2025

Maybe we could reuse the conftest.py in #168.



@pytest.mark.parametrize("model", models)
def test_image_to_audio(omni_runner: type[OmniRunner], model: str) -> None:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think that we could keep only one test that outputs audio to save resources.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should I add at least one audio test and one video/image test? The visual has more handling than audio. And qwen3-omni has some bug now don't support mixed modality.

timeout_in_minutes: 20
depends_on: image-build
commands:
- pytest -s -v tests/multi_stages/qwen3_omni/
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
- pytest -s -v tests/multi_stages/qwen3_omni/
- export VLLM_WORKER_MULTIPROC_METHOD=spawn
- pytest -s -v tests/multi_stages/qwen3_omni/

@R2-Y
Copy link
Contributor Author

R2-Y commented Dec 6, 2025

Maybe we could reuse the conftest.py in #168.

Sure, I refer your PR, can add one single modality test to conftest.py later after your's merged.

@ywang96
Copy link
Member

ywang96 commented Dec 6, 2025

Maybe it's a good idea to use 2xH100 for qwen3-omni e2e test for now unless we run into resource issues. Feel free to take a look at #221

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants