Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion examples/visual_gen/serve/benchmark_visual_gen.sh
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,10 @@
#
# Requirements:
# pip install git+https://github.com/huggingface/diffusers.git
# pip install av
#
# Optional (for MP4/H.264 video output):
# apt-get install ffmpeg # or: conda install ffmpeg
# Without ffmpeg, videos are saved as AVI/MJPEG using a pure-Python encoder.

set -euo pipefail

Expand Down
1 change: 0 additions & 1 deletion tests/integration/test_lists/test-db/l0_a10.yml
Original file line number Diff line number Diff line change
Expand Up @@ -174,7 +174,6 @@ l0_a10:
- unittest/trt/quantization
- unittest/trt/functional # 37 mins
- llmapi/test_llm_examples.py::test_llmapi_quickstart_atexit
- examples/test_visual_gen.py::test_visual_gen_quickstart
- unittest/api_stability
- unittest/bindings
- unittest/test_model_runner_cpp.py
Expand Down
1 change: 0 additions & 1 deletion tests/integration/test_lists/test-db/l0_gb203.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,6 @@ l0_gb203:
# - examples/test_qwen.py::test_llm_qwen1_5_7b_single_gpu_lora[qwen1.5_7b_chat-Qwen1.5-7B-Chat-750Mb-lora] # https://nvbugs/5234573
# - examples/test_qwen.py::test_llm_qwen_single_gpu_summary[qwen2.5_1.5b_instruct-enable_paged_kv_cache-enable_remove_input_padding-enable_weight_only-enable_fmha_fp32_acc] # https://nvbugs/5234573
- llmapi/test_llm_examples.py::test_llmapi_quickstart
- examples/test_visual_gen.py::test_visual_gen_quickstart
- llmapi/test_llm_examples.py::test_llmapi_example_inference
- llmapi/test_llm_examples.py::test_llmapi_example_inference_async
- llmapi/test_llm_examples.py::test_llmapi_example_inference_async_streaming
Expand Down
1 change: 0 additions & 1 deletion tests/integration/test_lists/test-db/l0_sanity_check.yml
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,6 @@ l0_sanity_check:
linux_distribution_name: ubuntu*
tests:
- llmapi/test_llm_examples.py::test_llmapi_quickstart
- examples/test_visual_gen.py::test_visual_gen_quickstart
- llmapi/test_llm_examples.py::test_llmapi_example_inference
- llmapi/test_llm_examples.py::test_llmapi_example_inference_async
- llmapi/test_llm_examples.py::test_llmapi_example_inference_async_streaming
Expand Down
1 change: 0 additions & 1 deletion tests/integration/test_lists/waives.txt
Original file line number Diff line number Diff line change
Expand Up @@ -376,5 +376,4 @@ unittest/auto_deploy/multigpu/transformations/library/test_tp_sharding.py::test_
perf/test_perf_sanity.py::test_e2e[disagg_upload-gen_only-gb200_gpt-oss-120b-fp4_1k1k_con64_ctx1_tp1_gen1_tp4_eplb0_mtp0_ccb-UCX] SKIP (https://nvbugs/4846166)
perf/test_perf_sanity.py::test_e2e[disagg_upload-gen_only-gb200_deepseek-r1-fp4_1k1k_con3072_ctx1_dep4_gen1_dep4_eplb0_mtp1_ccb-UCX] SKIP (https://nvbugs/5846166)
full:RTXPro6000D/accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_nvfp4_4gpus[moe_backend=CUTLASS-mtp_nextn=0-ep4-fp8kv=True-attention_dp=True-cuda_graph=True-overlap_scheduler=True-torch_compile=True] SKIP (https://nvbugs/5961814)
examples/test_visual_gen.py::test_visual_gen_quickstart SKIP (https://nvbugs/5963896)
full:RTXPro6000D/accuracy/test_llm_api_pytorch.py::TestDeepSeekV3Lite::test_nvfp4_4gpus[moe_backend=CUTLASS-mtp_nextn=2-ep4-fp8kv=True-attention_dp=True-cuda_graph=True-overlap_scheduler=True-torch_compile=False] SKIP (https://nvbugs/5961814)