@@ -301,6 +301,7 @@ steps:
301
301
parallelism : 4
302
302
303
303
- label : PyTorch Compilation Unit Tests
304
+ torch_nightly : true
304
305
source_file_dependencies :
305
306
- vllm/
306
307
- tests/compile
@@ -310,6 +311,7 @@ steps:
310
311
- pytest -v -s compile/test_sequence_parallelism.py
311
312
312
313
- label : PyTorch Fullgraph Smoke Test # 9min
314
+ torch_nightly : true
313
315
source_file_dependencies :
314
316
- vllm/
315
317
- tests/compile
@@ -320,6 +322,7 @@ steps:
320
322
- pytest -v -s compile/piecewise/test_toy_llama.py
321
323
322
324
- label : PyTorch Fullgraph Test # 18min
325
+ torch_nightly : true
323
326
source_file_dependencies :
324
327
- vllm/
325
328
- tests/compile
@@ -401,12 +404,13 @@ steps:
401
404
commands :
402
405
- pytest -v -s benchmarks/
403
406
404
- - label : Quantization Test # 33min
407
+ - label : Quantization Test
405
408
source_file_dependencies :
406
409
- csrc/
407
410
- vllm/model_executor/layers/quantization
408
411
- tests/quantization
409
- command : VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
412
+ commands :
413
+ - VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
410
414
411
415
- label : LM Eval Small Models # 53min
412
416
working_dir : " /vllm-workspace/.buildkite/lm-eval-harness"
@@ -447,88 +451,85 @@ steps:
447
451
# #### models test #####
448
452
449
453
- label : Basic Models Test # 24min
454
+ torch_nightly : true
450
455
source_file_dependencies :
451
456
- vllm/
452
457
- tests/models
453
458
commands :
454
459
- pytest -v -s models/test_transformers.py
455
460
- pytest -v -s models/test_registry.py
461
+ - pytest -v -s models/test_utils.py
462
+ - pytest -v -s models/test_vision.py
456
463
# V1 Test: https://github.com/vllm-project/vllm/issues/14531
457
464
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2'
458
465
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'llama4'
459
466
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'plamo2'
460
467
461
- - label : Language Models Test (Standard) # 32min
468
+ - label : Language Models Test (Standard)
462
469
# mirror_hardwares: [amd]
463
470
source_file_dependencies :
464
471
- vllm/
465
- - tests/models/decoder_only/language
466
- - tests/models/embedding/language
467
- - tests/models/encoder_decoder/language
472
+ - tests/models/language
468
473
commands :
469
474
# Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
470
475
-
pip install 'git+https://github.com/Dao-AILab/[email protected] '
471
- - pytest -v -s models/decoder_only/language -m 'core_model or quant_model'
472
- - pytest -v -s models/embedding/language -m core_model
476
+ - pytest -v -s models/language -m core_model
473
477
474
- - label : Language Models Test (Extended) # 1h10min
478
+ - label : Language Models Test (Extended)
475
479
optional : true
476
480
source_file_dependencies :
477
481
- vllm/
478
- - tests/models/decoder_only/language
479
- - tests/models/embedding/language
480
- - tests/models/encoder_decoder/language
482
+ - tests/models/language
481
483
commands :
482
484
# Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
483
- - pip install causal-conv1d
484
- - pytest -v -s models/decoder_only/language -m 'not core_model and not quant_model'
485
- - pytest -v -s models/embedding/language -m 'not core_model'
485
+ -
pip install 'git+https://github.com/Dao-AILab/[email protected] '
486
+ - pytest -v -s models/language -m 'not core_model'
486
487
487
- - label : Multi-Modal Models Test (Standard) # 40min
488
+ - label : Multi-Modal Models Test (Standard)
488
489
# mirror_hardwares: [amd]
489
490
source_file_dependencies :
490
491
- vllm/
491
- - tests/models/decoder_only/audio_language
492
- - tests/models/decoder_only/vision_language
493
- - tests/models/embedding/vision_language
494
- - tests/models/encoder_decoder/audio_language
495
- - tests/models/encoder_decoder/vision_language
492
+ - tests/models/multimodal
496
493
commands :
497
494
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
498
- - pytest -v -s models/multimodal
499
- - pytest -v -s models/decoder_only/audio_language -m 'core_model or quant_model'
500
- - pytest -v -s models/decoder_only/vision_language -m 'core_model or quant_model'
501
- - pytest -v -s models/embedding/vision_language -m core_model
502
- - pytest -v -s models/encoder_decoder/audio_language -m core_model
503
- - pytest -v -s models/encoder_decoder/language -m core_model
504
- - pytest -v -s models/encoder_decoder/vision_language -m core_model
505
- - pytest -v -s models/decoder_only/vision_language/test_interleaved.py
506
-
507
- - label : Multi-Modal Models Test (Extended) 1 # 48m
495
+ - pytest -v -s models/multimodal/processing
496
+ - pytest -v -s --ignore models/multimodal/generation/test_whisper.py models/multimodal -m core_model
497
+ - cd .. && pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work
498
+
499
+ - label : Multi-Modal Models Test (Extended) 1
508
500
optional : true
509
501
source_file_dependencies :
510
502
- vllm/
511
- - tests/models/decoder_only/audio_language
512
- - tests/models/decoder_only/vision_language
513
- - tests/models/embedding/vision_language
514
- - tests/models/encoder_decoder/vision_language
503
+ - tests/models/multimodal
515
504
commands :
516
505
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
517
- - pytest -v -s models/decoder_only/audio_language -m 'not core_model and not quant_model'
518
- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=0) and not core_model and not quant_model'
519
- - pytest -v -s --ignore models/decoder_only/vision_language/test_models.py models/decoder_only/vision_language -m 'not core_model and not quant_model'
520
- - pytest -v -s models/embedding/vision_language -m 'not core_model'
521
- - pytest -v -s models/encoder_decoder/language -m 'not core_model'
522
- - pytest -v -s models/encoder_decoder/vision_language -m 'not core_model'
523
-
524
- - label : Multi-Modal Models Test (Extended) 2 # 38m
506
+ - pytest -v -s --ignore models/multimodal/generation/test_common.py --ignore models/multimodal/processing models/multimodal -m 'not core_model'
507
+
508
+ - label : Multi-Modal Models Test (Extended) 2
525
509
optional : true
526
510
source_file_dependencies :
527
511
- vllm/
528
- - tests/models/decoder_only/vision_language
512
+ - tests/models/multimodal
529
513
commands :
530
514
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
531
- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=1) and not core_model and not quant_model'
515
+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=0) and not core_model'
516
+
517
+ - label : Multi-Modal Models Test (Extended) 3
518
+ optional : true
519
+ source_file_dependencies :
520
+ - vllm/
521
+ - tests/models/multimodal
522
+ commands :
523
+ - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
524
+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=1) and not core_model'
525
+
526
+ - label : Quantized Models Test
527
+ # mirror_hardwares: [amd]
528
+ source_file_dependencies :
529
+ - vllm/model_executor/layers/quantization
530
+ - tests/models/quantization
531
+ commands :
532
+ - pytest -v -s models/quantization
532
533
533
534
# This test is used only in PR development phase to test individual models and should never run on main
534
535
- label : Custom Models Test
@@ -598,9 +599,8 @@ steps:
598
599
- TARGET_TEST_SUITE=L4 pytest basic_correctness/ -v -s -m 'distributed(num_gpus=2)'
599
600
# Avoid importing model tests that cause CUDA reinitialization error
600
601
- pytest models/test_transformers.py -v -s -m 'distributed(num_gpus=2)'
601
- - pytest models/encoder_decoder/language/test_bart.py -v -s -m 'distributed(num_gpus=2)'
602
- - pytest models/encoder_decoder/vision_language/test_broadcast.py -v -s -m 'distributed(num_gpus=2)'
603
- - pytest models/decoder_only/vision_language/test_models.py -v -s -m 'distributed(num_gpus=2)'
602
+ - pytest models/language -v -s -m 'distributed(num_gpus=2)'
603
+ - pytest models/multimodal -v -s -m 'distributed(num_gpus=2)'
604
604
# test sequence parallel
605
605
- pytest -v -s distributed/test_sequence_parallel.py
606
606
# this test fails consistently.
0 commit comments