Fastcheck

Public

Fast CI for PRs

Queue Paused

tpu_queueKevin LuuKevin Luu

[Misc] Clean up MiniCPM-V/O code

Cyrus Leung canceled after 5m 10s
bootstrap
:docker: build image
Run Neuron Test
Neuron Test
Documentation Build
Basic Correctness Test
Core Test
Entrypoints Test
Run Async Engine, Inputs, Utils, Worker Test
Async Engine, Inputs, Utils, Worker Test
Run Python-only Installation Test
Python-only Installation Test
Run Chunked Prefill Test
Chunked Prefill Test
Run Distributed Tests (4 GPUs)
Distributed Tests (4 GPUs)
Run Metrics, Tracing Test
Metrics, Tracing Test
Run Regression Test
Regression Test
Run Engine Test
Engine Test
Run V1 Test
V1 Test
Run Examples Test
Examples Test
Run Prefix Caching Test
Prefix Caching Test
Run Samplers Test
Samplers Test
Run LogitsProcessor Test
LogitsProcessor Test
Run Speculative decoding tests
Speculative decoding tests
Run LoRA Test %N
Run PyTorch Fullgraph Smoke Test
PyTorch Fullgraph Smoke Test
Run PyTorch Fullgraph Test
PyTorch Fullgraph Test
Run Kernels Test %N
Run Tensorizer Test
Tensorizer Test
Run Benchmarks
Benchmarks
Run Quantization Test
Quantization Test
Run LM Eval Small Models
LM Eval Small Models
Run OpenAI API correctness
OpenAI API correctness
Run Encoder Decoder tests
Encoder Decoder tests
Run OpenAI-Compatible Tool Use
OpenAI-Compatible Tool Use
Run Basic Models Test
Basic Models Test
Run Language Models Test (Standard)
Language Models Test (Standard)
Run Language Models Test (Extended)
Language Models Test (Extended)
Run Multi-Modal Models Test (Standard)
Multi-Modal Models Test (Standard)
Run Multi-Modal Models Test (Extended) 1
Multi-Modal Models Test (Extended) 1
Run Multi-Modal Models Test (Extended) 2
Multi-Modal Models Test (Extended) 2
Run Custom Models Test
Custom Models Test
Run Distributed Comm Ops Test
Distributed Comm Ops Test
Run Distributed Tests (2 GPUs)
Distributed Tests (2 GPUs)
Run Plugin Tests (2 GPUs)
Plugin Tests (2 GPUs)
Run Multi-step Tests (4 GPUs)
Multi-step Tests (4 GPUs)
Run Pipeline Parallelism Test
Pipeline Parallelism Test
Run LoRA TP Test (Distributed)
LoRA TP Test (Distributed)
Run Weight Loading Multiple GPU Test
Weight Loading Multiple GPU Test
Run 2 Node Tests (4 GPUs in total)
2 Node Tests (4 GPUs in total)
Run A100 tests
Weight Loading Multiple GPU Test - Large Models
Distributed Tests (A100)
LM Eval Large Models
Run TPU V0 Test
TPU V0 Test
Run TPU V1 Test
TPU V1 Test
Run GH200 Test
GH200 Test
bootstrapcurl -sSL https://raw.githubusercontent.com/vllm-project/buildkite-ci/main/scripts/ci_aws_bootstrap.sh | bash
Waited 41s
·
Ran in 19s
:docker: build imageaws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/q9t5s3a7 && #!/bin/bash && if [[ -z $(docker manifest inspect public.ecr.aws/q9t5s3a7/vllm-ci-test-repo:1dfc902ddb2530b5a976262db40faee9957180d6) ]]; then && echo "Image not found, proceeding with build..." && else && echo "Image found" && exit 0 && fi && docker build --build-arg max_jobs=16 --build-arg buildkite_commit=1dfc902ddb2530b5a976262db40faee9957180d6 --build-arg USE_SCCACHE=1 --tag public.ecr.aws/q9t5s3a7/vllm-ci-test-repo:1dfc902ddb2530b5a976262db40faee9957180d6 --target test --progress plain . && docker push public.ecr.aws/q9t5s3a7/vllm-ci-test-repo:1dfc902ddb2530b5a976262db40faee9957180d6
Canceled
Waited 51s
·
Ran in 4m 17s
Neuron Testbash .buildkite/run-neuron-test.sh
Canceled
Documentation Build
Canceled
Basic Correctness Test
Canceled
Core Test
Canceled
Entrypoints Test
Canceled
Async Engine, Inputs, Utils, Worker Test
Canceled
Python-only Installation Test
Canceled
Chunked Prefill Test
Canceled
Distributed Tests (4 GPUs)
Canceled
Metrics, Tracing Test
Canceled
Regression Test
Canceled
Engine Test
Canceled
V1 Test
Canceled
Examples Test
Canceled
Prefix Caching Test
Canceled
Samplers Test
Canceled
LogitsProcessor Test
Canceled
Speculative decoding tests
Canceled
1/4
LoRA Test 1
Canceled
2/4
LoRA Test 2
Canceled
3/4
LoRA Test 3
Canceled
4/4
LoRA Test 4
Canceled
PyTorch Fullgraph Smoke Test
Canceled
PyTorch Fullgraph Test
Canceled
1/4
Kernels Test 1
Canceled
2/4
Kernels Test 2
Canceled
3/4
Kernels Test 3
Canceled
4/4
Kernels Test 4
Canceled
Tensorizer Test
Canceled
Benchmarks
Canceled
Quantization Test
Canceled
LM Eval Small Models
Canceled
OpenAI API correctness
Canceled
Encoder Decoder tests
Canceled
OpenAI-Compatible Tool Use
Canceled
Basic Models Test
Canceled
Language Models Test (Standard)
Canceled
Language Models Test (Extended)
Canceled
Multi-Modal Models Test (Standard)
Canceled
Multi-Modal Models Test (Extended) 1
Canceled
Multi-Modal Models Test (Extended) 2
Canceled
Custom Models Test
Canceled
Distributed Comm Ops Test
Canceled
Distributed Tests (2 GPUs)
Canceled
Plugin Tests (2 GPUs)
Canceled
Multi-step Tests (4 GPUs)
Canceled
Pipeline Parallelism Test
Canceled
LoRA TP Test (Distributed)
Canceled
Weight Loading Multiple GPU Test
Canceled
2 Node Tests (4 GPUs in total)./.buildkite/run-multi-node-test.sh /vllm-workspace/tests 2 2 public.ecr.aws/q9t5s3a7/vllm-ci-test-repo:1dfc902ddb2530b5a976262db40faee9957180d6 "VLLM_TEST_SAME_HOST=0 torchrun --nnodes 2 --nproc-per-node=2 --rdzv_backend=c10d --rdzv_endpoint=192.168.10.10 distributed/test_same_node.py | grep 'Same node test passed' && VLLM_MULTI_NODE=1 pytest -v -s distributed/test_multi_node_assignment.py && VLLM_MULTI_NODE=1 pytest -v -s distributed/test_pipeline_parallel.py" "VLLM_TEST_SAME_HOST=0 torchrun --nnodes 2 --nproc-per-node=2 --rdzv_backend=c10d --rdzv_endpoint=192.168.10.10 distributed/test_same_node.py | grep 'Same node test passed'"
Canceled
Weight Loading Multiple GPU Test - Large Models
Canceled
Distributed Tests (A100)
Canceled
LM Eval Large Models
Canceled
TPU V0 Testif [[ -f ".buildkite/run-tpu-test.sh" ]]; then bash .buildkite/run-tpu-test.sh; fi && yes | docker system prune -a
Canceled
TPU V1 Testif [[ -f ".buildkite/run-tpu-v1-test.sh" ]]; then bash .buildkite/run-tpu-v1-test.sh; fi && yes | docker system prune -a
Canceled
GH200 Testnvidia-smi && bash .buildkite/run-gh200-test.sh
Canceled
AMD: :docker: build imagegrep -i 'from base as test' Dockerfile.rocm && docker build --build-arg max_jobs=16 --tag rocm/vllm-ci:1dfc902ddb2530b5a976262db40faee9957180d6 -f Dockerfile.rocm --target test --progress plain . || docker build --build-arg max_jobs=16 --tag rocm/vllm-ci:1dfc902ddb2530b5a976262db40faee9957180d6 -f Dockerfile.rocm --progress plain . && docker push rocm/vllm-ci:1dfc902ddb2530b5a976262db40faee9957180d6
Canceled
AMD: Core Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s core"
Canceled
AMD: Entrypoints Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; export VLLM_WORKER_MULTIPROC_METHOD=spawn && pytest -v -s entrypoints/llm --ignore=entrypoints/llm/test_lazy_outlines.py --ignore=entrypoints/llm/test_generate.py --ignore=entrypoints/llm/test_generate_multiple_loras.py --ignore=entrypoints/llm/test_guided_generate.py --ignore=entrypoints/llm/test_collective_rpc.py && pytest -v -s entrypoints/llm/test_lazy_outlines.py && pytest -v -s entrypoints/llm/test_generate.py && pytest -v -s entrypoints/llm/test_generate_multiple_loras.py && VLLM_USE_V1=0 pytest -v -s entrypoints/llm/test_guided_generate.py && pytest -v -s entrypoints/openai --ignore=entrypoints/openai/test_oot_registration.py --ignore=entrypoints/openai/correctness/ && pytest -v -s entrypoints/test_chat_utils.py && VLLM_USE_V1=0 pytest -v -s entrypoints/offline_mode"
Canceled
AMD: Regression Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pip install modelscope && pytest -v -s test_regression.py"
Canceled
AMD: Engine Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s engine test_sequence.py test_config.py test_logger.py && pytest -v -s tokenization"
Canceled
AMD: Prefix Caching Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s prefix_caching"
Canceled
AMD: LogitsProcessor Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s test_logits_processor.py && pytest -v -s model_executor/test_guided_processors.py"
Canceled
AMD: LoRA Test %Nbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s lora --shard-id=$BUILDKITE_PARALLEL_JOB --num-shards=$BUILDKITE_PARALLEL_JOB_COUNT --ignore=lora/test_long_context.py --ignore=lora/test_chatglm3_tp.py --ignore=lora/test_llama_tp.py --ignore=lora/test_minicpmv_tp.py --ignore=lora/test_transfomers_model.py"
Canceled
AMD: Kernels Test %Nbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s kernels --shard-id=$BUILDKITE_PARALLEL_JOB --num-shards=$BUILDKITE_PARALLEL_JOB_COUNT"
Canceled
AMD: Tensorizer Testbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; apt-get update && apt-get install -y curl libsodium23 && export VLLM_WORKER_MULTIPROC_METHOD=spawn && pytest -v -s tensorizer_loader"
Canceled
AMD: Benchmarksbash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/.buildkite ; bash run-benchmarks.sh"
Canceled
AMD: OpenAI-Compatible Tool Usebash .buildkite/run-amd-test.sh "(command rocm-smi || true) && export VLLM_LOGGING_LEVEL=DEBUG && export VLLM_ALLOW_DEPRECATED_BEAM_SEARCH=1 && cd /vllm-workspace/tests ; pytest -v -s tool_use"
Canceled
Total Job Run Time: 4m 36s