[CI][Hardware][Intel GPU] add Intel GPU(XPU) ci pipeline (#5616)
This commit is contained in:
parent
f5dda63eb5
commit
cf90ae0123
@ -42,12 +42,18 @@ steps:
|
||||
command: bash .buildkite/run-neuron-test.sh
|
||||
soft_fail: false
|
||||
|
||||
- label: "Intel Test"
|
||||
- label: "Intel CPU Test"
|
||||
depends_on: ~
|
||||
agents:
|
||||
queue: intel
|
||||
queue: intel-cpu
|
||||
command: bash .buildkite/run-cpu-test.sh
|
||||
|
||||
- label: "Intel GPU Test"
|
||||
depends_on: ~
|
||||
agents:
|
||||
queue: intel-gpu
|
||||
command: bash .buildkite/run-xpu-test.sh
|
||||
|
||||
{% for step in steps %}
|
||||
{% if step.gpu == "a100" %}
|
||||
- label: "{{ step.label }}"
|
||||
|
||||
@ -59,7 +59,7 @@ vLLM is flexible and easy to use with:
|
||||
- Tensor parallelism support for distributed inference
|
||||
- Streaming outputs
|
||||
- OpenAI-compatible API server
|
||||
- Support NVIDIA GPUs, AMD GPUs, and Intel CPUs
|
||||
- Support NVIDIA GPUs, AMD GPUs, Intel CPUs and GPUs
|
||||
- (Experimental) Prefix caching support
|
||||
- (Experimental) Multi-lora support
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user