[CI][Hardware][Intel GPU] add Intel GPU(XPU) ci pipeline (#5616)
This commit is contained in:
parent
f5dda63eb5
commit
cf90ae0123
@ -42,12 +42,18 @@ steps:
|
|||||||
command: bash .buildkite/run-neuron-test.sh
|
command: bash .buildkite/run-neuron-test.sh
|
||||||
soft_fail: false
|
soft_fail: false
|
||||||
|
|
||||||
- label: "Intel Test"
|
- label: "Intel CPU Test"
|
||||||
depends_on: ~
|
depends_on: ~
|
||||||
agents:
|
agents:
|
||||||
queue: intel
|
queue: intel-cpu
|
||||||
command: bash .buildkite/run-cpu-test.sh
|
command: bash .buildkite/run-cpu-test.sh
|
||||||
|
|
||||||
|
- label: "Intel GPU Test"
|
||||||
|
depends_on: ~
|
||||||
|
agents:
|
||||||
|
queue: intel-gpu
|
||||||
|
command: bash .buildkite/run-xpu-test.sh
|
||||||
|
|
||||||
{% for step in steps %}
|
{% for step in steps %}
|
||||||
{% if step.gpu == "a100" %}
|
{% if step.gpu == "a100" %}
|
||||||
- label: "{{ step.label }}"
|
- label: "{{ step.label }}"
|
||||||
|
|||||||
@ -59,7 +59,7 @@ vLLM is flexible and easy to use with:
|
|||||||
- Tensor parallelism support for distributed inference
|
- Tensor parallelism support for distributed inference
|
||||||
- Streaming outputs
|
- Streaming outputs
|
||||||
- OpenAI-compatible API server
|
- OpenAI-compatible API server
|
||||||
- Support NVIDIA GPUs, AMD GPUs, and Intel CPUs
|
- Support NVIDIA GPUs, AMD GPUs, Intel CPUs and GPUs
|
||||||
- (Experimental) Prefix caching support
|
- (Experimental) Prefix caching support
|
||||||
- (Experimental) Multi-lora support
|
- (Experimental) Multi-lora support
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user