sgl-kernel and JIT kernels used by diffusion workloads.
Standard Installation (NVIDIA GPUs)
Method 1: With pip or uv
It is recommended to use uv for a faster installation:Command
Method 2: From source
Command
Method 3: Using Docker
The Docker images are available on Docker Hub at lmsysorg/sglang, built from the Dockerfile. Replace<secret> below with your HuggingFace Hub token.
Command
Platform-Specific: ROCm (AMD GPUs)
For AMD Instinct GPUs (e.g., MI300X), you can use the ROCm-enabled Docker image:Command
Platform-Specific: MUSA (Moore Threads GPUs)
For Moore Threads GPUs (MTGPU) with the MUSA software stack, please follow the instructions below to install from source:Command
Platform-Specific: Intel XPU
For Intel Data Center GPU Max or Arc GPUs, follow the XPU installation guide to set up the base environment, then install diffusion dependencies:Command
Platform-Specific: Ascend NPU
For Ascend NPU, please follow the NPU installation guide. Quick test:Command
Platform-Specific: Apple MPS
For Apple MPS, please follow the instructions below to install from source:Command
