This guide provides detailed instructions for installing vLLM on HPC systems using our installer.
- HPC system with ROCm support
- AMD GPU (MI300X or compatible)
- Git
- Anaconda/Miniconda installation
- Module system with required modules:
- Anaconda (anaconda3/2023.09-0)
- ROCm (amd-mixed/6.2.1)
- CPE (cpe/24.07)
-
Prepare the Environment
# Clone the repository git clone https://github.com/AI-DarwinLabs/vllm-hpc-installer.git cd vllm-hpc-installer # Make the installer executable chmod +x install.sh
-
Configuration (Optional)
- Edit
config/default_config.sh
if you need to customize:- Python version
- ROCm version
- GPU architecture
- Conda environment name
- Edit
-
Run the Installer
./install.sh
-
Verify Installation
# Activate the conda environment conda activate vllm # Verify vLLM installation python -c 'import vllm; print(vllm.__version__)'
The installer sets up:
- Conda environment with Python 3.11
- PyTorch with ROCm support
- vLLM
- Flash Attention
- Additional dependencies
-
Module Load Errors
- Ensure all required modules are available on your system
- Check module versions match your configuration
-
CUDA/ROCm Issues
- Verify ROCm installation
- Check GPU architecture setting matches your hardware
-
Installation Logs
- Check logs in the
logs/
directory for detailed error messages
- Check logs in the
If you encounter issues:
- Check the logs in
logs/
directory - Consult the GitHub Issues page
- Submit a new issue with:
- Error messages
- System information
- Installation logs