Skip to content

Conversation

@groxaxo
Copy link

@groxaxo groxaxo commented Sep 2, 2025

This PR resolves critical CUDA compatibility issues that were preventing the application from running with large models on GPU. The main problem was the use of CUDA 12.9.1, which is not officially supported by PyTorch.

Summary of Changes

  • Dockerfile Updates: Changed base images to CUDA 12.1.1 (officially supported by PyTorch) and updated PyTorch installations to use CUDA 12.1 compatible packages
  • Documentation Improvements: Added comprehensive CUDA setup guide and updated README with driver requirements and usage instructions
  • Dependency Management: Added minimum version requirements for torch (>=2.1.0) and torchaudio (>=2.1.0) in pyproject.toml
  • Testing Infrastructure: Added test files to verify CUDA compatibility in containerized environments

Why CUDA 12.1?

  • Officially supported by current PyTorch versions
  • Compatible with widely available NVIDIA drivers (515+)
  • Provides good performance for Whisper models
  • Balances compatibility with performance

Testing

To verify the fixes:

# Test CUDA compatibility
docker build -t cuda-test -f Dockerfile.test .
docker run --gpus all cuda-test

# Test with actual GPU container
docker build -t whisperlivekit-gpu -f Dockerfile.gpu .
docker run --gpus all -p 8000:8000 whisperlivekit-gpu --model large-v3

These changes ensure WhisperLiveKit works reliably with GPU acceleration, including support for large models.

Related Documentation

Roo and others added 2 commits September 2, 2025 10:00
This PR addresses CUDA compatibility issues by:

- Updating base images from CUDA 12.9.1 to CUDA 12.1.1 which is officially supported by PyTorch

- Updating PyTorch installation to use CUDA 12.1 compatible packages

- Adding comprehensive GPU setup documentation

- Adding CUDA compatibility test scripts

- Updating README with correct Docker instructions and requirements

- Adding minimum version requirements for torch and torchaudio
@ronaldvh95
Copy link

Thnx

@Vladimir-Anfimov
Copy link

Hi, when will this be deployed?

@QuentinFuxa
Copy link
Owner

Hi, when will this be deployed?

Hi, you probably used some AI coding for the fix since the statement that torch only supports cuda up to 12.1.1 is false:

image

So that is not True anymore, no need for the downgrade

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants