Skip to content

Conversation

@elasticdotventures
Copy link
Owner

Summary

Adds custom Dockerfile for vLLM to include missing Python dependencies required by the DeepSeek-OCR model.

Changes

  • Dockerfile.vllm: New custom Dockerfile based on vllm/vllm-openai:v0.8.5.post1
    • Installs missing packages: addict, easydict, matplotlib
  • docker-compose.yml: Updated vllm service to build from custom Dockerfile instead of using pre-built image
    • Changed from using pre-built vllm/vllm-openai:v0.8.5.post1 to building blender-mcp-vllm:local

Problem Solved

The vLLM container was failing to start with error:

ImportError: This modeling file requires the following packages that were not found in your environment: addict, matplotlib

Then after adding those dependencies:

ImportError: This modeling file requires the following packages that were not found in your environment: easydict

Testing

  • ✅ Docker image builds successfully with all dependencies
  • ✅ All three required packages (addict, easydict, matplotlib) are installed
  • ⚠️ Container starts but encounters a secondary compatibility issue (see Known Issues)

Known Issues

⚠️ Container still fails with transformers compatibility issue:

ImportError: cannot import name 'LlamaFlashAttention2' from 'transformers.models.llama.modeling_llama'

This appears to be a deeper compatibility issue between DeepSeek-OCR model code and the transformers version in vllm v0.8.5.post1. The model's custom code expects LlamaFlashAttention2 which is not available in the installed transformers version.

Next Steps

  • Investigate transformers version compatibility with DeepSeek-OCR
  • Consider pinning transformers to a specific version
  • Explore alternative vllm versions
  • Test with a compatible model or wait for model/vllm updates

Related

🤖 Generated with Claude Code

elasticdotventures and others added 8 commits October 17, 2025 16:25
Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com>
Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com>
- Add Dockerfile for containerized MCP server
- Add GitHub Actions workflow for multi-arch builds (amd64, arm64)
- Add .dockerignore for optimized builds
- Add README.Docker.md with comprehensive usage guide
- Configure GHCR image registry
- Auto-build on push to main and version tags

🤖 Generated with Claude Code
Co-Authored-By: Claude <[email protected]>
- Create custom Dockerfile based on vllm/vllm-openai:v0.8.5.post1
- Add missing Python packages: addict, easydict, matplotlib
- Update docker-compose.yml to build from custom Dockerfile
- Addresses ImportError for addict and easydict packages

Note: Still investigating LlamaFlashAttention2 compatibility issue

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <[email protected]>
DeepSeek-OCR blocked on vLLM MLA architecture support.
Using llava-hf/llava-1.5-7b-hf as interim solution.

Changes:
- Switch from DeepSeek-OCR to LLaVA 1.5 7B model
- Use vLLM V0 engine for better compatibility
- Remove custom Dockerfile (use official vllm image)
- Document DeepSeek-OCR blockers in FIX-TODO-DeepSeekOCR.md

See FIX-TODO-DeepSeekOCR.md for details on complexity and blockers.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <[email protected]>
@elasticdotventures elasticdotventures merged commit c7709c6 into main Oct 23, 2025
1 check failed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants