-
Notifications
You must be signed in to change notification settings - Fork 1.8k
[None] [test] Add MNNVL AlltoAll tests to pre-merge #8601
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
)" This reverts commit 84aa3c9.
|
/bot run --only-multi-gpu-test |
📝 WalkthroughWalkthroughThe changes introduce a new Changes
Sequence DiagramsequenceDiagram
participant DeepSeekV3
participant MoEInterface
participant WideEPMoE
participant FusedMoEWideEP
DeepSeekV3->>MoEInterface: forward(x, ..., alltoall_result_do_sum=False)
MoEInterface->>WideEPMoE: forward_impl(x, ..., alltoall_result_do_sum=False)
WideEPMoE->>FusedMoEWideEP: forward_chunk(x, ..., alltoall_result_do_sum=False)
FusedMoEWideEP->>FusedMoEWideEP: alltoall_combine(result, ..., do_reduce=False)
FusedMoEWideEP-->>FusedMoEWideEP: Skip summation reduction based on flag
FusedMoEWideEP-->>WideEPMoE: Return combined output
WideEPMoE-->>MoEInterface: Return result
MoEInterface-->>DeepSeekV3: Return final output
Estimated code review effort🎯 3 (Moderate) | ⏱️ ~25 minutes The changes introduce a new parameter that is consistently threaded through multiple method signatures in a clear pattern. However, the diff spans five distinct files with varying change types—from signature updates to test infrastructure modifications—requiring separate reasoning for each cohort. The logic itself is straightforward (parameter passing and conditional reduction control), but the breadth of affected components and the need to verify parameter propagation across layers increases the review effort. Pre-merge checks and finishing touches❌ Failed checks (2 warnings)
✅ Passed checks (1 passed)
✨ Finishing touches
🧪 Generate unit tests (beta)
📜 Recent review detailsConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro 📒 Files selected for processing (6)
🧰 Additional context used📓 Path-based instructions (3)**/*.{h,hpp,hh,hxx,cpp,cxx,cc,cu,cuh,py}📄 CodeRabbit inference engine (CODING_GUIDELINES.md)
Files:
**/*.py📄 CodeRabbit inference engine (CODING_GUIDELINES.md)
Files:
**/*.{cpp,cxx,cc,h,hpp,hh,hxx,cu,cuh,py}📄 CodeRabbit inference engine (CODING_GUIDELINES.md)
Files:
🧬 Code graph analysis (3)tensorrt_llm/_torch/models/modeling_deepseekv3.py (1)
tensorrt_llm/_torch/modules/fused_moe/fused_moe_wide_ep.py (2)
tests/unittest/_torch/modules/test_fused_moe.py (1)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (2)
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
|
PR_Github #22234 [ run ] triggered by Bot. Commit: |
|
PR_Github #22234 [ run ] completed with state |
|
/bot run --only-multi-gpu-test |
|
PR_Github #22353 [ run ] triggered by Bot. Commit: |
|
PR_Github #22353 [ run ] completed with state |
|
/bot run --add-multi-gpu-test --disable-fail-fast |
|
PR_Github #22397 [ run ] triggered by Bot. Commit: |
|
PR_Github #22397 [ run ] completed with state |
Reverts #8106
Summary by CodeRabbit
New Features
Tests