Implement TTA Batch Processing to Improve Inference Speed#2153
Implement TTA Batch Processing to Improve Inference Speed#2153PengchengShi1220 wants to merge 4 commits intoMIC-DKFZ:masterfrom
Conversation
|
Hi,
Best, |
|
Hi Fabian, Thanks for your feedback! Based on your suggestions, I have now made the "use_batch_tta" an optional parameter in the nnUNetPredictor class, which can be controlled via the parser argument "disable_batch_tta". This allows users to opt-in or out of batch TTA based on their VRAM capacity and priorities. Please let me know if further adjustments are required. Best, |
|
Frankly given the limited speed improvement of allowing batched prediction I would prefer to keep things simple. Apologies for dragging this out so long. I appreciate the work you did! This was an important thing to try, even if the improvements were much smaller than one might anticipate. |
Summary:
Proposing the integration of Test Time Augmentation (TTA) with batch processing in nnUNet to enhance inference efficiency, particularly evident in larger 3D datasets. Demonstrated improvements of 5%-8% in speed with validated results on the AMOS2022 dataset.
Implementation Details:
Results:
Mirror Axes (0, 1, 2):
Mirror Axes (0, 1):
Mirror Axis (0):
VRAM Usage:
Recommendations:
The TTA batch processing approach has been thoroughly tested on the AMOS2022 dataset, showing consistent results with the original setup.