Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ROCm] Enable gemm fusion autotuner. #103

Open
wants to merge 1 commit into
base: rocm-jaxlib-v0.4.31-qa
Choose a base branch
from

Conversation

zoranjovanovic-ns
Copy link

No description provided.

@hsharsha
Copy link

hsharsha commented Feb 6, 2025

@zoranjovanovic-ns Is there a PR against 0.4.35 to fix common test failures between 0.4.31 and 0.4.35?

@zoranjovanovic-ns
Copy link
Author

@hsharsha Do you have log file or list of failing test cases?

@i-chaochen
Copy link

I still can see there are 7 failed unit tests @hsharsha

//xla/service/gpu:dot_dimension_sorter_test_gpu_amd_any
//xla/service/gpu:triton_fusion_numerics_verifier_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_fusion_emitter_parametrized_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_support_legacy_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_support_test
//xla/service/gpu/tests:gpu_triton_custom_call_test_gpu_amd_any
//xla/tests:dot_operation_test_gpu_amd_any

Executed 276 out of 276 tests: 269 tests pass and 7 fail locally

wondering is it expectable?

@zoranjovanovic-ns
Copy link
Author

Will go through the list and investigate.

@hsharsha
Copy link

hsharsha commented Feb 7, 2025

@zoranjovanovic-ns
Copy link
Author

@hsharsha Should I create PR for https://github.com/ROCm/xla/tree/r0.4.31-qa-test-fix then?

@hsharsha
Copy link

hsharsha commented Feb 7, 2025

@hsharsha Should I create PR for https://github.com/ROCm/xla/tree/r0.4.31-qa-test-fix then?

I need to merge it to 0.4.31-qa. i cleaned up few cuda specific tests in there. With that these tests are failing with triton autotuner

//xla/service/gpu:triton_fusion_numerics_verifier_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_fusion_emitter_parametrized_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_support_legacy_test_gpu_amd_any
//xla/service/gpu/fusions/triton:triton_support_test

//xla/tools/multihost_hlo_runner:functional_hlo_runner_test_gpu_amd_any

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants