Skip to content

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell) #22029

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

sergey-kozub
Copy link
Contributor

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.

@sergey-kozub sergey-kozub requested a review from reedwm January 29, 2025 17:16
@sergey-kozub sergey-kozub self-assigned this Jan 29, 2025
copybara-service bot pushed a commit that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR #22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

FUTURE_COPYBARA_INTEGRATE_REVIEW=#22029 from openxla:devel/sm120a be59b7a
PiperOrigin-RevId: 721049239
copybara-service bot pushed a commit that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR #22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

FUTURE_COPYBARA_INTEGRATE_REVIEW=#22029 from openxla:devel/sm120a be59b7a
PiperOrigin-RevId: 721049239
copybara-service bot pushed a commit to tensorflow/tensorflow that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR openxla/xla#22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a51721637d880207e7adb69a18c3a92bea by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

FUTURE_COPYBARA_INTEGRATE_REVIEW=openxla/xla#22029 from openxla:devel/sm120a be59b7a51721637d880207e7adb69a18c3a92bea
PiperOrigin-RevId: 721049239
copybara-service bot pushed a commit that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR #22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

FUTURE_COPYBARA_INTEGRATE_REVIEW=#22029 from openxla:devel/sm120a be59b7a
PiperOrigin-RevId: 721049239
copybara-service bot pushed a commit to tensorflow/tensorflow that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR openxla/xla#22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a51721637d880207e7adb69a18c3a92bea by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

FUTURE_COPYBARA_INTEGRATE_REVIEW=openxla/xla#22029 from openxla:devel/sm120a be59b7a51721637d880207e7adb69a18c3a92bea
PiperOrigin-RevId: 721049239
copybara-service bot pushed a commit to tensorflow/tensorflow that referenced this pull request Jan 29, 2025
…(Blackwell)

Imported from GitHub PR openxla/xla#22029

In addition to SM120a, also add SM101a mentioned in the PTX 8.7 spec (https://docs.nvidia.com/cuda/parallel-thread-execution/#release-notes), which is a slight variation of SM100a.

Bumping the max supported PTX version to 8.7, as the LLVM PR (llvm/llvm-project#124155) adding the support is now integrated to OpenXLA.
Copybara import of the project:

--
be59b7a51721637d880207e7adb69a18c3a92bea by Sergey Kozub <skozub@nvidia.com>:

[XLA:GPU] Add support for SM101a and SM120a architectures (Blackwell)

Merging this change closes #22029

PiperOrigin-RevId: 721088886
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants