Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ROCm] Update C++ Extension Tutorial for AMD GPU #3197

Merged
merged 5 commits into from
Mar 19, 2025

Conversation

naromero77amd
Copy link
Contributor

Fixes #2342

Description

This is a minor update to the C++ Extension with a note about AMD GPU support.

Checklist

  • The issue that is being fixed is referred in the description (see above "Fixes #ISSUE_NUMBER")
  • Only one issue is addressed in this pull request
  • Labels from the issue that this PR is fixing are added to this pull request
  • No unnecessary issues are included into this pull request.

Copy link

pytorch-bot bot commented Dec 20, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/tutorials/3197

Note: Links to docs will display an error until the docs builds have been completed.

❌ 1 New Failure, 15 Pending

As of commit c91a69b with merge base 08a9bcd (image):

NEW FAILURE - The following job has failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot
Copy link
Contributor

Hi @naromero77amd!

Thank you for your pull request and welcome to our community.

Action Required

In order to merge any pull request (code, docs, etc.), we require contributors to sign our Contributor License Agreement, and we don't seem to have one on file for you.

Process

In order for us to review and merge your suggested changes, please sign at https://code.facebook.com/cla. If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA.

Once the CLA is signed, our tooling will perform checks and validations. Afterwards, the pull request will be tagged with CLA signed. The tagging process may take up to 1 hour after signing. Please give it that time before contacting us about it.

If you have received this in error or have any questions, please contact us at cla@meta.com. Thanks!

@facebook-github-bot
Copy link
Contributor

Thank you for signing our Contributor License Agreement. We can now accept your code for this (and any) Meta Open Source project. Thanks!

@naromero77amd
Copy link
Contributor Author

@pytorchbot rebase

Copy link

pytorch-bot bot commented Dec 20, 2024

You don't have permissions to rebase this PR since you are a first time contributor. If you think this is a mistake, please contact PyTorch Dev Infra.

@naromero77amd
Copy link
Contributor Author

@svekars @malfet Can you please assign a reviewer? Thanks.

Copy link
Contributor

@malfet malfet left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@naromero77amd it's OK to reference the article at the end of the tutorial, but imo it would be good to incorporate parts of its context into this tutorial, otherwise, in my humble opinion, it's not very helpful for the end user

@svekars
Copy link
Contributor

svekars commented Jan 16, 2025

I agree with @malfet. Also, all references should go to the Conclusion section as to where the user can go after completing this tutorial.

@naromero77amd
Copy link
Contributor Author

@svekars @malfet

My apologies for taking so long to follow-up on this PR.

I should have added more background as to the purpose of this PR.

The point of the PR was to just add a reference to the ROCm blog since this tutorial is already deprecated.

This tutorial is superseded by this newer tutorial:
https://pytorch.org/tutorials/advanced/cpp_custom_ops.html#cpp-custom-ops-tutorial

which in the December 2024 time frame I created a separate PR to add a NOTE about AMD ROCm support.
#3198

The NOTE itself is quite boring because there is nothing extra needed to support a custom HIP kernel in a PyTorch C++ extension. You just write a CUDA kernel and the hipify script takes care of the rest.

I can certainly move the NOTE to the conclusion section if you prefer that. Let me know.

@svekars
Copy link
Contributor

svekars commented Mar 11, 2025

@naromero77amd Please move to the conclusion section - thank you!

@naromero77amd
Copy link
Contributor Author

Done

@naromero77amd
Copy link
Contributor Author

@svekars There is a pending check-- does that require your approval?

@naromero77amd
Copy link
Contributor Author

@svekars ping

@naromero77amd naromero77amd requested a review from malfet March 19, 2025 16:16
@svekars svekars merged commit 6053b2a into pytorch:main Mar 19, 2025
18 of 19 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Custom rocm hip and c++ extensions
4 participants