-
Notifications
You must be signed in to change notification settings - Fork 25.7k
[ROCM] fix bug in #60313 #61073
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ROCM] fix bug in #60313 #61073
Conversation
💊 CI failures summary and remediationsAs of commit 3b2a6a9 (more details on the Dr. CI page and at hud.pytorch.org/pr/61073):
🕵️ 1 new failure recognized by patternsThe following CI failures do not appear to be due to upstream breakages:
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This affects ROCm only, so I trust you know what you are doing.
But frankly, it feels like rather than documenting limitations of ROCfft support, PR is trying to avoid testing those corner cases
| # rocFFT requires/assumes that the input to hipfftExecC2R or hipfftExecZ2D | ||
| # is of the form that is a valid output from a real to complex transform | ||
| # (i.e. it cannot be a set of random numbers) | ||
| # So for ROCm, call np.fft.rfftn and use its output as the input |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This comment seems out of date, isn't it?
|
@malfet has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
This PR fixes a bug in #60313. Where the tensors generated by _generate_valid_rocfft_input are on the cpu instead of the gpu. This was due to using numpy to generate tensors and converting it to pytorch using torch.from_numpy. This leads to the generated tensors staying on the cpu. We now generate the tensors using pytorch itself which carries over the device type of the input tensors to the generated tensor.