KEMBAR78
Add test for consistency between meta and CPU devices. by ysiraichi · Pull Request #138515 · pytorch/pytorch · GitHub
Skip to content

Conversation

@ysiraichi
Copy link
Collaborator

@ysiraichi ysiraichi commented Oct 21, 2024

Stack from ghstack (oldest at bottom):

Reference: #138399

This PR introduces an OpInfo test that checks whether running each out= operation
using meta inputs is consistent with using concrete (e.g. CPU) inputs. More specifically,
it tests the case where the output tensors are not of the expected data type. According to
the out= specification, some operations should error.

I have added XFAIL to the set of operations that are currently failing.

[ghstack-poisoned]
@ysiraichi ysiraichi requested a review from mruberry as a code owner October 21, 2024 22:07
@pytorch-bot
Copy link

pytorch-bot bot commented Oct 21, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/138515

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 8716a9b with merge base 4dd04db (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@ezyang
Copy link
Contributor

ezyang commented Oct 24, 2024

It's kind of up in the air whether or not CPU is a good default, but I guess it's easy to test and if you really care you use FakeTensor, so sure why no

@ezyang
Copy link
Contributor

ezyang commented Oct 24, 2024

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Oct 24, 2024
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

The merge job was canceled or timed out. This most often happen if two merge requests were issued for the same PR, or if merge job was waiting for more than 6 hours for tests to finish. In later case, please do not hesitate to reissue the merge command
For more information see pytorch-bot wiki.

@ysiraichi
Copy link
Collaborator Author

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

The merge job was canceled or timed out. This most often happen if two merge requests were issued for the same PR, or if merge job was waiting for more than 6 hours for tests to finish. In later case, please do not hesitate to reissue the merge command
For more information see pytorch-bot wiki.

@ysiraichi
Copy link
Collaborator Author

I've just realized that the CI failure I was getting in #138520 wasn't being run in this PR. I believe this PR also fails on linux-jammy-cpu-py3.9-gcc11-inductor / test (inductor_avx512, 2, 2, linux.12xlarge). I will wait for the ciflow/inductor to run, and see if that's the case.

@ysiraichi ysiraichi marked this pull request as draft October 24, 2024 17:23
@ysiraichi
Copy link
Collaborator Author

Moving this to "draft", in order to debug said CI failure.

@ysiraichi
Copy link
Collaborator Author

Apparently the reason for the CI failure was that the test was being run under inductor. Added a skip for inductor (and dynamo) runs.

@ysiraichi ysiraichi marked this pull request as ready for review October 28, 2024 16:28
@ysiraichi
Copy link
Collaborator Author

The CI failure looks not related, so merging anyway.

@ysiraichi
Copy link
Collaborator Author

@pytorchbot merge -i

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged while ignoring the following 1 checks: Lint / lintrunner-noclang / linux-job

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@huydhn
Copy link
Contributor

huydhn commented Oct 28, 2024

@pytorchbot revert -m 'Sorry for reverting your change, but the test is failing in trunk, maybe a landrace' -c landrace

test_ops.py::TestCommonCPU::test_meta_consistency_out_dtype_mismatch_permute_copy_cpu_float32 GH job link HUD commit link

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a revert job. Check the current status here.
Questions? Feedback? Please reach out to the PyTorch DevX Team

pytorchmergebot added a commit that referenced this pull request Oct 28, 2024
…)"

This reverts commit 006130d.

Reverted #138515 on behalf of https://github.com/huydhn due to Sorry for reverting your change, but the test is failing in trunk, maybe a landrace ([comment](#138515 (comment)))
@pytorchmergebot
Copy link
Collaborator

@ysiraichi your PR has been successfully reverted.

[ghstack-poisoned]
[ghstack-poisoned]
pytorchmergebot pushed a commit that referenced this pull request Oct 30, 2024
Pull Request resolved: #138520
Approved by: https://github.com/ezyang
ghstack dependencies: #138515
rahulsingh-intel pushed a commit to rahulsingh-intel/pytorch that referenced this pull request Nov 5, 2024
Reference: pytorch#138399

This PR introduces an `OpInfo` test that checks whether running each `out=` operation
using meta inputs is consistent with using concrete (e.g. CPU) inputs. More specifically,
it tests the case where the output tensors are not of the expected data type. According to
the `out=` specification, some operations should error.

I have added XFAIL to the set of operations that are currently failing.
Pull Request resolved: pytorch#138515
Approved by: https://github.com/ezyang
rahulsingh-intel pushed a commit to rahulsingh-intel/pytorch that referenced this pull request Nov 5, 2024
@github-actions github-actions bot deleted the gh/ysiraichi/63/head branch November 30, 2024 02:07
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants