KEMBAR78
[Inductor XPU] Fix wrong device check before skip concat linear. by etaf · Pull Request #140916 · pytorch/pytorch · GitHub
Skip to content

Conversation

[ghstack-poisoned]
@pytorch-bot
Copy link

pytorch-bot bot commented Nov 18, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/140916

Note: Links to docs will display an error until the docs builds have been completed.

❌ 2 New Failures, 2 Unrelated Failures

As of commit d09f20d with merge base 3614d13 (image):

NEW FAILURES - The following jobs have failed:

FLAKY - The following job failed but was likely due to flakiness present on trunk:

UNSTABLE - The following job failed but was likely due to flakiness present on trunk and has been marked as unstable:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@etaf etaf changed the title [Inductor XPU] Fix the missed XPU device check before skip enabling concat linear. [Inductor XPU] Fix wrong device assign before skip enabling concat linear. Nov 18, 2024
@etaf etaf changed the title [Inductor XPU] Fix wrong device assign before skip enabling concat linear. [Inductor XPU] Fix wrong device assign before skip concat linear. Nov 18, 2024
[ghstack-poisoned]
etaf added a commit that referenced this pull request Nov 18, 2024
…oncat linear.

ghstack-source-id: e953960
Pull Request resolved: #140916
@etaf etaf added ciflow/xpu Run XPU CI tasks topic: not user facing topic category labels Nov 18, 2024
@etaf etaf requested review from eellison and zhuhaozhe November 18, 2024 02:20
@etaf etaf changed the title [Inductor XPU] Fix wrong device assign before skip concat linear. [WIP][Inductor XPU] Fix wrong device assign before skip concat linear. Nov 18, 2024
@etaf etaf marked this pull request as draft November 18, 2024 06:11
Copy link
Contributor

@eellison eellison left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The device initialization doesn't really matter here. We ignore devices in the generated pattern matching searches. If you want to skip for cpp, you should update the check functions and check device there.

[ghstack-poisoned]
etaf added a commit that referenced this pull request Nov 19, 2024
…oncat linear.

ghstack-source-id: e2216ad
Pull Request resolved: #140916
@etaf etaf requested a review from eellison November 19, 2024 01:20
@etaf
Copy link
Collaborator Author

etaf commented Nov 19, 2024

The device initialization doesn't really matter here. We ignore devices in the generated pattern matching searches. If you want to skip for cpp, you should update the check functions and check device there.

Thanks @eellison , I've updated the code, please have a review.

@etaf etaf changed the title [WIP][Inductor XPU] Fix wrong device assign before skip concat linear. [Inductor XPU] Fix wrong device assign before skip concat linear. Nov 19, 2024
@etaf etaf marked this pull request as ready for review November 19, 2024 01:21
@etaf etaf requested review from EikanWang and jgong5 November 19, 2024 01:22
@etaf etaf changed the title [Inductor XPU] Fix wrong device assign before skip concat linear. [Inductor XPU] Fix wrong device check before skip concat linear. Nov 19, 2024
if not config.cpp.enable_concat_linear:
return

# workaround https://github.com/pytorch/pytorch/issues/97894
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this workaround comment is not relevant..

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

resolved, thanks.

@etaf
Copy link
Collaborator Author

etaf commented Nov 20, 2024

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here

[ghstack-poisoned]
@pytorchmergebot
Copy link
Collaborator

Successfully rebased gh/etaf/67/orig onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via ghstack checkout https://github.com/pytorch/pytorch/pull/140916)

[ghstack-poisoned]
etaf added a commit that referenced this pull request Nov 20, 2024
…oncat linear.

ghstack-source-id: e292c7f
Pull Request resolved: #140916
[ghstack-poisoned]
etaf added a commit that referenced this pull request Nov 24, 2024
…oncat linear.

ghstack-source-id: 55b1ca0
Pull Request resolved: #140916
@etaf
Copy link
Collaborator Author

etaf commented Nov 26, 2024

xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 3, 4, linux.idc.xpu) (gh)
inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_maximize_xpu
xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 4, 4, linux.idc.xpu) (gh)
inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_weight_decay_xpu
The failed job is known issue: #141466

@pytorchbot merge -i

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Nov 26, 2024
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged while ignoring the following 4 checks: xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 1, 4, linux.idc.xpu), xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 3, 4, linux.idc.xpu), xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 4, 4, linux.idc.xpu), inductor / cuda12.4-py3.10-gcc9-sm86 / test (inductor_timm, 1, 2, linux.g5.4xlarge.nvidia.gpu)

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

The merge job was canceled or timed out. This most often happen if two merge requests were issued for the same PR, or if merge job was waiting for more than 6 hours for tests to finish. In later case, please do not hesitate to reissue the merge command
For more information see pytorch-bot wiki.

@etaf
Copy link
Collaborator Author

etaf commented Nov 26, 2024

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: 2 jobs have failed, first few of them are: xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 3, 4, linux.idc.xpu), xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 4, 4, linux.idc.xpu)

Details for Dev Infra team Raised by workflow job

@etaf
Copy link
Collaborator Author

etaf commented Nov 26, 2024

@pytorchbot merge -i

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged while ignoring the following 4 checks: xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 1, 4, linux.idc.xpu), xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 3, 4, linux.idc.xpu), xpu / linux-jammy-xpu-2025_0-py3.9 / test (default, 4, 4, linux.idc.xpu), inductor / cuda12.4-py3.10-gcc9-sm86 / test (inductor_timm, 1, 2, linux.g5.4xlarge.nvidia.gpu)

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

5 participants