KEMBAR78
[Inductor] [Quant] Enable QConv2d Unary int8-mixed-bf16 Lowering by leslie-fang-intel · Pull Request #112550 · pytorch/pytorch · GitHub
Skip to content

Conversation

leslie-fang-intel
Copy link
Collaborator

@leslie-fang-intel leslie-fang-intel commented Nov 1, 2023

@pytorch-bot
Copy link

pytorch-bot bot commented Nov 1, 2023

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/112550

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit c6718ec with merge base 8bdce9b (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

leslie-fang-intel added a commit to leslie-fang-intel/pytorch that referenced this pull request Nov 2, 2023
…wering"


**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```


cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx peterbell10 ipiszy yf225 chenyang78 kadeng muchulee8 aakhundov ColinPeppler

[ghstack-poisoned]
leslie-fang-intel added a commit to leslie-fang-intel/pytorch that referenced this pull request Nov 3, 2023
…wering"


**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```


cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx peterbell10 ipiszy yf225 chenyang78 kadeng muchulee8 aakhundov ColinPeppler

[ghstack-poisoned]
leslie-fang-intel added a commit to leslie-fang-intel/pytorch that referenced this pull request Nov 3, 2023
…wering"


**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```


cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx peterbell10 ipiszy yf225 chenyang78 kadeng muchulee8 aakhundov ColinPeppler

[ghstack-poisoned]
…wering"


**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```


cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx peterbell10 ipiszy yf225 chenyang78 kadeng muchulee8 aakhundov ColinPeppler

[ghstack-poisoned]
@leslie-fang-intel
Copy link
Collaborator Author

Hi @eellison @jerryzh168, Could you kindly help to review these 3 PRs in this ghstack for the Inductor lowering of int8-mixed-bf16 quantization case?

@eellison
Copy link
Contributor

eellison commented Nov 7, 2023

@jerryzh168 want to review this one ?

Copy link
Contributor

@jerryzh168 jerryzh168 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LG overall, please add more some docs

…wering"


**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```


cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx peterbell10 ipiszy yf225 chenyang78 kadeng muchulee8 aakhundov ColinPeppler

[ghstack-poisoned]
@leslie-fang-intel
Copy link
Collaborator Author

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

pytorchmergebot pushed a commit that referenced this pull request Nov 10, 2023
…12551)

**Summary**
- PR 6 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor #111640.
- Enable the QConv2d Binary int8-mixed-bf16 post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```

Pull Request resolved: #112551
Approved by: https://github.com/jgong5, https://github.com/eellison, https://github.com/jerryzh168
ghstack dependencies: #112550
@facebook-github-bot facebook-github-bot deleted the gh/leslie-fang-intel/40/head branch November 13, 2023 15:25
Skylion007 pushed a commit to Skylion007/pytorch that referenced this pull request Nov 14, 2023
…orch#112550)

**Summary**
- PR 5 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor pytorch#111640.
- Enable the QConv2d Unary int8-mixed-bf16 weight prepack and post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```

Pull Request resolved: pytorch#112550
Approved by: https://github.com/jgong5, https://github.com/jerryzh168
Skylion007 pushed a commit to Skylion007/pytorch that referenced this pull request Nov 14, 2023
…torch#112551)

**Summary**
- PR 6 for enabling Int8-Mixed-BF16 PT2E PTQ Quantization with Inductor pytorch#111640.
- Enable the QConv2d Binary int8-mixed-bf16 post grad lowering inside inductor.

**TestPlan**
```
python -m pytest test_mkldnn_pattern_matcher.py -k test_qconv2d
```

Pull Request resolved: pytorch#112551
Approved by: https://github.com/jgong5, https://github.com/eellison, https://github.com/jerryzh168
ghstack dependencies: pytorch#112550
k-artem added a commit to ROCm/pytorch that referenced this pull request Jun 13, 2025
* Most part of testcases work properly on Navi48(gfx1201) with
  TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1, in this commit enable it
  for this arch. No support of AOTriton currently for Navci44(gfx1200),
  so these testcases just skipped.
* test_qconv2d_int8_mixed_bf16 skipped because it was originally skipped in
  pytorch#112550 but later lost.
* test_sac_ilp_case1 skipped as per SWDEV-509011
* test_distributed_checkpoint_state_dict_type[0-1]_cuda fixed bug
  with arguments.
pruthvistony pushed a commit to ROCm/pytorch that referenced this pull request Jul 7, 2025
… SDPA or Navi4x (#2213)

[release/2.6][SWDEV-523736] Skip some testcases for archs without SDPA
or Navi4x

[SWDEV-523736] Fix some unittests for Navi4x
* Most part of testcases work properly on Navi48(gfx1201) with
  TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1, in this commit enable it
  for this arch. No support of AOTriton currently for Navci44(gfx1200),
  so these testcases just skipped.
* test_qconv2d_int8_mixed_bf16 skipped because it was originally skipped
in
  pytorch#112550 but later lost.
* test_sac_ilp_case1 skipped as per SWDEV-509011
* test_distributed_checkpoint_state_dict_type[0-1]_cuda fixed bug
  with arguments.

Fixes #SWDEV-523736
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants