KEMBAR78
Add BF16 in FP8 quantize ops by sryap · Pull Request #1961 · pytorch/FBGEMM · GitHub
Skip to content

Conversation

@sryap
Copy link
Contributor

@sryap sryap commented Aug 21, 2023

Summary:

  • Added output_dtype for half, bfloat16 and float as output in
    dequantization functions; currently it's an integer value defined by
    Sparse_dtype (float:0, half:1, bfloat16:5)
  • Added type conversion in quant and dequant kernels by using native
    CUDA/HIP functions for half to float conversion and writing
    everything explicitly.

Differential Revision: D47904459

@netlify
Copy link

netlify bot commented Aug 21, 2023

Deploy Preview for pytorch-fbgemm-docs canceled.

Name Link
🔨 Latest commit 19fb8e1
🔍 Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/64e6a4ae3f99cf000870964e

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

sryap added a commit to sryap/FBGEMM that referenced this pull request Aug 21, 2023
Summary:
Pull Request resolved: pytorch#1961

- Added output_dtype for half, bfloat16 and float as output in
  dequantization functions; currently it's an integer value defined by
  Sparse_dtype (float:0, half:1, bfloat16:5)
- Added type conversion in quant and dequant kernels by using native
  CUDA/HIP functions for half to float conversion and writing
  everything explicitly.

Differential Revision: D47904459

fbshipit-source-id: 41d3f0c50365d0482aab912c202f458a787419d8
@sryap sryap force-pushed the export-D47904459 branch from 8e44a0d to facb7ed Compare August 21, 2023 22:22
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

sryap added a commit to sryap/FBGEMM that referenced this pull request Aug 23, 2023
Summary:
Pull Request resolved: pytorch#1961

- Added output_dtype for half, bfloat16 and float as output in
  dequantization functions; currently it's an integer value defined by
  Sparse_dtype (float:0, half:1, bfloat16:5)
- Added type conversion in quant and dequant kernels by using native
  CUDA/HIP functions for half to float conversion and writing
  everything explicitly.

Reviewed By: jianyuh

Differential Revision: D47904459

fbshipit-source-id: f608d7da5dcf05ff78a6e0eb13d985ed99207d1a
@sryap sryap force-pushed the export-D47904459 branch from facb7ed to 56e870d Compare August 23, 2023 23:25
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

sryap added a commit to sryap/FBGEMM that referenced this pull request Aug 24, 2023
Summary:
Pull Request resolved: pytorch#1961

- Added output_dtype for half, bfloat16 and float as output in
  dequantization functions; currently it's an integer value defined by
  Sparse_dtype (float:0, half:1, bfloat16:5)
- Added type conversion in quant and dequant kernels by using native
  CUDA/HIP functions for half to float conversion and writing
  everything explicitly.

Reviewed By: jianyuh

Differential Revision: D47904459

fbshipit-source-id: 3fdca310b5262c249e7dc552070e27a569c9af23
@sryap sryap force-pushed the export-D47904459 branch from 56e870d to 7d5b278 Compare August 24, 2023 00:20
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

sryap added a commit to sryap/FBGEMM that referenced this pull request Aug 24, 2023
Summary:
Pull Request resolved: pytorch#1961

- Added output_dtype for half, bfloat16 and float as output in
  dequantization functions; currently it's an integer value defined by
  Sparse_dtype (float:0, half:1, bfloat16:5)
- Added type conversion in quant and dequant kernels by using native
  CUDA/HIP functions for half to float conversion and writing
  everything explicitly.

Reviewed By: jianyuh

Differential Revision: D47904459

fbshipit-source-id: 4f5bbf71cb3c5f0ec2f4ef0048f30c6cdf48cd2e
@sryap sryap force-pushed the export-D47904459 branch from 7d5b278 to a6ee85a Compare August 24, 2023 00:25
Summary:
Pull Request resolved: pytorch#1961

- Added output_dtype for half, bfloat16 and float as output in
  dequantization functions; currently it's an integer value defined by
  Sparse_dtype (float:0, half:1, bfloat16:5)
- Added type conversion in quant and dequant kernels by using native
  CUDA/HIP functions for half to float conversion and writing
  everything explicitly.

Reviewed By: jianyuh

Differential Revision: D47904459

fbshipit-source-id: d48da0fc7b0b158c46628952a7c7ec8e1aa502df
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D47904459

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 4920770.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants