KEMBAR78
Improve MHA docs by jbschlosser · Pull Request #61977 · pytorch/pytorch · GitHub
Skip to content

Conversation

@jbschlosser
Copy link
Contributor

@jbschlosser jbschlosser commented Jul 21, 2021

Fixes #60831
Also clarifies the relationship between embed_dim and num_heads (see #60853 and #60445).
Formatting was overhauled to remove some redundancy between the input docs and shape docs; suggestions / comments welcome!

Link to rendered docs here: https://14912919-65600975-gh.circle-artifacts.com/0/docs/generated/torch.nn.MultiheadAttention.html

@facebook-github-bot
Copy link
Contributor

facebook-github-bot commented Jul 21, 2021

🔗 Helpful links

💊 CI failures summary and remediations

As of commit bf3d733 (more details on the Dr. CI page):


  • 1/1 failures introduced in this PR

1 failure not recognized by patterns:

Job Step Action
GitHub Actions Linux CI (pytorch-linux-bionic-py3.8-gcc9-coverage) / render_test_results (default) Checkout PyTorch 🔁 rerun

This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.

Please report bugs/suggestions to the (internal) Dr. CI Users group.

Click here to manually regenerate this comment.

Copy link

@bhosmer bhosmer left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This looks great!

@facebook-github-bot
Copy link
Contributor

@jbschlosser has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@facebook-github-bot
Copy link
Contributor

@jbschlosser has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@facebook-github-bot
Copy link
Contributor

@jbschlosser merged this pull request in e408af0.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Misleading documentation in torch.nn.MultiheadAttention

3 participants