KEMBAR78
[TP] Enable embedding sharding in TP API by fduwjj · Pull Request #111177 · pytorch/pytorch · GitHub
Skip to content

Conversation

@fduwjj
Copy link
Contributor

@fduwjj fduwjj commented Oct 13, 2023

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 13, 2023

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/111177

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit cd69e2d with merge base 35750bf (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

fduwjj added a commit that referenced this pull request Oct 13, 2023
ghstack-source-id: e1c0bf4
Pull Request resolved: #111177
@fduwjj fduwjj added ciflow/trunk Trigger trunk jobs on your pull request module: dtensor distributed tensor tag release notes: distributed (dtensor) release notes category labels Oct 13, 2023
Copy link
Collaborator

@wanchaol wanchaol left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you also update the doc of ColwiseParallel and RowwiseParallel style to explicitly mention what modules are supported?

@fduwjj fduwjj requested a review from wanchaol October 13, 2023 21:38
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
fduwjj added a commit that referenced this pull request Oct 13, 2023
ghstack-source-id: 25fa245
Pull Request resolved: #111177
Copy link
Collaborator

@wanchaol wanchaol left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

stamp to unblock, some suggestions about the doc

We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
@fduwjj
Copy link
Contributor Author

fduwjj commented Oct 15, 2023

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: 1 mandatory check(s) failed. The first few are:

Dig deeper by viewing the failures on hud

Details for Dev Infra team Raised by workflow job

Failing merge rule: Core Maintainers

We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.


[ghstack-poisoned]
fduwjj added a commit that referenced this pull request Oct 15, 2023
ghstack-source-id: 979ddcb
Pull Request resolved: #111177
@fduwjj
Copy link
Contributor Author

fduwjj commented Oct 15, 2023

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@wanchaol
Copy link
Collaborator

@fduwjj can you fix the doc errors? Looks like the doc generated from this PR have multiple issues, see https://docs-preview.pytorch.org/pytorch/pytorch/111177/distributed.tensor.parallel.html#torch.distributed.tensor.parallel.style.ColwiseParallel

All the examples in the doc (i.e. Colwise/Rowwise) are not properly formatted, and also parallelize_module example can probably remove Pairwise and use ColwiseParallel

@fduwjj
Copy link
Contributor Author

fduwjj commented Oct 16, 2023

@wanchaol Fix in #111346.

pytorchmergebot pushed a commit that referenced this pull request Oct 16, 2023
yeounoh pushed a commit to yeounoh/pytorch that referenced this pull request Oct 16, 2023
We see use cases where embedding sharding is also needed in TP API so we enabled it in the API since DTensor already support colwise embedding sharding.

Pull Request resolved: pytorch#111177
Approved by: https://github.com/wanchaol
ghstack dependencies: pytorch#111160, pytorch#111166, pytorch#111176
yeounoh pushed a commit to yeounoh/pytorch that referenced this pull request Oct 16, 2023
@facebook-github-bot facebook-github-bot deleted the gh/fduwjj/110/head branch October 18, 2023 14:25
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/trunk Trigger trunk jobs on your pull request Merged module: dtensor distributed tensor tag release notes: distributed (dtensor) release notes category

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants