-
Notifications
You must be signed in to change notification settings - Fork 227
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add batch dim idx to support latest deepspeed DistributedAttention #1725
Conversation
…Attention (#37) * Temp patch for batch_dim * Adding batch_dim_idx as per latest deepspeed * Update modeling_llama.py
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
@regisss Please review it and merge along with other 1.20 Dependent PRs |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I left one comments to address + @mlapinskix's comments
@regisss Addressed review comments. Can you review and merge |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
https://github.com/microsoft/DeepSpeed/commits/master/deepspeed/sequence/layer.py
With latest changes in DistributedAttention, we need to add batch_dim_idx and rotary_pos_emb to Distributed Attention.
This change brings changes to add the batch_dim_idx in line with new changes.
deepspeedai/DeepSpeed@ffe0af2
Fixes # (issue)
Before submitting