Skip to content

Conversation

@kexinyu
Copy link

@kexinyu kexinyu commented Jul 2, 2019

Currently, the stream parameter is not set when launching these two kernels: softmax_warp_forward() and softmax_warp_backward(), i.e. the kernels are always put on the default stream, which may fail to respect the stream that was set previously. Add at::cuda::getCurrentCUDAStream() as a launch argument to fix this issue.

@pytorchbot pytorchbot added module: cuda Related to torch.cuda, and CUDA support in general module: operators labels Jul 2, 2019
@ngimel
Copy link
Collaborator

ngimel commented Jul 3, 2019

@pytorchbot merge this please

@pytorchbot pytorchbot added the merge-this-please Was marked for merge with @pytorchbot merge this please label Jul 3, 2019
Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@izdeby has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@ezyang
Copy link
Contributor

ezyang commented Jul 5, 2019

Related #21589

zdevito pushed a commit to zdevito/ATen that referenced this pull request Jul 5, 2019
Summary:
Currently, the **stream** parameter is not set when launching these two kernels: softmax_warp_forward() and softmax_warp_backward(), i.e. the kernels are always put on the default stream, which may fail to respect the stream that was set previously. Add **at::cuda::getCurrentCUDAStream()** as a launch argument to fix this issue.
Pull Request resolved: pytorch/pytorch#22470

Differential Revision: D16115051

Pulled By: izdeby

fbshipit-source-id: 38b27e768bb5fcecc1a06143ab5d63b0e68a279e
@facebook-github-bot
Copy link
Contributor

@izdeby merged this pull request in e5d6403.

xzhu1900 pushed a commit to xzhu1900/pytorch that referenced this pull request Jul 5, 2019
Summary:
Currently, the **stream** parameter is not set when launching these two kernels: softmax_warp_forward() and softmax_warp_backward(), i.e. the kernels are always put on the default stream, which may fail to respect the stream that was set previously. Add **at::cuda::getCurrentCUDAStream()** as a launch argument to fix this issue.
Pull Request resolved: pytorch#22470

Differential Revision: D16115051

Pulled By: izdeby

fbshipit-source-id: 38b27e768bb5fcecc1a06143ab5d63b0e68a279e
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

merge-this-please Was marked for merge with @pytorchbot merge this please Merged module: cuda Related to torch.cuda, and CUDA support in general open source

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants