Skip to content

Conversation

@ngimel
Copy link
Collaborator

@ngimel ngimel commented Jun 10, 2019

Stream is not respected on range/linspace/logspace functions, which contributes to #21589 (this is not a complete solution for that issue).

@pytorchbot pytorchbot added module: cuda Related to torch.cuda, and CUDA support in general module: operators labels Jun 10, 2019
@ngimel
Copy link
Collaborator Author

ngimel commented Jun 11, 2019

@iotamudelta any advice on fixing

/var/lib/jenkins/workspace/aten/src/ATen/native/hip/RangeFactories.hip:61:35: error: no member named 'par' in namespace 'thrust::cuda'
22:30:07       auto policy = thrust::cuda::par.on(stream);

?
In other places (e.g. Embedding.cu) the following construct is used

auto policy = thrust::cuda::par(allocator).on(stream);

but I don't need allocator for the ranges, and thrust::cuda::par.on(stream) is valid syntax.

@iotamudelta
Copy link
Contributor

@ngimel looks to me like a shortcoming of the hipified Thrust. Could this be a construct that didn't exist in Thrust some two years ago? Either way, we are working on updating things in that realm, so could you work around it via __HIP_COMPILER_HCC__ ifdefs (either w/ allocator or serial execution, I guess) for now and leave a TODO in the code so that I can fix it once the bigger picture there is sorted out? Thanks!

@ngimel
Copy link
Collaborator Author

ngimel commented Jun 11, 2019

rocm builds were fixed by including header, caffe2_onnx failure seems unrelated. BTW, what happened to cuda 10 ci?

@ngimel ngimel requested a review from gchanan June 11, 2019 20:29
@ezyang
Copy link
Contributor

ezyang commented Jun 11, 2019

BTW, what happened to cuda 10 ci?

It's run on master, but not on every commit. If you like, I can make it possible to explicitly request cuda 10 CI on pull requests, if you're futzing around with CUDA 10 specific code.

Copy link
Contributor

@ezyang ezyang left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Very nice.

@ezyang
Copy link
Contributor

ezyang commented Jun 11, 2019

I'm going to go ahead and merge this, but ideally we'd also turn back on the failing tests when we make fixes like this. (This might be hard/annoying to do.)

Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ezyang is landing this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@ngimel
Copy link
Collaborator Author

ngimel commented Jun 11, 2019

But it does not fix the tests :-( Most of the tests fail because of backward running on the wrong stream. This is something that I noticed looking on the profile, and that has to be fixed anyway.

zdevito pushed a commit to zdevito/ATen that referenced this pull request Jun 12, 2019
Summary:
Stream is not respected on range/linspace/logspace functions, which contributes to pytorch/pytorch#21589 (this is not a complete solution for that issue).
Pull Request resolved: pytorch/pytorch#21619

Differential Revision: D15769666

Pulled By: ezyang

fbshipit-source-id: 7c036f7aecb3119430c4d432775cad98a5028fa8
@facebook-github-bot
Copy link
Contributor

@ezyang merged this pull request in cf5c3bb.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Merged module: cuda Related to torch.cuda, and CUDA support in general open source

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants