Skip to content

Conversation

@ptrblck
Copy link
Collaborator

@ptrblck ptrblck commented Jul 30, 2019

  • Swapped CUBLAS_OP_N for 'n'
  • added a test

This PR should fix #23545.

Thanks at AlphabetMan for reporting the initial issue reported in the forum as well as @ngimel for the guidance.

@pytorchbot pytorchbot added module: cuda Related to torch.cuda, and CUDA support in general module: nn Related to torch.nn labels Jul 30, 2019
test/test_nn.py Outdated

@unittest.skipIf(not TEST_CUDA, 'CUDA not available')
def test_ConvTranspose2d_half_cublas_gemm(self):
inputs = torch.randn(1, 1, 16, 16, device='cuda', dtype=torch.half)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

you can also force CuDNN to be off (for future safety) using the context manager we have available (i think it's called with torch.backends.cudnn something something

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! I've added the context manager.

@ssnl
Copy link
Collaborator

ssnl commented Jul 31, 2019

@pytorchbot merge this please

also @soumith , this probably wants to be in 1.2 too i suppose?

@pytorchbot pytorchbot added the merge-this-please Was marked for merge with @pytorchbot merge this please label Jul 31, 2019
Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ezyang is landing this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@ptrblck ptrblck deleted the cublas branch July 31, 2019 17:08
zdevito pushed a commit to zdevito/ATen that referenced this pull request Jul 31, 2019
Summary:
* Swapped `CUBLAS_OP_N` for `'n'`
* added a test

This PR should fix pytorch/pytorch#23545.

Thanks at AlphabetMan for reporting the initial issue reported in [the forum](https://discuss.pytorch.org/t/cuda-10-1-error-using-transposeconv2d-with-output-padding-1/51414?u=ptrblck) as well as ngimel for the guidance.
Pull Request resolved: pytorch/pytorch#23552

Differential Revision: D16580986

Pulled By: ezyang

fbshipit-source-id: abc0bce1e84d9c9d96d44ae0296951725adc8424
@facebook-github-bot
Copy link
Contributor

@ezyang merged this pull request in 9130ab3.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

merge-this-please Was marked for merge with @pytorchbot merge this please Merged module: cuda Related to torch.cuda, and CUDA support in general module: nn Related to torch.nn open source

Projects

None yet

Development

Successfully merging this pull request may close these issues.

THCUNN convolutions fail with half inputs/weights

7 participants