Skip to content

Conversation

@XiaobingSuper
Copy link
Collaborator

@XiaobingSuper XiaobingSuper commented May 16, 2019

mkldnn backward ops list:

@pytorchbot pytorchbot added module: autograd Related to torch.autograd, and the autograd engine in general module: internals Related to internal abstractions in c10 and ATen module: mkldnn Related to Intel IDEEP or oneDNN (a.k.a. mkldnn) integration module: operators labels May 16, 2019
@li-roy li-roy added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label May 16, 2019
facebook-github-bot pushed a commit that referenced this pull request Jun 13, 2019
Summary:
### mkldnn backward ops list:
 - [ ] \(#20567) Add aten mkldnn conv2d backward operator 💛
 - [ ] \(#20570) Add aten mkldnn backward ops: relu, linear and reshape 💛
 - [ ] \(#20571) Add aten mkldnn backward ops: max_pool2d, avg_pool2d and adaptive_avg_poo2d 💛
 - [ ] \(#20572) Add aten mkldnn batchnorm backward operator 💛
 - [ ] \(#20573) Add aten mkldnn zero_ operator:yellow_heart:
 - [ ] \(#20575) Add mkldnn mul operator 💛
Pull Request resolved: #20575

Differential Revision: D15799529

Pulled By: bddppq

fbshipit-source-id: 4887d8ef1a0e316ad9db199b657d9481fc13e486
zdevito pushed a commit to zdevito/ATen that referenced this pull request Jun 13, 2019
Summary:
### mkldnn backward ops list:
 - [ ] \(pytorch/pytorch#20567) Add aten mkldnn conv2d backward operator 💛
 - [ ] \(pytorch/pytorch#20570) Add aten mkldnn backward ops: relu, linear and reshape 💛
 - [ ] \(pytorch/pytorch#20571) Add aten mkldnn backward ops: max_pool2d, avg_pool2d and adaptive_avg_poo2d 💛
 - [ ] \(pytorch/pytorch#20572) Add aten mkldnn batchnorm backward operator 💛
 - [ ] \(pytorch/pytorch#20573) Add aten mkldnn zero_ operator:yellow_heart:
 - [ ] \(pytorch/pytorch#20575) Add mkldnn mul operator 💛
Pull Request resolved: pytorch/pytorch#20575

Differential Revision: D15799529

Pulled By: bddppq

fbshipit-source-id: 4887d8ef1a0e316ad9db199b657d9481fc13e486
facebook-github-bot pushed a commit that referenced this pull request Jun 14, 2019
Summary:
### mkldnn backward ops list:
 - [ ] \(#20567) Add aten mkldnn conv2d backward operator 💛
 - [ ] \(#20570) Add aten mkldnn backward ops: relu, linear and reshape 💛
 - [ ] \(#20571) Add aten mkldnn backward ops: max_pool2d, avg_pool2d and adaptive_avg_poo2d 💛
 - [ ] \(#20572) Add aten mkldnn batchnorm backward operator 💛
 - [ ] \(#20573) Add aten mkldnn zero_ operator:yellow_heart:
 - [ ] \(#20575) Add mkldnn mul operator 💚
Pull Request resolved: #20573

Differential Revision: D15820477

Pulled By: bddppq

fbshipit-source-id: 35d95f5b4e013c8db1911f52148550a2e40a2e68
zdevito pushed a commit to zdevito/ATen that referenced this pull request Jun 14, 2019
Summary:
### mkldnn backward ops list:
 - [ ] \(pytorch/pytorch#20567) Add aten mkldnn conv2d backward operator 💛
 - [ ] \(pytorch/pytorch#20570) Add aten mkldnn backward ops: relu, linear and reshape 💛
 - [ ] \(pytorch/pytorch#20571) Add aten mkldnn backward ops: max_pool2d, avg_pool2d and adaptive_avg_poo2d 💛
 - [ ] \(pytorch/pytorch#20572) Add aten mkldnn batchnorm backward operator 💛
 - [ ] \(pytorch/pytorch#20573) Add aten mkldnn zero_ operator:yellow_heart:
 - [ ] \(pytorch/pytorch#20575) Add mkldnn mul operator 💚
Pull Request resolved: pytorch/pytorch#20573

Differential Revision: D15820477

Pulled By: bddppq

fbshipit-source-id: 35d95f5b4e013c8db1911f52148550a2e40a2e68
@Jianhui-Li
Copy link

@gottbrath @bddppq @dzhulgakov @wesolwsk @VitalyFedyunin @ezyang This batch of PRs enabled the MKL-DNN for the training path and is expected to improve training models like resnext101 by 2x on CPU with AVX512 (also benefit AVX2). With more DL accelerating instructions like bfp16 in CPU roadmap, we think it is important to accelerate the Pytorch training path as we do for the inference path so the community get the benefits.

@dzhulgakov
Copy link
Collaborator

Hey, @XiaobingSuper - is it possible to rebase these PRs on top of current master and make sure the CI passes?

@XiaobingSuper XiaobingSuper force-pushed the mkldnn_pool_bwd branch 4 times, most recently from 92aa6eb to 93c3595 Compare July 18, 2019 05:11
@dariogarcia
Copy link

My research group is really interested in this. We are using PyTorch on CPUs because our research has extreme memory requirements, and TensorFlow has a limited tensor size whereas PyTorch does not.

@gbaydin
Copy link

gbaydin commented Sep 6, 2019

Hi, our group is interested in this PR to improve CPU training in our large-scale probabilistic programming work.

@wbhimji
Copy link

wbhimji commented Oct 11, 2019

I'd like to also support that we need these optimizations for the PyTorch we run in support of large-scale science at the NERSC supercomputing center. Currently we blocking on using more recent PyTorch 1.2 and 1.3 for some projects because of this. Thanks!

@XiaobingSuper
Copy link
Collaborator Author

@dariogarcia, @gbaydin, @wbhimji, thank you for your interest in our work. I will rebase the code after get the reponse from Facebook team.

@XiaobingSuper
Copy link
Collaborator Author

@VitalyFedyunin

@DavidRuhe
Copy link

We at SURF (national super-computing facilities in the Netherlands) are also really interested in better MKL support in PyTorch. In our case specifically 3D operations.

@XiaobingSuper
Copy link
Collaborator Author

@VitalyFedyunin

Copy link
Contributor

@VitalyFedyunin VitalyFedyunin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Overall looks good, can you please rebase for final tests. Also what is the plan to start renaming functions to onednn?

@XiaobingSuper
Copy link
Collaborator Author

Overall looks good, can you please rebase for final tests. Also what is the plan to start renaming functions to onednn?

@VitalyFedyunin , the renaming function will be ok at next step which will upgrade DNNL to v1.4, so I will rebase those PRs which stiil using the old name, thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

module: autograd Related to torch.autograd, and the autograd engine in general module: internals Related to internal abstractions in c10 and ATen module: mkldnn Related to Intel IDEEP or oneDNN (a.k.a. mkldnn) integration open source triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.