Skip to content

BaoWangMath/DNN-DataDependentActivation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

40 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DNN-DataDependentActivation

This repository consists PyTorch code for deep neural networks with graph interpolating function as output activation function

Place the pyflann library in your current directory to replace the pyflann folder

Cifar10-Natural

Code for reproducing results of naturally trained ResNets on the Cifar10

Cifar10-Robust

Code for reproducing results of PGD adversarial training for ResNets on the Cifar10

MNIST-Robust

Code for reproducing results of PGD adversarial training for Small-CNN on the MNIST

If you find this work useful and use it on you own research, please cite our paper

@incollection{NIPS2018_7355,
title = {Deep Neural Nets with Interpolating Function as Output Activation},
author = {Wang, Bao and Luo, Xiyang and Li, Zhen and Zhu, Wei and Shi, Zuoqiang and Osher, Stanley},
booktitle = {Advances in Neural Information Processing Systems 31},
editor = {S. Bengio and H. Wallach and H. Larochelle and K. Grauman and N. Cesa-Bianchi and R. Garnett},
pages = {743--753},
year = {2018},
publisher = {Curran Associates, Inc.},
url = {http://papers.nips.cc/paper/7355-deep-neural-nets-with-interpolating-function-as-output-activation.pdf}
}

And the longer version is available at

@ARTICLE{Wang:2019Interpolation,
       author = {B. Wang and S. Osher},
        title = "{Graph Interpolating Activation Improves Both Natural and Robust Accuracies in Data-Efficient Deep Learning}",
      journal = {arXiv e-prints},
         year = "2019",
        month = "Jul",
          eid = {arXiv:1907.06800},
        pages = {arXiv:1907.06800},
archivePrefix = {arXiv},
       eprint = {},
 primaryClass = {stat.ML}
}

Dependence

PyTorch 0.4.1

Releases

No releases published

Packages

No packages published

Languages