Skip to content

ThomasMiconi/DiffRNN

Repository files navigation

This is the source code for the arXiv preprint "Neural networks with differentiable structure".

This code implements recurrent neural networks with differentiable structure: the number of neurons in the network undergoes gradient descent, just like the weights of the network. The network adjusts its number of neurons to the complexity of the task at hand.

This code is based on Andrej Karpathy's min-char-rnn.py program.

rnn.py is the main program. You can run it "as is" (python rnn.py) to run the model on the "hard" problem for 100000 cycles. It will generate an output file called output.txt, updated every 1000 cycles, which logs the current cycle number, position in the input file, loss, number of neurons, and total absolute sum of multipliers. (see code).

Other python files in the repository generate inputs or figures, or submit jobs to a cluster.

About

RNN with differentiable structure (number of neurons)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors