Skip to content

Generic implementations of neural network models and training alogs from scratch

Notifications You must be signed in to change notification settings

madvn/neural_nets_from_scratch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Implementations of Neural Network models and training algorithms from scratch [WIP]

Install

# Navigate to this repo in the terminal
pip install -e .

Usage

from neural_nets.backprop import BackPropNet

Models and Algos

  • Feedforward network of arbitrary size and activation functions

  • Backprop with arbitrary cost function

  • Feedback alignment based training for FNNs [1]

  • Recurrent network with arbitrary size and activtion functions

  • Backprop through time [WIP]

  • RNN traninng with feedback alignment [2]

Dependencies

Numpy and Matplotlib - planning to use JAX for some of the grad operations.

References

[1] Lillicrap, Timothy P., et al. Random synaptic feedback weights support error backpropagation for deep learning. Nature communications 7 (2016): 13276

[2] Murray, J. M. (2018). Local online learning in recurrent networks with random feedback. BioRxiv, 458570.

About

Generic implementations of neural network models and training alogs from scratch

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages