diff --git a/README.md b/README.md index fbca6602913392b71c170f2ed5e4997e7e81f3b7..965d381370306cfb3b99f23ab975b1f55a99f13f 100644 --- a/README.md +++ b/README.md @@ -4,6 +4,16 @@ We study these binary activations with two datasets: [Part1: MNIST](#part1-mnist This repository uses Pytorch library. +Colaboratory notebooks for part1 et part2 contains some differents visualization tools to compare networks with binary weights and network with no binary weights, e.g.: +- visualization activations values for a specific data +- visualization filters trained by network +- visualization heatmap for prediction data +- visualization regions that maximizes a specific filter +- visualization generated images for activation maximization with gradient a ascent +- visualization 1 nearest neighbor classification with different global representation of data + +Learning networks code use Pytorch Ignite (in "experiments/MNIST_binary_Run_Notebook.ipynb). + # Introduction: train discrete variables To train a neural network with discrete variables, we can use two methods: REINFORCE (E (Williams, 1992; Mnih & Gregor,2014) and the straight-through estimator (Hinton, 2012; Bengio et al., 2013).