Skip to content

Latest commit

 

History

History
48 lines (28 loc) · 1.93 KB

README.md

File metadata and controls

48 lines (28 loc) · 1.93 KB

Training a MNIST digit recognizer

The aim here is to achieve the following results with MNIST digit recognizer

  • 99.4% validation accuracy

  • Less than 20k Parameters

  • Less than 20 Epochs

  • Used Batch Normalization, Dropout, a Fully connected layer after have used Global Average Pooling.

Model

The summary of the model defined is shown below

Forwarding
Figure 1.a : Model Summary

Data Augmentation is very important to regularize your network and increase the size of your training set. Data Augmentation strategy used here is random affine transformation. This is done to ensure that the model does not overfit on the training data set and can generalize well on testing data set. In Euclidean geometry, an affine transformation, is a geometric transformation that preserves lines and parallelism (but not necessarily distances and angles).

<iframe src="//commons.wikimedia.org/wiki/File:Affine_transformations.ogv?embedplayer=yes" width="480" height="480" frameborder="0" webkitAllowFullScreen mozallowfullscreen allowFullScreen></iframe>

Training Log

Forwarding
Figure 1.b : Training log

Result

An accurarcy of 99.46 is achieved in 19th epoch with model of 12,674 parameters for the MNIST data

Forwarding
Figure 1.c : Loss and accuracy plots