neural-reader
Overview
My first adventure into machine learning was an MNIST classifier I wrote after learning about neural networks in Dr. Terry Pilling’s advanced engineering math course at UMary.
In true UMary Engineering style, the full training loop and classifier are written from scratch in C with only basic libraries.
The approach and architecture are far from ideal, but with persistence, I was able to achieve >95% accuracy. The network has four layers, each 784 neurons tall. The height comes from the 28x28 resolution of the MNIST dataset. There are, of course, $28^2 = 784$ pixels in each image, each needing to be fed into the network. For code simplicity, I chose to keep each layer the same height. Each neuron in the output layer is trained to output a value between 0.0 and 0.9, which maps to 0-9. A final output is produced by averaging across the output layer.

Downloading and Compiling
The code is available on my GitHub page. You can try it yourself by following these instructions:
- Clone the repository:
git clone https://github.com/johnsoupir/neural-reader.git
- Change to the project directory:
cd neural-reader
- Compile the code:
make all
If this fails because make
is not installed, run:
sudo apt install make
Training and Recognition
A pre-trained network is included, but to train your own:
./train
The network can continue training from a previous parameter file by setting the oldDogNewTricks
variable to true
.
To recognize using the pre-trained parameters, run:
./recognize