Rate [e.g 0.1, 0.01, 0.001] • Momentum [e.g 0.0, 0.2, 0.4, 0.6] • Number of units in an hidden layer [e.g min = 32 , max = 512] • Number of Hidden Layers [e.g 1 to 4] • Activation function [e.g tanh, relu, sigmoid, softmax] • Droupout(Regularization technique) Rate [e.g 0.0, 0.1, 0.2] Geez, That’s a whole bunch of manual time saved! Imagine doing combinations of these. To know more and understand about these hyperparameters refer to good ol Wikipeida. 12