Optimizing the architecture of a DL classifier
When creating a NN model for a given ML task, one crucial design decision that needs to be made is the configuration of the network architecture. In the case of an MLP, the number of nodes in the input and output layers is determined by the characteristics of the problem at hand. Therefore, the choices to be made are about the hidden layers—how many layers, and how many nodes are in each layer. Some rules of thumb can be employed for making these decisions, but in many cases, identifying the best choices can turn into a cumbersome trial-and-error process.
One way to handle network architecture parameters is to consider them as hyperparameters of the model since they need to be determined before training is done and, consequently, affect the training’s results. In this section, we are going to apply this approach and use a genetic algorithm to search for the best combination of hidden layers, in a similar manner to the...