The first architectural constraints that you must consider while building a network with dense layers are its the depth and width. Then, you need to define an input layer with the appropriate shape, and successively choose from different activation functions to use per layer.
As we did for our MNIST example, we simply import the sequential model and the dense layer structure. Then we proceed by initializing an empty sequential model and progressively add hidden layers until we reach the output layer. Do note that our input layer always requires a specific input shape, which for us corresponds to the 12,000 - dimensional one-hot encoded vectors that we will be feeding it. In our current model, the output layer only has one neuron, which will ideally fire if the sentiment in a given review is positive; otherwise, it won't. We will choose Rectified Linear...