A linear model classifies samples using separating hyperplanes, so a problem is linearly separable if it's possible to find a linear model whose accuracy overcomes a predetermined threshold. Logistic regression is one of most famous linear classifiers, based on the principle of maximizing the probability of a sample belonging to the right class. SGD classifiers are a more generic family of algorithms, identified by the different loss functions that are adopted. SGD allows partial fitting, particularly when the amount of data is too large to be loaded in memory. A Perceptron is a particular instance of SGD, representing a linear neural network that cannot solve the XOR problem (for this reason, multi-layer perceptrons became the first choice for non-linear classification). However, in general, its performance is comparable to a logistic regression model.
The performances...