#Proofinatoot Sigmoid and softmax learn an equivalent classifier.

Let \[\sigma(wx + b) = 0.5\] be the decision boundary for a sigmoid classifier.

Then, for a softmax, \[\exp(w_1 x + b_1) / [(\exp(w_1x + b_1) + \exp(w_2x + b_2)] = 0.5\] implies \[\exp(w_1x + b_1) = \exp(w_2x + b_2)\] implies \[w_1x + b_1 = w_2x + b_2\] implies \[(w_1 - w_2)x + (b_1 - b_2) = 0.\]

Source for this toot: https://stats.stackexchange.com/questions/198038/cross-entropy-or-log-likelihood-in-output-layer

Why does SGD work? We might finally know.

(1) A sufficient represenation of data is invariant to nuissance if and only if it holds the lowest possible amount of information about the data.

(2) The information of a representation is bounded by the information in the weights.

(3) The information in the weights is lowest at flat minima rather than sharp minima. Because SGD seeks flat minima, it forces generalization.

arxiv.org/abs/1706.01350

Masters in data science - looking to start a PhD in graphical models. Posts about statistical learning theory and neural networks.

Joined Jun 2017