Like any Machine Learning model, a Multi Layer Perceptron has an associated risk that is sought to be minimized. Because once again the distribution of the input and output are generally unknown, the empirical risk over some data is used instead.
\(\mathcal{N}\) | This is the symbol used for a function approximator, typically a neural network. |
\(y\) | This symbol stands for the ground truth of a sample. In supervised learning this is often paired with the corresponding input. |
\(\Theta\) | This is the symbol for the set of all possible model parameters \( \htmlId{tooltip-weightVector}{\theta} \). |
\(\theta\) | This is the symbol we use for model weights/parameters. |
\(L\) | This is the symbol for a loss function. It is a function that calculates how wrong a model's inference is compared to where it should be. |
\(u\) | This symbol denotes the input of a model. |
Note: Other terms such as can be added to the formulation in similar ways: see Loss Minimization with Regularization.
Was this page helpful?