Machine learning in computational mechanics (draft)
1. Machine learning
2. Deep learning
Neural networks (NNs) are powerful function approximators capable of modeling any continuous function. A neural network, parameterized by learnable parameters (typically weights and biases ), learns a function that approximates the relationship . NNs are built using nested linear transformations combined with non-linear activation functions . The simplest form, fully connected neural networks, achieve this through layers of fully connected neurons. The activation of each neuron (the th neuron in layer ) is computed by applying a non-linear activation function to a linear combination of the activations from the previous layer.
The most basic NNs: Fully connected NNs achieve this with layers of fully connected neurons, where the activation of each neuron (the th neuron of layer ) is obtained through linear combinations of the previous layer and the non-linear activation function :

If more than one layer (excluding input and output layer ) is employed, the NN is considered a deep NN, and its training process is thereby deep learning. The evaluation of the NN, i.e., the prediction is referred to as forward propagation. The quality of prediction is determined by a cost function , which is to be minimized. Its gradients with respect to the parameters are obtained with automatic differentiation, specifically referred to as backward propagation in the context of NNs. The gradients are used within a gradient-based optimization to update the parameters and thereby improve the prediction .