Inputs are first handed via some absolutely linked layer, to a double-layer residual multihead notice as demonstrated in Fig. seven. Residual networks (Kaiming He, 2016), integrate feedforward to prevent neurons from suffering from exploding or vanishing gradients in the course of the training approach. The fully linked levels in the residual block