With our definition of the targets y, the outputs of the network ˜y and the inputs x we define now the activation zlj of node/neuron/unit j of the l-th layer as a function of the bias, the weights which add up from the previous layer l−1 and the forward passes/outputs al−1 from the previous layer as
zlj=Ml−1∑i=1wlijal−1i+blj,where blk are the biases from layer l. Here Ml−1 represents the total number of nodes/neurons/units of layer l−1. The figure in the whiteboard notes illustrates this equation. We can rewrite this in a more compact form as the matrix-vector products we discussed earlier,
zl=(Wl)Tal−1+bl.