Here is a fun visual for activation functions and hidden layers.
It is too computationally expensive to train neural networks on vectorized images. Instead we have to use convolution.
Convolution works by sliding a kernel over an image. For each neuron, it learns its own filter (kernel) and convolve it with the image. The result of this convolution process is a feature map.
This is known as a convolution neural network. We decide how many filters and layers to train.
This original convolution function is transformed to
where layer number, kernel size, # of channels (input) or filters (depth)
If you are interested in learning more about neural networks, I recommend reading my article on neural network layers!