π΄ Problem: MLPs ignore the structure of the data
Warning
Traditional neural networks (such as MLPs) require the input to be a flat vector (1D).
If the input data has dimensionality , this implies a flattening operation ().
The flattening step leads to the loss of information encoded in the structure of the input data (prior knowledge).
β MLPs completely ignore the dataβs structure.
case: Images πΌοΈ
In the case of images, the value of a pixel is typically correlated with the values of neighboring pixels.
This local correlation underlies a spatial structure that MLPs, due to input flattening, are unable to capture.
- βLocal correlations between pixels are lost β the spatial coherence inherent in the image disappears
- βThe spatial prior must be learned again: the layers downstream of the input layer have to (re)learn the structure of the data
- βMLPs cannot learn spatial patterns, which are characteristic of structured data such as images, audio, and time series

β Solution
The introduction of the Local Receptive Field concept makes it possible to preserve the spatial structure by observing sub-regions of the input.