How Deep Learning Decodes Faces Into Important Characteristics Like Age

Deep learning uncontrolled studies show that the brain divides faces into semantically important characteristics, such as age.

How Deep Learning Decodes Faces Into Important Characteristics Like Age

A new study in unsupervised deep learning shows that the brain classifies faces according to semantically important factors, such as age, at one level of a  neural network.

The Active Appearance Model (AAM) is essentially a built-in structure that does not follow common learning principles. Deep neural networks are a popular computational model in abdominal flow in monkeys in recent years. Unlike AAM, these models are domain independent and custom distributions are based on data-driven training. Modern deep networks are trained to perform multifaceted entity identification tasks using high-density training signals, generating multidimensional representations that closely mimic those found in biological systems.

Deep Classifiers

On the other hand, deep classifiers no longer account for the responses of individual neurons in monkey face spots and AAM. Also, the representative form of Deep Classifier and AAM is different. Whereas deep classifiers generate multidimensional multiplexed representations for many simulated neurons, AAM uses a low-dimensional coding scheme that contains orthogonal information in one dimension.

It is a long-standing belief that the visual system uses introspection to determine sensory input. Although these interpreted structures may seem deceptively simple to humans, they are difficult to reconstruct in practice due to the highly complex nonlinear transformation of the input signal at the pixel level.

Source: Unsupervised deep learning

Recent advances in machine learning have provided a blueprint for implementing this theory, suggesting generative models with deep self-control. For example, beta-variant autoencoders (VAEs) can reliably reconstruct sensory data  from low-dimensional embeddings while prompting individual network blocks to encode semantically important elements such as objects, colors, faces, gender, and scene layout. It is a model to learn.

Deep Generative Models

These deep generative models continue the  long legacy of the neuroscience community of generating autonomously controlled visual models, enabling powerful generalizations, imagination, abstract reasoning, constructive reasoning, and other characteristics of biological visual perception. This study aims to find out whether the overall learning goal can lead to coding similar to the coding used by real neurons. The results show that the VAE-optimized unwind target tests the mechanism by which  ventral visual flow produces visible low-dimensional face images.


Unlike previous studies, the results show that unsupervised deep learning can intelligently process these codes at the neural level. Furthermore, this study demonstrates that the mutational axis of an individual IT neuron corresponds to a single "unwinding" latent unit. Thus, this study expands on previous studies of the coding function of single neurons in the facial macular region of monkeys demonstrating a one-to-one correspondence between model units and neurons, rather than the small agreement previously reported.

For more information, refer to the article .


The raw responses of all models to the 2,162 face photos generated in this survey were pushed to the figshare repository. .

The following databases contain the face image data used in this study:

FERET face database ,

CVL face database ,

MR2 face database ,

PEAL face database


Due to the  complexity of the research and partial dependence on commercial libraries, the code supporting the results is available upon request by Irina Higgins ( Additionally, VAE models, concordance scores, and  UDR measures are  available as open source implementations at: .