Back to Glossary
Machine Learning

Disentangled representation

A disentangled representation is a way of structuring data representations such that individual factors of variation in the data are captured by distinct and independent dimensions in the representation space. Ideally, each dimension corresponds to a single, interpretable feature or attribute of the data.

Explanation

The goal of disentangled representation learning is to create representations where changes in one dimension of the representation space correspond to changes in only one underlying factor of variation in the data. For example, in a dataset of faces, a disentangled representation might have one dimension that controls the pose (rotation), another the lighting, and another the identity of the person. This contrasts with entangled representations where multiple factors are intertwined across dimensions, making it difficult to isolate and manipulate individual attributes. Methods for learning disentangled representations often involve specialized training objectives and architectures, such as variational autoencoders (VAEs) with specific regularization terms that encourage independence or sparsity in the learned representations. The benefit of disentangled representations is improved interpretability, control, and generalization in downstream tasks. They enable targeted manipulation of specific attributes, facilitate transfer learning by isolating relevant factors, and can lead to more robust and explainable AI systems.

Related Terms