Dimension Reduction

title: Dimension Reduction

Dimension Reduction

Dealing with a lot of dimensions can be painful for machine learning algorithms. High dimensionality will increase the computational complexity, increase the risk of overfitting (as your algorithm has more degrees of freedom) and the sparsity of the data will grow. Hence, dimensionality reduction will project the data in a space with less dimension to limit these phenomena.

Why is dimensionality reduction useful?

  • Projection into two dimensions is often used to facilitate the visualization of high dimensional data sets.
  • When the dimensions can be given a meaningful interpretation, projection along that dimension can be used to explain certain behaviors.
  • In the supervised learning case, dimensionality reduction can be used to reduce the dimension of the features, potentially leading to better performance for the learning algorithm.

Dimensionality Reduction Techniques

  • Linear Discriminant Analysis LDA
  • Principal Components Analysis PCA
  • Kernel PCA
  • Graph-based kernel PCA
  • t-Distributed Stochastic Neighbor Embedding t-SNE
  • Auto Encoders
  • Generalized discriminant analysis (GDA)
  • autoencoders

More Information:

This article needs improvement. You can help improve this article. You can also write similar articles and help the community.