Logo
Unionpedia
Communication
Get it on Google Play
New! Download Unionpedia on your Android™ device!
Free
Faster access than browser!
 

Dimensionality reduction

Index Dimensionality reduction

In statistics, machine learning, and information theory, dimensionality reduction or dimension reduction is the process of reducing the number of random variables under consideration by obtaining a set of principal variables. [1]

65 relations: Autoencoder, Backpropagation, Canonical correlation, Circumstellar disc, Combinatorial optimization, Correlation and dependence, Covariance, CUR matrix approximation, Curse of dimensionality, Data analysis, Data transformation (statistics), Diffusion map, Dimension, Eigenvalues and eigenvectors, Embedding, Feature (machine learning), Feature extraction, Feature selection, Hyperparameter optimization, Information gain in decision trees, Information theory, Isomap, Johnson–Lindenstrauss lemma, K-nearest neighbors algorithm, Kernel method, Kernel principal component analysis, Latent semantic analysis, Linear discriminant analysis, Local tangent space alignment, Locality-sensitive hashing, Machine learning, Matrix (mathematics), Maximally informative dimensions, Methods of detecting exoplanets, MinHash, MIT Press, Multidimensional scaling, Multifactor dimensionality reduction, Multilinear principal component analysis, Multilinear subspace learning, Mutual information, Nature (journal), Nearest neighbor search, Neural network, Neuroscience, Non-negative matrix factorization, Nonlinear dimensionality reduction, Principal component analysis, Random projection, Regression analysis, ..., Restricted Boltzmann machine, Sammon mapping, Semantic mapping (statistics), Semidefinite embedding, Semidefinite programming, Singular-value decomposition, Statistical classification, Statistics, Sufficient dimension reduction, T-distributed stochastic neighbor embedding, Tensor, Time series, Topological data analysis, VLDB, Weighted correlation network analysis. Expand index (15 more) »

Autoencoder

An autoencoder is a type of artificial neural network used to learn efficient data codings in an unsupervised manner.

New!!: Dimensionality reduction and Autoencoder · See more »

Backpropagation

Backpropagation is a method used in artificial neural networks to calculate a gradient that is needed in the calculation of the weights to be used in the network.

New!!: Dimensionality reduction and Backpropagation · See more »

Canonical correlation

In statistics, canonical-correlation analysis (CCA) is a way of inferring information from cross-covariance matrices.

New!!: Dimensionality reduction and Canonical correlation · See more »

Circumstellar disc

A circumstellar disc (or circumstellar disk) is a torus, pancake or ring-shaped accumulation of matter composed of gas, dust, planetesimals, asteroids or collision fragments in orbit around a star.

New!!: Dimensionality reduction and Circumstellar disc · See more »

Combinatorial optimization

In applied mathematics and theoretical computer science, combinatorial optimization is a topic that consists of finding an optimal object from a finite set of objects.

New!!: Dimensionality reduction and Combinatorial optimization · See more »

Correlation and dependence

In statistics, dependence or association is any statistical relationship, whether causal or not, between two random variables or bivariate data.

New!!: Dimensionality reduction and Correlation and dependence · See more »

Covariance

In probability theory and statistics, covariance is a measure of the joint variability of two random variables.

New!!: Dimensionality reduction and Covariance · See more »

CUR matrix approximation

A CUR matrix approximation is a set of three matrices that, when multiplied together, closely approximate a given matrix.

New!!: Dimensionality reduction and CUR matrix approximation · See more »

Curse of dimensionality

The curse of dimensionality refers to various phenomena that arise when analyzing and organizing data in high-dimensional spaces (often with hundreds or thousands of dimensions) that do not occur in low-dimensional settings such as the three-dimensional physical space of everyday experience.

New!!: Dimensionality reduction and Curse of dimensionality · See more »

Data analysis

Data analysis is a process of inspecting, cleansing, transforming, and modeling data with the goal of discovering useful information, informing conclusions, and supporting decision-making.

New!!: Dimensionality reduction and Data analysis · See more »

Data transformation (statistics)

In statistics, data transformation is the application of a deterministic mathematical function to each point in a data set — that is, each data point zi is replaced with the transformed value yi.

New!!: Dimensionality reduction and Data transformation (statistics) · See more »

Diffusion map

Diffusion maps is a dimensionality reduction or feature extraction algorithm introduced by Coifman and Lafon which computes a family of embeddings of a data set into Euclidean space (often low-dimensional) whose coordinates can be computed from the eigenvectors and eigenvalues of a diffusion operator on the data.

New!!: Dimensionality reduction and Diffusion map · See more »

Dimension

In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it.

New!!: Dimensionality reduction and Dimension · See more »

Eigenvalues and eigenvectors

In linear algebra, an eigenvector or characteristic vector of a linear transformation is a non-zero vector that changes by only a scalar factor when that linear transformation is applied to it.

New!!: Dimensionality reduction and Eigenvalues and eigenvectors · See more »

Embedding

In mathematics, an embedding (or imbedding) is one instance of some mathematical structure contained within another instance, such as a group that is a subgroup.

New!!: Dimensionality reduction and Embedding · See more »

Feature (machine learning)

In machine learning and pattern recognition, a feature is an individual measurable property or characteristic of a phenomenon being observed.

New!!: Dimensionality reduction and Feature (machine learning) · See more »

Feature extraction

In machine learning, pattern recognition and in image processing, feature extraction starts from an initial set of measured data and builds derived values (features) intended to be informative and non-redundant, facilitating the subsequent learning and generalization steps, and in some cases leading to better human interpretations.

New!!: Dimensionality reduction and Feature extraction · See more »

Feature selection

In machine learning and statistics, feature selection, also known as variable selection, attribute selection or variable subset selection, is the process of selecting a subset of relevant features (variables, predictors) for use in model construction.

New!!: Dimensionality reduction and Feature selection · See more »

Hyperparameter optimization

In machine learning, hyperparameter optimization or tuning is the problem of choosing a set of optimal hyperparameters for a learning algorithm.

New!!: Dimensionality reduction and Hyperparameter optimization · See more »

Information gain in decision trees

In information theory and machine learning, information gain is a synonym for Kullback–Leibler divergence.

New!!: Dimensionality reduction and Information gain in decision trees · See more »

Information theory

Information theory studies the quantification, storage, and communication of information.

New!!: Dimensionality reduction and Information theory · See more »

Isomap

Isomap is a nonlinear dimensionality reduction method.

New!!: Dimensionality reduction and Isomap · See more »

Johnson–Lindenstrauss lemma

In mathematics, the Johnson–Lindenstrauss lemma is a result named after William B. Johnson and Joram Lindenstrauss concerning low-distortion embeddings of points from high-dimensional into low-dimensional Euclidean space.

New!!: Dimensionality reduction and Johnson–Lindenstrauss lemma · See more »

K-nearest neighbors algorithm

In pattern recognition, the k-nearest neighbors algorithm (k-NN) is a non-parametric method used for classification and regression.

New!!: Dimensionality reduction and K-nearest neighbors algorithm · See more »

Kernel method

In machine learning, kernel methods are a class of algorithms for pattern analysis, whose best known member is the support vector machine (SVM).

New!!: Dimensionality reduction and Kernel method · See more »

Kernel principal component analysis

In the field of multivariate statistics, kernel principal component analysis (kernel PCA) is an extension of principal component analysis (PCA) using techniques of kernel methods.

New!!: Dimensionality reduction and Kernel principal component analysis · See more »

Latent semantic analysis

Latent semantic analysis (LSA) is a technique in natural language processing, in particular distributional semantics, of analyzing relationships between a set of documents and the terms they contain by producing a set of concepts related to the documents and terms.

New!!: Dimensionality reduction and Latent semantic analysis · See more »

Linear discriminant analysis

Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics, pattern recognition and machine learning to find a linear combination of features that characterizes or separates two or more classes of objects or events.

New!!: Dimensionality reduction and Linear discriminant analysis · See more »

Local tangent space alignment

Local tangent space alignment (LTSA) is a method for manifold learning, which can efficiently learn a nonlinear embedding into low-dimensional coordinates from high-dimensional data, and can also reconstruct high-dimensional coordinates from embedding coordinates.

New!!: Dimensionality reduction and Local tangent space alignment · See more »

Locality-sensitive hashing

Locality-sensitive hashing (LSH) reduces the dimensionality of high-dimensional data.

New!!: Dimensionality reduction and Locality-sensitive hashing · See more »

Machine learning

Machine learning is a subset of artificial intelligence in the field of computer science that often uses statistical techniques to give computers the ability to "learn" (i.e., progressively improve performance on a specific task) with data, without being explicitly programmed.

New!!: Dimensionality reduction and Machine learning · See more »

Matrix (mathematics)

In mathematics, a matrix (plural: matrices) is a rectangular array of numbers, symbols, or expressions, arranged in rows and columns.

New!!: Dimensionality reduction and Matrix (mathematics) · See more »

Maximally informative dimensions

Maximally informative dimensions is a dimensionality reduction technique used in the statistical analyses of neural responses.

New!!: Dimensionality reduction and Maximally informative dimensions · See more »

Methods of detecting exoplanets

Any planet is an extremely faint light source compared to its parent star.

New!!: Dimensionality reduction and Methods of detecting exoplanets · See more »

MinHash

In computer science and data mining, MinHash (or the min-wise independent permutations locality sensitive hashing scheme) is a technique for quickly estimating how similar two sets are.

New!!: Dimensionality reduction and MinHash · See more »

MIT Press

The MIT Press is a university press affiliated with the Massachusetts Institute of Technology (MIT) in Cambridge, Massachusetts (United States).

New!!: Dimensionality reduction and MIT Press · See more »

Multidimensional scaling

Multidimensional scaling (MDS) is a means of visualizing the level of similarity of individual cases of a dataset.

New!!: Dimensionality reduction and Multidimensional scaling · See more »

Multifactor dimensionality reduction

Multifactor dimensionality reduction (MDR) is a statistical approach, also used in machine learning automatic approaches, for detecting and characterizing combinations of attributes or independent variables that interact to influence a dependent or class variable.

New!!: Dimensionality reduction and Multifactor dimensionality reduction · See more »

Multilinear principal component analysis

Multilinear principal component analysis (MPCA) is a multilinear extension of principal component analysis (PCA).

New!!: Dimensionality reduction and Multilinear principal component analysis · See more »

Multilinear subspace learning

Multilinear subspace learning is an approach to dimensionality reduction.

New!!: Dimensionality reduction and Multilinear subspace learning · See more »

Mutual information

In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables.

New!!: Dimensionality reduction and Mutual information · See more »

Nature (journal)

Nature is a British multidisciplinary scientific journal, first published on 4 November 1869.

New!!: Dimensionality reduction and Nature (journal) · See more »

Nearest neighbor search

Nearest neighbor search (NNS), as a form of proximity search, is the optimization problem of finding the point in a given set that is closest (or most similar) to a given point.

New!!: Dimensionality reduction and Nearest neighbor search · See more »

Neural network

The term neural network was traditionally used to refer to a network or circuit of neurons.

New!!: Dimensionality reduction and Neural network · See more »

Neuroscience

Neuroscience (or neurobiology) is the scientific study of the nervous system.

New!!: Dimensionality reduction and Neuroscience · See more »

Non-negative matrix factorization

Non-negative matrix factorization (NMF or NNMF), also non-negative matrix approximation is a group of algorithms in multivariate analysis and linear algebra where a matrix is factorized into (usually) two matrices and, with the property that all three matrices have no negative elements.

New!!: Dimensionality reduction and Non-negative matrix factorization · See more »

Nonlinear dimensionality reduction

High-dimensional data, meaning data that requires more than two or three dimensions to represent, can be difficult to interpret.

New!!: Dimensionality reduction and Nonlinear dimensionality reduction · See more »

Principal component analysis

Principal component analysis (PCA) is a statistical procedure that uses an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of linearly uncorrelated variables called principal components.

New!!: Dimensionality reduction and Principal component analysis · See more »

Random projection

In mathematics and statistics, random projection is a technique used to reduce the dimensionality of a set of points which lie in Euclidean space.

New!!: Dimensionality reduction and Random projection · See more »

Regression analysis

In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships among variables.

New!!: Dimensionality reduction and Regression analysis · See more »

Restricted Boltzmann machine

A restricted Boltzmann machine (RBM) is a generative stochastic artificial neural network that can learn a probability distribution over its set of inputs.

New!!: Dimensionality reduction and Restricted Boltzmann machine · See more »

Sammon mapping

Sammon mapping or Sammon projection is an algorithm that maps a high-dimensional space to a space of lower dimensionality (see multidimensional scaling) by trying to preserve the structure of inter-point distances in high-dimensional space in the lower-dimension projection.

New!!: Dimensionality reduction and Sammon mapping · See more »

Semantic mapping (statistics)

Semantic mapping (SM) is a method in statistics for dimensionality reduction that can be used in a set of multidimensional vectors of features to extract a few new features that preserves the main data characteristics.

New!!: Dimensionality reduction and Semantic mapping (statistics) · See more »

Semidefinite embedding

Semidefinite embedding (SDE) or maximum variance unfolding (MVU) is an algorithm in computer science that uses semidefinite programming to perform non-linear dimensionality reduction of high-dimensional vectorial input data.

New!!: Dimensionality reduction and Semidefinite embedding · See more »

Semidefinite programming

Semidefinite programming (SDP) is a subfield of convex optimization concerned with the optimization of a linear objective function (a user-specified function that the user wants to minimize or maximize) over the intersection of the cone of positive semidefinite matrices with an affine space, i.e., a spectrahedron.

New!!: Dimensionality reduction and Semidefinite programming · See more »

Singular-value decomposition

In linear algebra, the singular-value decomposition (SVD) is a factorization of a real or complex matrix.

New!!: Dimensionality reduction and Singular-value decomposition · See more »

Statistical classification

In machine learning and statistics, classification is the problem of identifying to which of a set of categories (sub-populations) a new observation belongs, on the basis of a training set of data containing observations (or instances) whose category membership is known.

New!!: Dimensionality reduction and Statistical classification · See more »

Statistics

Statistics is a branch of mathematics dealing with the collection, analysis, interpretation, presentation, and organization of data.

New!!: Dimensionality reduction and Statistics · See more »

Sufficient dimension reduction

In statistics, sufficient dimension reduction (SDR) is a paradigm for analyzing data that combines the ideas of dimension reduction with the concept of sufficiency.

New!!: Dimensionality reduction and Sufficient dimension reduction · See more »

T-distributed stochastic neighbor embedding

T-distributed Stochastic Neighbor Embedding (t-SNE) is a machine learning algorithm for visualization developed by Laurens van der Maaten and Geoffrey Hinton.

New!!: Dimensionality reduction and T-distributed stochastic neighbor embedding · See more »

Tensor

In mathematics, tensors are geometric objects that describe linear relations between geometric vectors, scalars, and other tensors.

New!!: Dimensionality reduction and Tensor · See more »

Time series

A time series is a series of data points indexed (or listed or graphed) in time order.

New!!: Dimensionality reduction and Time series · See more »

Topological data analysis

In applied mathematics, topological data analysis (TDA) is an approach to the analysis of datasets using techniques from topology.

New!!: Dimensionality reduction and Topological data analysis · See more »

VLDB

VLDB is an annual conference held by the non-profit Very Large Data Base Endowment Inc. The mission of VLDB is to promote and exchange scholarly work in databases and related fields throughout the world.

New!!: Dimensionality reduction and VLDB · See more »

Weighted correlation network analysis

Weighted correlation network analysis, also known as weighted gene co-expression network analysis (WGCNA), is a widely used data mining method especially for studying biological networks based on pairwise correlations between variables.

New!!: Dimensionality reduction and Weighted correlation network analysis · See more »

Redirects here:

Dimension reduction, Dimensionality Reduction, Dimensionality reduction algorithm.

References

[1] https://en.wikipedia.org/wiki/Dimensionality_reduction

OutgoingIncoming
Hey! We are on Facebook now! »