74 relations: Aleksandr Gorban, Autoencoder, Backpropagation, Charles-Augustin de Coulomb, Concentration of measure, Coulomb's law, Curse of dimensionality, Degree matrix, Diffusion map, Dimension, Dimensionality reduction, Distance, Distance (graph theory), Distance matrix, Dynamic time warping, Dynamical system, Eigendecomposition of a matrix, Elastic map, Embedding, Euclidean distance, Expectation–maximization algorithm, Factor analysis, Feature extraction, Feature learning, Floyd–Warshall algorithm, Fourier series, Gaussian process, Generative topographic map, Geodesic, Global cascades model, Graduated optimization, Growing self-organizing map, Hamming space, Independent component analysis, Isomap, Journal of Machine Learning Research, K-nearest neighbors algorithm, Kernel method, Kernel principal component analysis, Klein bottle, Laplace–Beltrami operator, Latent variable model, Linear discriminant analysis, Local tangent space alignment, Machine learning, Manifold, Manifold alignment, Markov chain, MIT Press, Multidimensional scaling, ..., Neural network, Nonlinear dimensionality reduction, Ohio State University, Partha Niyogi, Principal component analysis, Projective space, Radial basis function network, Random walk, Reproducing kernel Hilbert space, Restricted Boltzmann machine, Sammon mapping, Self-organizing map, Semidefinite embedding, Similarity measure, Singular-value decomposition, Sparse matrix, Sphere, Stress majorization, Swiss roll, T-distributed stochastic neighbor embedding, Torus, Trevor Hastie, University of Chicago, Yale University. Expand index (24 more) » « Shrink index
Alexander Nikolaevich Gorban (Александр Николаевич Горба́нь.) is a scientist of Soviet origin, working in the United Kingdom.
An autoencoder is a type of artificial neural network used to learn efficient data codings in an unsupervised manner.
Backpropagation is a method used in artificial neural networks to calculate a gradient that is needed in the calculation of the weights to be used in the network.
Charles-Augustin de Coulomb (14 June 1736 – 23 August 1806) was a French military engineer and physicist.
In mathematics, concentration of measure (about a median) is a principle that is applied in measure theory, probability and combinatorics, and has consequences for other fields such as Banach space theory.
Coulomb's law, or Coulomb's inverse-square law, is a law of physics for quantifying the amount of force with which stationary electrically charged particles repel or attract each other.
The curse of dimensionality refers to various phenomena that arise when analyzing and organizing data in high-dimensional spaces (often with hundreds or thousands of dimensions) that do not occur in low-dimensional settings such as the three-dimensional physical space of everyday experience.
In the mathematical field of graph theory the degree matrix is a diagonal matrix which contains information about the degree of each vertex—that is, the number of edges attached to each vertex.
Diffusion maps is a dimensionality reduction or feature extraction algorithm introduced by Coifman and Lafon which computes a family of embeddings of a data set into Euclidean space (often low-dimensional) whose coordinates can be computed from the eigenvectors and eigenvalues of a diffusion operator on the data.
In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it.
In statistics, machine learning, and information theory, dimensionality reduction or dimension reduction is the process of reducing the number of random variables under consideration by obtaining a set of principal variables.
Distance is a numerical measurement of how far apart objects are.
In the mathematical field of graph theory, the distance between two vertices in a graph is the number of edges in a shortest path (also called a graph geodesic) connecting them.
In mathematics, computer science and especially graph theory, a distance matrix is a square matrix (two-dimensional array) containing the distances, taken pairwise, between the elements of a set.
In time series analysis, dynamic time warping (DTW) is one of the algorithms for measuring similarity between two temporal sequences, which may vary in speed.
In mathematics, a dynamical system is a system in which a function describes the time dependence of a point in a geometrical space.
In linear algebra, eigendecomposition or sometimes spectral decomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors.
Elastic maps provide a tool for nonlinear dimensionality reduction.
In mathematics, an embedding (or imbedding) is one instance of some mathematical structure contained within another instance, such as a group that is a subgroup.
In mathematics, the Euclidean distance or Euclidean metric is the "ordinary" straight-line distance between two points in Euclidean space.
In statistics, an expectation–maximization (EM) algorithm is an iterative method to find maximum likelihood or maximum a posteriori (MAP) estimates of parameters in statistical models, where the model depends on unobserved latent variables.
Factor analysis is a statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors.
In machine learning, pattern recognition and in image processing, feature extraction starts from an initial set of measured data and builds derived values (features) intended to be informative and non-redundant, facilitating the subsequent learning and generalization steps, and in some cases leading to better human interpretations.
In machine learning, feature learning or representation learning is a set of techniques that allows a system to automatically discover the representations needed for feature detection or classification from raw data.
In computer science, the Floyd–Warshall algorithm is an algorithm for finding shortest paths in a weighted graph with positive or negative edge weights (but with no negative cycles).
In mathematics, a Fourier series is a way to represent a function as the sum of simple sine waves.
In probability theory and statistics, a Gaussian process is a stochastic process (a collection of random variables indexed by time or space), such that every finite collection of those random variables has a multivariate normal distribution, i.e. every finite linear combination of them is normally distributed.
Generative topographic map (GTM) is a machine learning method that is a probabilistic counterpart of the self-organizing map (SOM), is probably convergent and does not require a shrinking neighborhood or a decreasing step size.
In differential geometry, a geodesic is a generalization of the notion of a "straight line" to "curved spaces".
Global cascades models are a class of models aiming to model large and rare cascades that are triggered by exogenous perturbations which are relatively small compared with the size of the system.
Graduated optimization is a global optimization technique that attempts to solve a difficult optimization problem by initially solving a greatly simplified problem, and progressively transforming that problem (while optimizing) until it is equivalent to the difficult optimization problem.
A growing self-organizing map (GSOM) is a growing variant of a self-organizing map (SOM).
In statistics and coding theory, a Hamming space is usually the set of all 2^N binary strings of length N. It is used in the theory of coding signals and transmission.
In signal processing, independent component analysis (ICA) is a computational method for separating a multivariate signal into additive subcomponents.
Isomap is a nonlinear dimensionality reduction method.
The Journal of Machine Learning Research is a peer-reviewed open access scientific journal covering machine learning.
In pattern recognition, the k-nearest neighbors algorithm (k-NN) is a non-parametric method used for classification and regression.
In machine learning, kernel methods are a class of algorithms for pattern analysis, whose best known member is the support vector machine (SVM).
In the field of multivariate statistics, kernel principal component analysis (kernel PCA) is an extension of principal component analysis (PCA) using techniques of kernel methods.
In topology, a branch of mathematics, the Klein bottle is an example of a non-orientable surface; it is a two-dimensional manifold against which a system for determining a normal vector cannot be consistently defined.
In differential geometry, the Laplace operator, named after Pierre-Simon Laplace, can be generalized to operate on functions defined on surfaces in Euclidean space and, more generally, on Riemannian and pseudo-Riemannian manifolds.
A latent variable model is a statistical model that relates a set of observable variables (so-called manifest variables) to a set of latent variables.
Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics, pattern recognition and machine learning to find a linear combination of features that characterizes or separates two or more classes of objects or events.
Local tangent space alignment (LTSA) is a method for manifold learning, which can efficiently learn a nonlinear embedding into low-dimensional coordinates from high-dimensional data, and can also reconstruct high-dimensional coordinates from embedding coordinates.
Machine learning is a subset of artificial intelligence in the field of computer science that often uses statistical techniques to give computers the ability to "learn" (i.e., progressively improve performance on a specific task) with data, without being explicitly programmed.
In mathematics, a manifold is a topological space that locally resembles Euclidean space near each point.
Manifold alignment is a class of machine learning algorithms that produce projections between sets of data, given that the original data sets lie on a common manifold.
A Markov chain is "a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event".
The MIT Press is a university press affiliated with the Massachusetts Institute of Technology (MIT) in Cambridge, Massachusetts (United States).
Multidimensional scaling (MDS) is a means of visualizing the level of similarity of individual cases of a dataset.
The term neural network was traditionally used to refer to a network or circuit of neurons.
High-dimensional data, meaning data that requires more than two or three dimensions to represent, can be difficult to interpret.
The Ohio State University, commonly referred to as Ohio State or OSU, is a large, primarily residential, public university in Columbus, Ohio.
Partha Niyogi (July 31, 1967 – October 1, 2010) was the Louis Block Professor in Computer Science and Statistics at the University of Chicago.
Principal component analysis (PCA) is a statistical procedure that uses an orthogonal transformation to convert a set of observations of possibly correlated variables into a set of values of linearly uncorrelated variables called principal components.
In mathematics, a projective space can be thought of as the set of lines through the origin of a vector space V. The cases when and are the real projective line and the real projective plane, respectively, where R denotes the field of real numbers, R2 denotes ordered pairs of real numbers, and R3 denotes ordered triplets of real numbers.
In the field of mathematical modeling, a radial basis function network is an artificial neural network that uses radial basis functions as activation functions.
A random walk is a mathematical object, known as a stochastic or random process, that describes a path that consists of a succession of random steps on some mathematical space such as the integers.
In functional analysis (a branch of mathematics), a reproducing kernel Hilbert space (RKHS) is a Hilbert space of functions in which point evaluation is a continuous linear functional.
A restricted Boltzmann machine (RBM) is a generative stochastic artificial neural network that can learn a probability distribution over its set of inputs.
Sammon mapping or Sammon projection is an algorithm that maps a high-dimensional space to a space of lower dimensionality (see multidimensional scaling) by trying to preserve the structure of inter-point distances in high-dimensional space in the lower-dimension projection.
A self-organizing map (SOM) or self-organizing feature map (SOFM) is a type of artificial neural network (ANN) that is trained using unsupervised learning to produce a low-dimensional (typically two-dimensional), discretized representation of the input space of the training samples, called a map, and is therefore a method to do dimensionality reduction.
Semidefinite embedding (SDE) or maximum variance unfolding (MVU) is an algorithm in computer science that uses semidefinite programming to perform non-linear dimensionality reduction of high-dimensional vectorial input data.
In statistics and related fields, a similarity measure or similarity function is a real-valued function that quantifies the similarity between two objects.
In linear algebra, the singular-value decomposition (SVD) is a factorization of a real or complex matrix.
In numerical analysis and computer science, a sparse matrix or sparse array is a matrix in which most of the elements are zero.
A sphere (from Greek σφαῖρα — sphaira, "globe, ball") is a perfectly round geometrical object in three-dimensional space that is the surface of a completely round ball (viz., analogous to the circular objects in two dimensions, where a "circle" circumscribes its "disk").
Stress majorization is an optimization strategy used in multidimensional scaling (MDS) where, for a set of n m-dimensional data items, a configuration X of n points in r(\sigma(X). Usually r is 2 or 3, i.e. the (n x r) matrix X lists points in 2- or 3-dimensional Euclidean space so that the result may be visualised (i.e. an MDS plot). The function \sigma is a cost or loss function that measures the squared differences between ideal (m-dimensional) distances and actual distances in r-dimensional space. It is defined as: where w_\ge 0 is a weight for the measurement between a pair of points (i,j), d_(X) is the euclidean distance between i and j and \delta_ is the ideal distance between the points (their separation) in the m-dimensional data space. Note that w_ can be used to specify a degree of confidence in the similarity between points (e.g. 0 can be specified if there is no information for a particular pair). A configuration X which minimizes \sigma(X) gives a plot in which points that are close together correspond to points that are also close together in the original m-dimensional data space. There are many ways that \sigma(X) could be minimized. For example, Kruskal recommended an iterative steepest descent approach. However, a significantly better (in terms of guarantees on, and rate of, convergence) method for minimizing stress was introduced by Jan de Leeuw.. De Leeuw's iterative majorization method at each step minimizes a simple convex function which both bounds \sigma from above and touches the surface of \sigma at a point Z, called the supporting point. In convex analysis such a function is called a majorizing function. This iterative majorization process is also referred to as the SMACOF algorithm ("Scaling by MAjorizing a COmplicated Function").
A Swiss roll, jelly roll, or cream roll is a type of sponge cake roll filled with whipped cream, jam, or icing.
T-distributed Stochastic Neighbor Embedding (t-SNE) is a machine learning algorithm for visualization developed by Laurens van der Maaten and Geoffrey Hinton.
In geometry, a torus (plural tori) is a surface of revolution generated by revolving a circle in three-dimensional space about an axis coplanar with the circle.
Trevor John Hastie (born 27 June 1953) is a South African and American statistician and computer scientist.
The University of Chicago (UChicago, U of C, or Chicago) is a private, non-profit research university in Chicago, Illinois.
Yale University is an American private Ivy League research university in New Haven, Connecticut.