1. PCA principal components analysis
主要是通过对协方差矩阵Covariance matrix进行特征分解,以得出数据的主成分(即特征向量eigenvector)与它们的权值(即特征值eigenvalue)。
PCA是最简单的以特征量分析多元统计分布的方法。其结果可以理解为对原数据中的方差variance做出解释:哪一个方向上的数据值对方差的影响最大?换而言之,PCA提供了一种降低数据维度的有效办法;如果分析者在原数据中除掉最小的特征值所对应的成分,那么所得的低维度数据必定是最优化的(也即,这样降低维度必定是失去讯息最少的方法)。
2. kmeans
The kernel k-means problem is an extension of the k-means problem where the input data points are mapped non-linearly into a higher-dimensional feature space via a kernel function $k(x_i,x_j) = \phi^T(x_i)\phi(x_j)$.
3. bayes
4. spectral clustering
In multivariate statistics and the clustering of data, spectral clustering techniques make use of the spectrum (eigenvalues) of the similarity matrix of the data to perform dimensionality reduction before clustering in fewer dimensions. The similarity matrix is provided as an input and consists of a quantitative assessment of the relative similarity of each pair of points in the dataset.
广义上来说,任何在演算法中用到SVD/特征值分解的,都叫Spectral Algorithm。 从很老很老的PCA/LDA,到比较近的Spectral Embedding/Clustering,都属于这类。
它的思想就是将聚类和图划分等同起来。
就是计算Laplacian matrix的算法不一样。
Given a simple graph G with n vertices, its Laplacian matrix $L:=(\ell_{i,j})_{n \times n}$ is defined as:
$L = D - A.$
That is, it is the difference of the degree matrix D and the adjacency matrix A of the graph. In the case of directed graphs, either the indegree or outdegree might be used, depending on the application.
5. svm
6. EM
To be learned:
6. deep learning
7. Spark
Classical method of machine learning,布布扣,bubuko.com
Classical method of machine learning
原文地址:http://www.cnblogs.com/linyx/p/3854949.html