码迷,mamicode.com
首页 > 其他好文 > 详细

Tikhonov regularization 吉洪诺夫 正则化

时间:2018-06-01 10:41:48      阅读:178      评论:0      收藏:0      [点我收藏+]

标签:html   product   llb   show   ble   The   strong   mat   XML   

        这个知识点很重要,但是,我不懂。

        第一个问题:为什么要做正则化?

In mathematics, statistics, and computer science, particularly in the fields of machine learning and inverse problems, regularization is a process of introducing additional information in order to solve an ill-posed problem or to prevent overfitting

And, what is ill-posed problem?... ...

And, what is overfitting? In statistics, overfitting is "the production of an analysis that corresponds too closely or exactly to a particular set of data, and may therefore fail to fit additional data or predict future observations reliably", as the next figure shows. 

技术分享图片

Figure 1.  The green curve represents an overfitted model and the black line represents a regularized model. While the green line best follows the training data, it is too dependent on that data and it is likely to have a higher error rate on new unseen data, compared to the black line.

        第二个问题:常用的正则化方法有哪些?

        第三个问题:The advantages fo Tikhonov regularizatioin

        The fourth question: Tikhonov regularization

        Tikhonov regularization, named for Andrey Tikhonov, is the most commonly used method of regularization of ill-posed problems. In statistics, the method is known as ridge regression, in machine learning it is known as weight decay, and with multiple independent discoveries, it is also variously known as the Tikhonov–Miller method, the Phillips–Twomey method, the constrained linear inversion method, and the method of linear regularization. It is related to the Levenberg–Marquardt algorithm for non-linear least-squares problems.

        Suppose that for a known matrix A and vector b, we wish to find a vector x such that:

        技术分享图片

        The standard approach is ordinary least squares linear regression. However, if no x satisfies the equation or more than one x does—that is, the solution is not unique—the problem is said to be ill posed. In such cases, ordinary least squares estimation leads to an overdetermined (over-fitted), or more often an underdetermined (under-fitted) system of equations. Most real-world phenomena have the effect of low-pass filters in the forward direction where A maps x to b. Therefore, in solving the inverse-problem, the inverse mapping operates as a high-pass filter that has the undesirable tendency of amplifying noise (eigenvalues / singular values are largest in the reverse mapping where they were smallest in the forward mapping). In addition, ordinary least squares implicitly nullifies every element of the reconstructed version of x that is in the null-space of A, rather than allowing for a model to be used as a prior for 技术分享图片. Ordinary least squares seeks to minimize the sum of squared residuals, which can be compactly written as:

        技术分享图片  where 技术分享图片 is the Euclidean norm

                                             技术分享图片

         In order to give preference to a particular solution with desirable properties, a regularization term can be included in this minimization:

        技术分享图片 for some suitably chosen Tikhonov matrix技术分享图片. In many cases, this matrix is chosen as a multiple of the identity matrix (技术分享图片), giving preference to solutions with smaller norms; this is known as L2 regularization.[1] In other cases, high-pass operators (e.g., a difference operator or a weighted Fourier operator) may be used to enforce smoothness if the underlying vector is believed to be mostly continuous. This regularization improves the conditioning of the problem, thus enabling a direct numerical solution. An explicit solution, denoted by 技术分享图片, is given by:

技术分享图片

The effect of regularization may be varied via the scale of matrix 技术分享图片. For 技术分享图片 this reduces to the unregularized least squares solution provided that (ATA)?1 exists.

L2 regularization is used in many contexts aside from linear regression, such as classification with logistic regression or support vector machines,[2] and matrix factorization.[3]

     

对于y=Xw,若X无解或有多个解,称这个问题是病态的。病态问题下,用最小二乘法求解会导致过拟合或欠拟合,用正则化来解决。

设X为m乘n矩阵:

  • 过拟合模型:m<<nm<<n,欠定方程,存在多解的可能性大;
  • 欠拟合模型:m>>nm>>n,超定方程,可能无解,或者有解但准确率很低

 技术分享图片

 

REF:

https://blog.csdn.net/darknightt/article/details/70179848

Tikhonov regularization 吉洪诺夫 正则化

标签:html   product   llb   show   ble   The   strong   mat   XML   

原文地址:https://www.cnblogs.com/shyang09/p/9120007.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!