In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. In the Bayesian setting, the term MMSE more specifically refers to estimation with quadratic loss function. In such case, the MMSE estimator is given by the posterior mean of the parameter to be estimated. Since the posterior mean is cumbersome to calculate, the form of the MMSE estimator is usually constrained to be within a certain class of functions. Linear MMSE estimators are a popular choice since they are easy to use, easy to calculate, and very versatile. It has given rise to many popular estimators such as the Wiener–Kolmogorov filter and Kalman filter.
Motivation
The term MMSE more specifically refers to estimation in a Bayesian setting with quadratic cost function. The basic idea behind the Bayesian approach to estimation stems from practical situations where we often have some prior information about the parameter to be estimated. For instance, we may have prior information about the range that the parameter can assume; or we may have an old estimate of the parameter that we want to modify when a new observation is made available; or the statistics of an actual random signal such as speech. This is in contrast to the non-Bayesian approach like minimum-variance unbiased estimator (MVUE) where absolutely nothing is assumed to be known about the parameter in advance and which does not account for such situations. In the Bayesian approach, such prior information is captured by the prior probability density function of the parameters; and based directly on Bayes' theorem, it allows us to make better posterior estimates as more observations become available. Thus unlike non-Bayesian approach where parameters of interest are assumed to be deterministic, but unknown constants, the Bayesian estimator seeks to estimate a parameter that is itself a random variable. Furthermore, Bayesian estimation can also deal with situations where the sequence of observations are not necessarily independent. Thus Bayesian estimation provides yet another alternative to the MVUE. This is useful when the MVUE does not exist or cannot be found.
Definition
Let  be a
 be a  hidden random vector variable, and let
 hidden random vector variable, and let  be a
 be a  known random vector variable (the measurement or observation), both of them not necessarily of the same dimension. An estimator
 known random vector variable (the measurement or observation), both of them not necessarily of the same dimension. An estimator  of
 of  is any function of the measurement
 is any function of the measurement  . The estimation error vector is given by
. The estimation error vector is given by  and its mean squared error (MSE) is given by the trace of error covariance matrix
 and its mean squared error (MSE) is given by the trace of error covariance matrix
 
where the expectation  is taken over
 is taken over  conditioned on
 conditioned on  . When
. When  is a scalar variable, the MSE expression simplifies to
 is a scalar variable, the MSE expression simplifies to  . Note that MSE can equivalently be defined in other ways, since
. Note that MSE can equivalently be defined in other ways, since
 
The MMSE estimator is then defined as the estimator achieving minimal MSE:
 
Properties
- When the means and variances are finite, the MMSE estimator is uniquely defined[1] and is given by:
 
 
- In other words, the MMSE estimator is the conditional expectation of  given the known observed value of the measurements. Also, since given the known observed value of the measurements. Also, since is the posterior mean, the error covariance matrix is the posterior mean, the error covariance matrix is equal to the posterior covariance is equal to the posterior covariance matrix, matrix, . .
 
- The MMSE estimator is unbiased (under the regularity assumptions mentioned above):
 
 
 
 
- where  is the Fisher information of is the Fisher information of . Thus, the MMSE estimator is asymptotically efficient. . Thus, the MMSE estimator is asymptotically efficient.
- The orthogonality principle: When  is a scalar, an estimator constrained to be of certain form is a scalar, an estimator constrained to be of certain form is an optimal estimator, i.e. is an optimal estimator, i.e. if and only if if and only if
 
 
- for all  in closed, linear subspace in closed, linear subspace of the measurements. For random vectors, since the MSE for estimation of a random vector is the sum of the MSEs of the coordinates, finding the MMSE estimator of a random vector decomposes into finding the MMSE estimators of the coordinates of X separately: of the measurements. For random vectors, since the MSE for estimation of a random vector is the sum of the MSEs of the coordinates, finding the MMSE estimator of a random vector decomposes into finding the MMSE estimators of the coordinates of X separately: 
 
- for all i and j. More succinctly put, the cross-correlation between the minimum estimation error  and the estimator and the estimator should be zero, should be zero, 
 
- If  and and are jointly Gaussian, then the MMSE estimator is linear, i.e., it has the form are jointly Gaussian, then the MMSE estimator is linear, i.e., it has the form for matrix for matrix and constant and constant . This can be directly shown using the Bayes' theorem. As a consequence, to find the MMSE estimator, it is sufficient to find the linear MMSE estimator. . This can be directly shown using the Bayes' theorem. As a consequence, to find the MMSE estimator, it is sufficient to find the linear MMSE estimator.
Linear MMSE estimator
In many cases, it is not possible to determine the analytical expression of the MMSE estimator. Two basic numerical approaches to obtain the MMSE estimate depends on either finding the conditional expectation  or finding the minima of MSE. Direct numerical evaluation of the conditional expectation is computationally expensive since it often requires multidimensional integration usually done via Monte Carlo methods. Another computational approach is to directly seek the minima of the MSE using techniques such as the stochastic gradient descent methods; but this method still requires the evaluation of expectation. While these numerical methods have been fruitful, a closed form expression for the MMSE estimator is nevertheless possible if we are willing to make some compromises.
 or finding the minima of MSE. Direct numerical evaluation of the conditional expectation is computationally expensive since it often requires multidimensional integration usually done via Monte Carlo methods. Another computational approach is to directly seek the minima of the MSE using techniques such as the stochastic gradient descent methods; but this method still requires the evaluation of expectation. While these numerical methods have been fruitful, a closed form expression for the MMSE estimator is nevertheless possible if we are willing to make some compromises.
One possibility is to abandon the full optimality requirements and seek a technique minimizing the MSE within a particular class of estimators, such as the class of linear estimators. Thus, we postulate that the conditional expectation of  given
 given  is a simple linear function of
 is a simple linear function of  ,
,  , where the measurement
, where the measurement  is a random vector,
 is a random vector,  is a matrix and
 is a matrix and  is a vector. This can be seen as the first order Taylor approximation of
 is a vector. This can be seen as the first order Taylor approximation of  . The linear MMSE estimator is the estimator achieving minimum MSE among all estimators of such form. That is, it solves the following optimization problem:
. The linear MMSE estimator is the estimator achieving minimum MSE among all estimators of such form. That is, it solves the following optimization problem: 
 
One advantage of such linear MMSE estimator is that it is not necessary to explicitly calculate the posterior probability density function of  . Such linear estimator only depends on the first two moments of
. Such linear estimator only depends on the first two moments of  and
 and  . So although it may be convenient to assume that
. So although it may be convenient to assume that  and
 and  are jointly Gaussian, it is not necessary to make this assumption, so long as the assumed distribution has well defined first and second moments. The form of the linear estimator does not depend on the type of the assumed underlying distribution.
 are jointly Gaussian, it is not necessary to make this assumption, so long as the assumed distribution has well defined first and second moments. The form of the linear estimator does not depend on the type of the assumed underlying distribution.
The expression for optimal  and
 and  is given by:
 is given by: 
 
 
where  ,
,  the
 the  is cross-covariance matrix between
 is cross-covariance matrix between  and
 and  , the
, the  is auto-covariance matrix of
 is auto-covariance matrix of  .
.
Thus, the expression for linear MMSE estimator, its mean, and its auto-covariance is given by
 
 
 
where the  is cross-covariance matrix between
 is cross-covariance matrix between  and
 and  .
.
Lastly, the error covariance and minimum mean square error achievable by such estimator is
 
 
Derivation using orthogonality principle
Let us have the optimal linear MMSE estimator given as  , where we are required to find the expression for
, where we are required to find the expression for  and
 and  . It is required that the MMSE estimator be unbiased. This means,
. It is required that the MMSE estimator be unbiased. This means,
 
Plugging the expression for  in above, we get
 in above, we get
 
where  and
 and  . Thus we can re-write the estimator as
. Thus we can re-write the estimator as
 
and the expression for estimation error becomes
 
From the orthogonality principle, we can have  , where we take
, where we take  . Here the left-hand-side term is
. Here the left-hand-side term is
 
When equated to zero, we obtain the desired expression for  as
 as
 
The  is cross-covariance matrix between X and Y, and
 is cross-covariance matrix between X and Y, and  is auto-covariance matrix of Y. Since
 is auto-covariance matrix of Y. Since  , the expression can also be re-written in terms of
, the expression can also be re-written in terms of  as
 as
 
Thus the full expression for the linear MMSE estimator is
 
Since the estimate  is itself a random variable with
 is itself a random variable with  , we can also obtain its auto-covariance as
, we can also obtain its auto-covariance as
 
Putting the expression for  and
 and  , we get
, we get
 
Lastly, the covariance of linear MMSE estimation error will then be given by
 
The first term in the third line is zero due to the orthogonality principle. Since  , we can re-write
, we can re-write  in terms of covariance matrices as
 in terms of covariance matrices as
 
This we can recognize to be the same as  Thus the minimum mean square error achievable by such a linear estimator is
 Thus the minimum mean square error achievable by such a linear estimator is
 . .
 Univariate case
For the special case when both  and
 and  are scalars, the above relations simplify to
 are scalars, the above relations simplify to
 
 
where  is the Pearson's correlation coefficient between
 is the Pearson's correlation coefficient between  and
 and  .
. 
The above two equations allows us to interpret the correlation coefficient either as normalized slope of linear regression 
 
or as square root of the ratio of two variances 
 . .
When  , we have
, we have  and
 and  . In this case, no new information is gleaned from the measurement which can decrease the uncertainty in
. In this case, no new information is gleaned from the measurement which can decrease the uncertainty in  . On the other hand, when
. On the other hand, when  , we have
, we have  and
 and  . Here
. Here  is completely determined by
 is completely determined by  , as given by the equation of straight line.
, as given by the equation of straight line.
Computation
Standard method like Gauss elimination can be used to solve the matrix equation for  . A more numerically stable method is provided by QR decomposition method. Since the matrix
. A more numerically stable method is provided by QR decomposition method. Since the matrix  is a symmetric positive definite matrix,
 is a symmetric positive definite matrix,  can be solved twice as fast with the Cholesky decomposition, while for large sparse systems conjugate gradient method is more effective. Levinson recursion is a fast method when
 can be solved twice as fast with the Cholesky decomposition, while for large sparse systems conjugate gradient method is more effective. Levinson recursion is a fast method when  is also a Toeplitz matrix. This can happen when
 is also a Toeplitz matrix. This can happen when  is a wide sense stationary process. In such stationary cases, these estimators are also referred to as Wiener–Kolmogorov filters.
 is a wide sense stationary process. In such stationary cases, these estimators are also referred to as Wiener–Kolmogorov filters.
Linear MMSE estimator for linear observation process
Let us further model the underlying process of observation as a linear process:  , where
, where  is a known matrix and
 is a known matrix and  is random noise vector with the mean
 is random noise vector with the mean  and cross-covariance
 and cross-covariance  . Here the required mean and the covariance matrices will be
. Here the required mean and the covariance matrices will be
 
 
 
Thus the expression for the linear MMSE estimator matrix  further modifies to
 further modifies to
 
Putting everything into the expression for  , we get
, we get
 
Lastly, the error covariance is
 
The significant difference between the estimation problem treated above and those of least squares and Gauss–Markov estimate is that the number of observations m, (i.e. the dimension of  ) need not be at least as large as the number of unknowns, n, (i.e. the dimension of
) need not be at least as large as the number of unknowns, n, (i.e. the dimension of  ). The estimate for the linear observation process exists so long as the m-by-m matrix
). The estimate for the linear observation process exists so long as the m-by-m matrix  exists; this is the case for any m if, for instance,
 exists; this is the case for any m if, for instance,  is positive definite. Physically the reason for this property is that since
 is positive definite. Physically the reason for this property is that since  is now a random variable, it is possible to form a meaningful estimate (namely its mean) even with no measurements. Every new measurement simply provides additional information which may modify our original estimate. Another feature of this estimate is that for m < n, there need be no measurement error. Thus, we may have
 is now a random variable, it is possible to form a meaningful estimate (namely its mean) even with no measurements. Every new measurement simply provides additional information which may modify our original estimate. Another feature of this estimate is that for m < n, there need be no measurement error. Thus, we may have  , because as long as
, because as long as  is positive definite, the estimate still exists. Lastly, this technique can handle cases where the noise is correlated.
 is positive definite, the estimate still exists. Lastly, this technique can handle cases where the noise is correlated.
An alternative form of expression can be obtained by using the matrix identity
 
which can be established by post-multiplying by  and pre-multiplying by
 and pre-multiplying by  to obtain
 to obtain
 
and
 
Since  can now be written in terms of
 can now be written in terms of  as
 as  , we get a simplified expression for
, we get a simplified expression for  as
 as
 
In this form the above expression can be easily compared with ridge regression, weighed least square and Gauss–Markov estimate. In particular, when  , corresponding to infinite variance of the apriori information concerning
, corresponding to infinite variance of the apriori information concerning  , the result
, the result  is identical to the weighed linear least square estimate with
 is identical to the weighed linear least square estimate with  as the weight matrix. Moreover, if the components of
 as the weight matrix. Moreover, if the components of  are uncorrelated and have equal variance such that
 are uncorrelated and have equal variance such that  where
 where  is an identity matrix, then
 is an identity matrix, then  is identical to the ordinary least square estimate. When apriori information is available as
 is identical to the ordinary least square estimate. When apriori information is available as  and the
 and the  are uncorrelated and have equal variance, we have
 are uncorrelated and have equal variance, we have  , which is identical to ridge regression solution.
, which is identical to ridge regression solution.
Sequential linear MMSE estimation
In many real-time applications, observational data is not available in a single batch. Instead the observations are made in a sequence. One possible approach is to use the sequential observations to update an old estimate as additional data becomes available, leading to finer estimates. One crucial difference between batch estimation and sequential estimation is that sequential estimation requires an additional Markov assumption. 
In the Bayesian framework, such recursive estimation is easily facilitated using Bayes' rule. Given  observations,
 observations,  , Bayes' rule gives us the posterior density of
, Bayes' rule gives us the posterior density of  as
 as
 
The  is called the posterior density,
 is called the posterior density,  is called the likelihood function, and
 is called the likelihood function, and  is the prior density of k-th time step. Here we have assumed the conditional independence of
 is the prior density of k-th time step. Here we have assumed the conditional independence of  from previous observations
 from previous observations  given
 given  as
 as 
 
This is the Markov assumption. 
The MMSE estimate  given the k-th observation is then the mean of the posterior density
 given the k-th observation is then the mean of the posterior density  . With the lack of dynamical information on how the state
. With the lack of dynamical information on how the state  changes with time, we will make a further stationarity assumption about the prior:
 changes with time, we will make a further stationarity assumption about the prior:
 
Thus, the prior density for k-th time step is the posterior density of (k-1)-th time step. This structure allows us to formulate a recursive approach to estimation.
In the context of linear MMSE estimator, the formula for the estimate will have the same form as before:  However, the mean and covariance matrices of
 However, the mean and covariance matrices of  and
 and  will need to be replaced by those of the prior density
 will need to be replaced by those of the prior density  and likelihood
 and likelihood  , respectively.
, respectively.
For the prior density  , its mean is given by the previous MMSE estimate,
, its mean is given by the previous MMSE estimate, 
![{\displaystyle {\bar {x}}_{k}=\mathrm {E} [x_{k}|y_{1},\ldots ,y_{k-1}]=\mathrm {E} [x_{k-1}|y_{1},\ldots ,y_{k-1}]={\hat {x}}_{k-1}}](./_assets_/831099b32c20d3fd925ced1a7bdb9246af3b2cbf.svg) , ,
and its covariance matrix is given by the previous error covariance matrix, 
 
as per by the properties of MMSE estimators and the stationarity assumption. 
Similarly, for the linear observation process, the mean of the likelihood  is given by
 is given by  and the covariance matrix is as before
 and the covariance matrix is as before
 . .
The difference between the predicted value of  , as given by
, as given by  , and its observed value
, and its observed value  gives the prediction error
 gives the prediction error  , which is also referred to as innovation or residual. It is more convenient to represent the linear MMSE in terms of the prediction error, whose mean and covariance are
, which is also referred to as innovation or residual. It is more convenient to represent the linear MMSE in terms of the prediction error, whose mean and covariance are ![{\displaystyle \mathrm {E} [{\tilde {y}}_{k}]=0}](./_assets_/b7333e06de208be900c7d618273db57bf678fb1f.svg) and
 and  .
.
Hence, in the estimate update formula, we should replace  and
 and  by
 by  and
 and  , respectively. Also, we should replace
, respectively. Also, we should replace  and
 and  by
 by  and
 and  . Lastly, we replace
. Lastly, we replace  by
 by 
 
Thus, we have the new estimate as new observation  arrives as
 arrives as
 
and the new error covariance as 
 
From the point of view of linear algebra, for sequential estimation, if we have an estimate  based on measurements generating space
 based on measurements generating space  , then after receiving another set of measurements, we should subtract out from these measurements that part that could be anticipated from the result of the first measurements. In other words, the updating must be based on that part of the new data which is orthogonal to the old data.
, then after receiving another set of measurements, we should subtract out from these measurements that part that could be anticipated from the result of the first measurements. In other words, the updating must be based on that part of the new data which is orthogonal to the old data.
The repeated use of the above two equations as more observations become available lead to recursive estimation techniques. The expressions can be more compactly written as
 
 
 
The matrix  is often referred to as the Kalman gain factor. The alternative formulation of the above algorithm will give
 is often referred to as the Kalman gain factor. The alternative formulation of the above algorithm will give
 
 
 
The repetition of these three steps as more data becomes available leads to an iterative estimation algorithm. The generalization of this idea to non-stationary cases gives rise to the Kalman filter. The three update steps outlined above indeed form the update step of the Kalman filter.
Special case: scalar observations
As an important special case, an easy to use recursive expression can be derived when at each k-th time instant the underlying linear observation process yields a scalar such that  , where
, where  is n-by-1 known column vector whose values can change with time,
 is n-by-1 known column vector whose values can change with time,  is n-by-1 random column vector to be estimated, and
 is n-by-1 random column vector to be estimated, and  is scalar noise term with variance
 is scalar noise term with variance  . After (k+1)-th observation, the direct use of above recursive equations give the expression for the estimate
. After (k+1)-th observation, the direct use of above recursive equations give the expression for the estimate  as:
 as:
 
where  is the new scalar observation and the gain factor
 is the new scalar observation and the gain factor  is n-by-1 column vector given by
 is n-by-1 column vector given by 
 
The  is n-by-n error covariance matrix given by
 is n-by-n error covariance matrix given by
 
Here, no matrix inversion is required. Also, the gain factor,  , depends on our confidence in the new data sample, as measured by the noise variance, versus that in the previous data. The initial values of
, depends on our confidence in the new data sample, as measured by the noise variance, versus that in the previous data. The initial values of  and
 and  are taken to be the mean and covariance of the aprior probability density function of
 are taken to be the mean and covariance of the aprior probability density function of  .
.
Alternative approaches: This important special case has also given rise to many other iterative methods (or adaptive filters), such as the least mean squares filter and recursive least squares filter, that directly solves the original MSE optimization problem using stochastic gradient descents. However, since the estimation error  cannot be directly observed, these methods try to minimize the mean squared prediction error
 cannot be directly observed, these methods try to minimize the mean squared prediction error  . For instance, in the case of scalar observations, we have the gradient
. For instance, in the case of scalar observations, we have the gradient  Thus, the update equation for the least mean square filter is given by
 Thus, the update equation for the least mean square filter is given by
 
where  is the scalar step size and the expectation is approximated by the instantaneous value
 is the scalar step size and the expectation is approximated by the instantaneous value  . As we can see, these methods bypass the need for covariance matrices.
. As we can see, these methods bypass the need for covariance matrices.
In many practical applications, the observation noise is uncorrelated. That is,  is a diagonal matrix. In such cases, it is advantageous to consider the components of
 is a diagonal matrix. In such cases, it is advantageous to consider the components of  as independent scalar measurements, rather than vector measurement. This allows us to reduce computation time by processing the
 as independent scalar measurements, rather than vector measurement. This allows us to reduce computation time by processing the  measurement vector as
 measurement vector as  scalar measurements. The use of scalar update formula avoids matrix inversion in the implementation of the covariance update equations, thus improving the numerical robustness against roundoff errors. The update can be implemented iteratively as:
 scalar measurements. The use of scalar update formula avoids matrix inversion in the implementation of the covariance update equations, thus improving the numerical robustness against roundoff errors. The update can be implemented iteratively as:
 
 
 
where  , using the initial values
, using the initial values  and
 and  . The intermediate variables
. The intermediate variables  is the
 is the  -th diagonal element of the
-th diagonal element of the  diagonal matrix
 diagonal matrix  ; while
; while  is the
 is the  -th row of
-th row of  matrix
 matrix  . The final values are
. The final values are  and
 and  .
.
Examples
Example 1
We shall take a linear prediction problem as an example. Let a linear combination of observed scalar random variables   and
 and  be used to estimate another future scalar random variable
 be used to estimate another future scalar random variable  such that
 such that  . If the random variables
. If the random variables ![{\displaystyle z=[z_{1},z_{2},z_{3},z_{4}]^{T}}](./_assets_/45cb1f9123fdf786074088616e42dfbcc1359d07.svg) are real Gaussian random variables with zero mean and its covariance matrix given by
 are real Gaussian random variables with zero mean and its covariance matrix given by
![{\displaystyle \operatorname {cov} (Z)=\operatorname {E} [zz^{T}]=\left[{\begin{array}{cccc}1&2&3&4\\2&5&8&9\\3&8&6&10\\4&9&10&15\end{array}}\right],}](./_assets_/90889a9de7192d317decb8381ca25cda49892b5c.svg) 
then our task is to find the coefficients  such that it will yield an optimal linear estimate
 such that it will yield an optimal linear estimate  .
.
In terms of the terminology developed in the previous sections, for this problem we have the observation vector ![{\displaystyle y=[z_{1},z_{2},z_{3}]^{T}}](./_assets_/7fdeb8eb600c7a066561d28e2b3f32fb5b1572b9.svg) , the estimator matrix
, the estimator matrix ![{\displaystyle W=[w_{1},w_{2},w_{3}]}](./_assets_/dd8e9343b228a9044dbb8208e9ceeb31270b04c1.svg) as a row vector, and the estimated variable
 as a row vector, and the estimated variable  as a scalar quantity. The autocorrelation matrix
 as a scalar quantity. The autocorrelation matrix  is defined as
 is defined as
![{\displaystyle C_{Y}=\left[{\begin{array}{ccc}E[z_{1},z_{1}]&E[z_{2},z_{1}]&E[z_{3},z_{1}]\\E[z_{1},z_{2}]&E[z_{2},z_{2}]&E[z_{3},z_{2}]\\E[z_{1},z_{3}]&E[z_{2},z_{3}]&E[z_{3},z_{3}]\end{array}}\right]=\left[{\begin{array}{ccc}1&2&3\\2&5&8\\3&8&6\end{array}}\right].}](./_assets_/50c0831c4af6cc131c13cd4da73e21ee08aac40c.svg) 
The cross correlation matrix  is defined as
 is defined as
![{\displaystyle C_{YX}=\left[{\begin{array}{c}E[z_{4},z_{1}]\\E[z_{4},z_{2}]\\E[z_{4},z_{3}]\end{array}}\right]=\left[{\begin{array}{c}4\\9\\10\end{array}}\right].}](./_assets_/e0b4e00619f85de1c0967817d68b6b154557af03.svg) 
We now solve the equation  by inverting
 by inverting  and pre-multiplying to get
 and pre-multiplying to get
![{\displaystyle C_{Y}^{-1}C_{YX}=\left[{\begin{array}{ccc}4.85&-1.71&-0.142\\-1.71&0.428&0.2857\\-0.142&0.2857&-0.1429\end{array}}\right]\left[{\begin{array}{c}4\\9\\10\end{array}}\right]=\left[{\begin{array}{c}2.57\\-0.142\\0.5714\end{array}}\right]=W^{T}.}](./_assets_/8f29d9505d5f630feae8a4566c23bae8dd82fd5d.svg) 
So we have  
  and
 and  as the optimal coefficients for
as the optimal coefficients for  . Computing the minimum
mean square error then gives
. Computing the minimum
mean square error then gives ![{\displaystyle \left\Vert e\right\Vert _{\min }^{2}=\operatorname {E} [z_{4}z_{4}]-WC_{YX}=15-WC_{YX}=.2857}](./_assets_/44558ab7a0b5ead1b9571853096d800773b38877.svg) .[2] Note that it is not necessary to obtain an explicit matrix inverse of
.[2] Note that it is not necessary to obtain an explicit matrix inverse of  to compute the value of
 to compute the value of  . The matrix equation can be solved by well known methods such as Gauss elimination method. A shorter, non-numerical example can be found in orthogonality principle.
. The matrix equation can be solved by well known methods such as Gauss elimination method. A shorter, non-numerical example can be found in orthogonality principle.
Example 2
Consider a vector  formed by taking
 formed by taking  observations of a fixed but unknown scalar parameter
 observations of a fixed but unknown scalar parameter  disturbed by white Gaussian noise. We can describe the process by a linear equation
 disturbed by white Gaussian noise. We can describe the process by a linear equation  , where
, where ![{\displaystyle 1=[1,1,\ldots ,1]^{T}}](./_assets_/43fec89f837a5e8d53869fb49ec95a9e56a788f0.svg) . Depending on context it will be clear if
. Depending on context it will be clear if  represents a scalar or a vector. Suppose that we know
 represents a scalar or a vector. Suppose that we know ![{\displaystyle [-x_{0},x_{0}]}](./_assets_/e79873ca5ddfd5d6b0168f6373b33c8bc3756c69.svg) to be the range within which the value of
 to be the range within which the value of  is going to fall in. We can model our uncertainty of
 is going to fall in. We can model our uncertainty of  by an aprior uniform distribution over an interval
 by an aprior uniform distribution over an interval ![{\displaystyle [-x_{0},x_{0}]}](./_assets_/e79873ca5ddfd5d6b0168f6373b33c8bc3756c69.svg) , and thus
, and thus  will have variance of
 will have variance of  . Let the noise vector
. Let the noise vector  be normally distributed as
 be normally distributed as  where
 where  is an identity matrix. Also
 is an identity matrix. Also  and
 and  are independent and
 are independent and  . It is easy to see that
. It is easy to see that 
 
Thus, the linear MMSE estimator is given by
 
We can simplify the expression by using the alternative form for  as
 as
 
where for ![{\displaystyle y=[y_{1},y_{2},\ldots ,y_{N}]^{T}}](./_assets_/8317c7045e5b1318cec0c4ee89727a02cdeecafc.svg) we have
 we have  
Similarly, the variance of the estimator is
 
Thus the MMSE of this linear estimator is
 
For very large  , we see that the MMSE estimator of a scalar with uniform aprior distribution can be approximated by the arithmetic average of all the observed data
, we see that the MMSE estimator of a scalar with uniform aprior distribution can be approximated by the arithmetic average of all the observed data 
 
while the variance will be unaffected by data  and the LMMSE of the estimate will tend to zero.
 and the LMMSE of the estimate will tend to zero.
However, the estimator is suboptimal since it is constrained to be linear. Had the random variable  also been Gaussian, then the estimator would have been optimal. Notice, that the form of the estimator will remain unchanged, regardless of the apriori distribution of
 also been Gaussian, then the estimator would have been optimal. Notice, that the form of the estimator will remain unchanged, regardless of the apriori distribution of  , so long as the mean and variance of these distributions are the same.
, so long as the mean and variance of these distributions are the same.
Example 3
Consider a variation of the above example: Two candidates are standing for an election. Let the fraction of votes that a candidate will receive on an election day be ![{\displaystyle x\in [0,1].}](./_assets_/1c44eb6b4643a03d3c166df0e61c4925b6d4d4f0.svg) Thus the fraction of votes the other candidate will receive will be
 Thus the fraction of votes the other candidate will receive will be  We shall take
 We shall take  as a random variable with a uniform prior distribution over
 as a random variable with a uniform prior distribution over ![{\displaystyle [0,1]}](./_assets_/738f7d23bb2d9642bab520020873cccbef49768d.svg) so that its mean is
 so that its mean is  and variance is
 and variance is  A few weeks before the election, two independent public opinion polls were conducted by two different pollsters. The first poll revealed that the candidate is likely to get
 A few weeks before the election, two independent public opinion polls were conducted by two different pollsters. The first poll revealed that the candidate is likely to get  fraction of votes. Since some error is always present due to finite sampling and the particular polling methodology adopted, the first pollster declares their estimate to have an error
 fraction of votes. Since some error is always present due to finite sampling and the particular polling methodology adopted, the first pollster declares their estimate to have an error  with zero mean and variance
 with zero mean and variance  Similarly, the second pollster declares their estimate to be
 Similarly, the second pollster declares their estimate to be  with an error
 with an error  with zero mean and variance
 with zero mean and variance  Note that except for the mean and variance of the error, the error distribution is unspecified. How should the two polls be combined to obtain the voting prediction for the given candidate?
 Note that except for the mean and variance of the error, the error distribution is unspecified. How should the two polls be combined to obtain the voting prediction for the given candidate?
As with previous example, we have
 
Here, both the  . Thus, we can obtain the LMMSE estimate as the linear combination of
. Thus, we can obtain the LMMSE estimate as the linear combination of  and
 and  as
 as
 
where the weights are given by
 
Here, since the denominator term is constant, the poll with lower error is given higher weight in order to predict the election outcome. Lastly, the variance of  is given by
 is given by
 
which makes  smaller than
 smaller than  Thus, the LMMSE is given by
 Thus, the LMMSE is given by 
 
In general, if we have  pollsters, then
 pollsters, then  where the weight for i-th pollster is given by
 where the weight for i-th pollster is given by  and the LMMSE is given by
 and the LMMSE is given by  
Example 4
Suppose that a musician is playing an instrument and that the sound is received by two microphones, each of them located at two different places. Let the attenuation of sound due to distance at each microphone be  and
 and  , which are assumed to be known constants. Similarly, let the noise at each microphone be
, which are assumed to be known constants. Similarly, let the noise at each microphone be  and
 and  , each with zero mean and variances
, each with zero mean and variances  and
 and  respectively. Let
 respectively. Let  denote the sound produced by the musician, which is a random variable with zero mean and variance
 denote the sound produced by the musician, which is a random variable with zero mean and variance  How should the recorded music from these two microphones be combined, after being synced with each other?
 How should the recorded music from these two microphones be combined, after being synced with each other?
We can model the sound received by each microphone as
 
Here both the  . Thus, we can combine the two sounds as
. Thus, we can combine the two sounds as 
 
where the i-th weight is given as
 
See also
Notes
Further reading
- Johnson, D. "Minimum Mean Squared Error Estimators". Connexions. Archived from Minimum Mean Squared Error Estimators the original on 25 July 2008. Retrieved 8 January 2013. 
- Jaynes, E.T. (2003). Probability Theory: The Logic of Science. Cambridge University Press. ISBN 978-0521592710.
- Bibby, J.; Toutenburg, H. (1977). Prediction and Improved Estimation in Linear Models. Wiley. ISBN 9780471016564.
- Lehmann, E. L.; Casella, G. (1998). "Chapter 4". Theory of Point Estimation (2nd ed.). Springer. ISBN 0-387-98502-6.
- Kay, S. M. (1993). Fundamentals of Statistical Signal Processing: Estimation Theory. Prentice Hall. pp. 344–350. ISBN 0-13-042268-1.
- Luenberger, D.G. (1969). "Chapter 4, Least-squares estimation". Optimization by Vector Space Methods (1st ed.). Wiley. ISBN 978-0471181170.
- Moon, T.K.; Stirling, W.C. (2000). Mathematical Methods and Algorithms for Signal Processing (1st ed.). Prentice Hall. ISBN 978-0201361865.
- Van Trees, H. L. (1968). Detection, Estimation, and Modulation Theory, Part I. New York: Wiley. ISBN 0-471-09517-6.
- Haykin, S.O. (2013). Adaptive Filter Theory (5th ed.). Prentice Hall. ISBN 978-0132671453.