Ref: Barfoot, Timothy D. State Estimation for Robotics. Cambridge University Press, 2017. Section 5.3

Deriving Cauchy M-Estimator For M.A.P Estimation

We would like to derive a common M-Estimator (Maximum Likelihood Type Estimator), the Cauchy cost function, from the point of view of covariance estimation. Specifically, we would like to derive it from the point-of-view of maximizing the a posteriori distribution.

We will begin with the traditional MAP estimation cost function.

where $r_n(x)$ is the residual of state $x$ at iteration $n$, and $\Sigma_n$ is the provided covariance at iteration $n$. This provides us with the traditional optimization problem, as provided below.

In the optimization problem provided above, it is assumed that the covariance is provided a priori. However, this is not always a valid assumption, so, it is desirable to be able to estimation both the state vector and the covariance estimate concurrently. To do so, we can augment our optimization problem as

which can be factorized as

Now, we need to provide a prior on our covariance matrix. A commonly utilized prior, due to the fact that it’s a conjugate proir to symmetric nonnegative-definite matrices, is the Inverse Wishart distribution, which is defined as,

where, $\nu_n$ is the degrees-of-freedom, and $\Psi_n$ is a scaling matrix.

If we plug the inverse wishart prior into the factorized objective function, we are left with,

Now, to find the optimal covariance estimate, we can set the partial derivative of our objective function w.r.t. $\Sigma^{-1}$ equal to zero, as provided below.

Setting the expression provided above equal to zero, we are left with the optimal covariance estimate.

Finally, if we plug our expression for the optimal $\Sigma_n$ back into the objective function, we are left with,

where $\alpha = \nu_n + \Sigma_n + 2$. Which is the same objective function as specified by the Cauchy M-estimator.