The following schedules are for Alex’s in-class students:

Let's improve our "big picture" view

1 Assessment
On the geodetic model and its solution

1 Assessment
In our previous lesson we looked at the overall process of solving a geomatics networks problem. We get to it as follows for the parametric case:

- By deriving the relevant functional model
- Then by linearizing it to get it into the form
- Then by solving that linear form under the constraint that

We also discussed how, fortunately, the general form of the latter solution has long been known under the name “estimation using a parametric least squares adjustment”.

That solution is as follows1I do not reference it specifically throughout the following, but one of my favourite treatments of this (and much more) is found in: Krakiwsky, E.J. (1990) *The Method of Least Squares: A Synthesis of Advances*, University of Calgary Geomatics Engineering Reports 10003.

Ever since our very first example, we’ve been building up to the way in which we can solve for estimates of the parameters and a set of adjusted measurements. Well here it is.

The vector containing the estimate of the unknown parameters, , is given by:

where

and where , , and are exactly the same design matrix, misclosure vector, and approximate coordinate vector that we studied in detail when we considered the linearized form of the functional model and when we practiced linearizing the fundamental observation equations.

The vector containing the estimated residuals, , is given by:

Recall our earlier treatments of what is a residual for the general parametric case. This is the least squares estimate of that residual. (And you might notice that it is a rearranged form of the original where is approximated by .)

From the above we can get the vector containing the adjusted measurements, , as follows:

As you can see, this is just our best guess at what the measurements should be – given our estimated residuals.

As we’ve discussed in class, one of the powers of the least squares approach is that it doesn’t just give you the parameters you’re after – it also gives you an estimate of their precision.

The least squares solution yields the following variance-covariance matrices.

Once we get to it, you will recognize that these come from not much more than the propagation of errors.

The variance-covariance matrix of the estimated unknown parameters is as follows:

The variance-covariance matrix of the adjusted measurements is as follows:

The variance-covariance matrix of the estimated residuals is as follows:

The variance-covariance matrix of the misclosure vector is just that of the observations:

As you can see from the solutions provided above, the whole least squares adjustment depends on the estimate of the variance-covariance matrix of the observations, . If this is not correct, then everything above is also not correct.

As we will also see again later in the course, this variance-covariance matrix is defined with an a-priori variance factor, which allows for statistical testing before the adjustment takes place (if is known).

Fortunately, if is not known, it’s possible to estimate the precision of the observations by looking at the residuals – which we can think about as the amount the observed values are “adjusted” by the estimation process. Put another way, we can estimate from the output of the adjustment itself:

If the factor is different from the a-priori factor by a statistically significantly amount then the covariance matrices should be scaled by it, e.g. for a-priori variance-covariance matrices:

and for variance-covariance matrices we have estimated:

For example, for the observations:

and for the estimated parameters:

At this point we can also defined the so-called normal matrix, , which depends on the datum characteristics and geometry of the network. It’s given by the following subset of what we saw above:

Which in turn means that some of the above equations can also be written as follows:

As a final note here, be aware that all of the above are just the algebraic forms of the expressions and do not necessarily indicate the best way to compute things in a practical situation.

Lesson Content

0% Complete
0/1 Steps

The following schedules are for Alex’s in-class students:

Welcome (back)!

Access to this site requires you to sign in. Use one of the icon(s) below:

This will take you to their secure sign in tools – we’ll never see your password and you don’t have to create and remember yet another one.

That said, we will receive your name and email address from them, which we will use to create and manage your account.