## Now let’s do it for our own model

In the last lesson we just learned that as long as we can get our hands on a decent approximate to what we’re after estimating, then we can turn our nonlinear functional models into linear approximations. Let’s put this to work for our situation where .

But how can we get approximate values of and ?

Just as in the general case (in the lesson How do we linearize?) we wrote:

we will now write the following for our desired unknown parameter, :

where:

is an approximation to

is an unknown correction to , or the difference between it and

And for our observed quantities, , we will just use the measurements themselves and our earlier equation:

where:

our measurements provide an approximation to

are our statistical errors, which should be familiar to you from our first lesson (What are errors and residuals? (And some other sanity checks out of the gate) as the difference between the measurement and the actual unknown observed quantity

Now we have the required approximations of the type but relevant for our situation and using them we can write the following:

Or (and this is the really awesome bit), we can write it as follows using what we now know about Taylor’s Theorem:

So we’ve taken our general functional model and expressed it in linear form!

It’s worth noting again that the terms and are the true (and unknown) corrections to the approximate values and that we used. (Just as was the true unknown correction to in Taylor’s Theorem.)

And it’s worth having a look at what each of the terms really means, which we will do next.

## Let’s have a look at this thing

Now we’ve got our linearize model. So let’s have a look at each of its components in turn.

is a vector containing the values of , computed at the known points . We will denote this as:

and call it the misclosure vector.

And is a matrix of size x that we will denote with the letter , where the row will be the partial differentials of with respect to , , , … .

where we’ve introduced the following slight shorthand for convenience of notation:

to indicate that the derivatives are evaluated at the approximate values, .

And, finally, is a matrix of size x that we will denote with the letter , where the rows contain the partial differentials of with respect to , , … .

where we’ve introduced the following slight shorthand for convenience of notation:

to indicate that the derivatives are evaluated at the measured values, .

## Summary of the general linearized model equation

We refer to the matrices and as design matrices. And with them in hand, we can write the function form of our functional math models:

where the misclosure vector is given by:

and the design matrices are given by:

## The special cases

### The linearized combined model

In the lesson Let’s linearize our general functional model, we arrived at the linearized model:

This is referred to as the combined linear model which we know has the non-linear form .

There are two special cases of this model, for each of the other types of equation we saw in So, to summarize functional modeling.

### The linearized parametric model

For the parametric model which has the form , it can be shown that the term

e.g.

so we get the following linearized form:

### The linearized condition model

And for the condition model which has the form , it can be shown that the term

(e.g. since there are no parameters!)

so we get the following linearized form:

## In summary

The following table summarizes the linearized functional models we’ve been dealing with.

Combined equations | Observation equations (Parametric model) |
Condition equations | |
---|---|---|---|

Nonlinear functional model | |||

number of equations | – (see note 1) | ||

number of observations | |||

number of parameters | (see note 1) | ||

degrees of freedom | – | – = – | |

linearized model (see note 2) |

Note 1: It might seem strange to say that there are no parameters in the condition equations (i.e. that = ) and then say that the number of equations is –. In this case, it is the number of parameters that *would* be used if the problem was solved using observation equations. And, as such, the degrees of freedom is in that case.

Note 2: I have used the error term to stay consistent with earlier lessons, e.g. right back to What are errors and residuals? (And some other sanity checks out of the gate), because I find there’s often a fundamental misunderstanding about what is and what it isn’t. That said, it’s not uncommon the see the symbols or used in its place. These are matters of preference and convention, and the meaning doesn’t change in either case.

**Lesson Assessments**