Proportional Odds Model Proofs

Property 1: The maximum value of the log-likelihood function LL occurs when the following r+k-1 equations hold for h = 1 to r-1.

Newton's method equations

or alternatively

Newton's method equations (alternative)and for j = 1 to k

More Newton's method equations

where

e_he'_h

Proof

The likelihood function is maximized when the first partial derivatives of LL with respect to the ah are zero for h = 1 to r-1 and the first partial derivatives of LL with respect to the bj are also zero for j = 1 to k. We define the following:

z_i and z_ih

Thus

p_ih

We also note that for 1 < h < r

p_ih - p_i,h-1 p_ih - p_i.h-1 continued

from which it follows that

ln(p_ih - p_i.h-1)

For h = 1 and h = r, we have

h = 0h = r

We next calculate various partial derivatives

Partial derivatives 1Partial derivative 2Partial derivative 3

For 1 < h < r, we have

Partial derivative 4Partial derivative 4bPartial derivative 4c

For h = 1 and h = r, we get

Partial derivative 5Partial derivative 6

Since pi0 = 0 and pir = 1, even when h = 1 or h = r

Partial derivative 7

And so this result holds for 1 ≤ h ≤ r.

For 1 < h < r, we have

Partial derivative 8b

For h = 1, we have

Parrtial derivative 9

and so for 1 ≤ hr-1

We next consider

Partial derivative 10aPartial derivative 10bPartial derivative 10c

which is valid for 0 < h < r-1. It is also valid for h = r-1 as can be seen from

Partial derivative 11

We also note that

Partial derivative 12

Finally, we now calculate the partial derivatives of LL using the results described above.

Partial derivative LL/b_jDerivative LL/b_j continued

Since

Partial derivative 12

it follows that for 0 < h < r

Derivative LL/b_j part 1Derivative LL/b_j part 2

which completes the proof.

Property A: The Jacobian matrix of the second partial derivatives of LL is an k+r-1 × k+r-1 symmetric matrix of form

Jacobian matrix J

where C = [chl] is an r-1 x r-1 matrix, D = [djg] is a k × k matrix and U = [ujh] is an k × r-1 matrix consisting of the following elements:

c_hhc_h,h+1Other c_hl elements

d_jgu_jh

where

g_h

Proof

We calculate the various second partial derivatives of LL using the properties described in the proof of Property 1. We start with the 

d terms

terms.

Second derivative LL/bb 2

Second derivative LL/bb 2Second derivative LL/bb 3Second derivative LL/bb 4

Second derivative LL/bb 5

Next, we calculate the

u terms

terms.

Second derivative LL/ba 1Second derivative LL/ba 2

Finally, we have the c termscase. First, we note that

Partial derivative e/a Partial derivative e'/a

which is also true in the following two cases:

Boundary cases

We now start with the case where l = h.

Second derivative LL/aa 1Second derivative LL/aa 2Second derivative LL/aa 3

If l = h + t where t ≥ 2, then

Second derivative LL/aa 4

and so

Second derivative LL/aa 5

Finally, we consider the case where  and 0 < h < r-1, but first we note that

Partial derivative e/aPartial derivative e'/a

It now follows that

Second derivative LL/aa

This completes the proof.

Property 2 (Newton’s method)

Proof: Let

Coefficient vector

F vector using derivatives

Jacobian using derivatives

Newton estimate of B

Using Newton’s method, we conclude that B* is a better estimate of the regression coefficients than B (i.e. it produces a larger value for LL), and so the sequence B, B*, B**, B***, etc. converges to the coefficient vector that maximizes LL and the corresponding Hessian matrix J-1 is a good estimate of the covariance matrix for the regression coefficients.

From Property 1, we see that the F vector can be expressed as

F vector

where

v_h

w_j

and by Property A, we see that the Jacobian matrix can be expressed as 

Jacobian matrix J

where C = [chl], D = [djg] and U = [ujh] are as described in Property A. This completes the proof.

Leave a Comment