From: Leonid Gibiansky <*lgibiansky*>

Date: Thu, 12 Feb 2015 10:09:08 -0500

I think we are back to the discussion of usefulness of SE/RSE provided

by Nonmem (search in archives will recover many e-mails on the subject).

I am reluctant to disagree with Lewis Sheiner (if this indeed was his

citation that was not taken out of context) but SEs are very useful (not

perfect but useful) as an indicator of identifiability of the problem

and some measure of precision. One cannot do bootstrap on each and every

step of the modeling (moreover, bootstrap CI are not that different from

asymptotic CI, especially for well-estimated parameters). Profiling is

also local, and unlikely to give something significantly different from

the Nonmem SE results unless RSEs are very large (and profiling is as

time-consuming as bootstrap). So SEs serve as a very useful indicator

whether data support the model.

Note that Nonmem SEs are local; they rely on approximation of the OF

surface, essentially, give the curvature of this surface, so they do not

take into account any constrains. In this sense, form of the model

(CL=THETA(1) or CL=log(THETA(2)) just define the rule of extrapolation

beyond the locality of the estimate. For practical purposes, SEs of

these two representations are related (approximately) as

SE(THETA2)=SE(THETA(1))/THETA(1)

(propagation of error rule). Given SE(THETA(1)) and THETA(1) you can

estimate SE(THETA(2)) and simulate accordingly.

Thanks

Leonid

--------------------------------------

Leonid Gibiansky, Ph.D.

President, QuantPharm LLC

web: www.quantpharm.com

e-mail: LGibiansky at quantpharm.com

tel: (301) 767 5566

On 2/12/2015 9:04 AM, Bob Leary wrote:

*> Dear Aziz -
*

*> The approximate likelihood methods in NONMEM such as FO, FOCE,and LAPLACE optimize an objective function than is parameterized internally
*

*> by the Cholesky factor L of Omega, regardless of whether the matrix is diagonal (the EM -based methods do something considerably different and work directly with Omega rather than
*

*> the Cholesky factor.)
*

*>
*

*> Thus for the approximate likelihood methods, the SE's computed internally by $COV from the Hessian or Sandwich or Fisher score methods
*

*> are first computed with respect to these Cholesky parameters , and then the corresponding SE's of the full Omega=LL' are computed by a 'propagation of errors' approach
*

*> which skews the results, particularly if the SE's are large. Thus in a sense regarding your dilemma about whether Model 1 or Model 2 is better with respect to applicability of $COV results, one answer is that both are fundamentally distorted by the propagation of errors method with respect to the Omega elements.
*

*>
*

*> But regarding your fundamental question 'can we trust the output of $COV '- all of this makes very little difference. Standard errors computed by $COV are inherently dubious - the applicability of the usual asymptotic arguments is very questionable for the types/sizes of data sets we often deal with.
*

*> As Lewis Sheiner used to say of these results, 'they are not worth the electrons used to compute them'. They are the best we can do for the level of computational investment put into them -
*

*> If you want something better, try a bootstrap or profiling method.
*

*>
*

*>
*

*>
*

*> ________________________________________
*

*>
*

*>
*

*> NOTICE: The information contained in this electronic mail message is intended only for the personal and confidential
*

*> use of the designated recipient(s) named above. This message may be an attorney-client communication, may be protected
*

*> by the work product doctrine, and may be subject to a protective order. As such, this message is privileged and
*

*> confidential. If the reader of this message is not the intended recipient or an agent responsible for delivering it to
*

*> the intended recipient, you are hereby notified that you have received this message in error and that any review,
*

*> dissemination, distribution, or copying of this message is strictly prohibited. If you have received this
*

*> communication in error, please notify us immediately by telephone and e-mail and destroy any and all copies of this
*

*> message in your possession (whether hard copies or electronically stored copies). Thank you.
*

*>
*

*> buSp9xeMeKEbrUze
*

*>
*

*>*

Received on Thu Feb 12 2015 - 10:09:08 EST

Date: Thu, 12 Feb 2015 10:09:08 -0500

I think we are back to the discussion of usefulness of SE/RSE provided

by Nonmem (search in archives will recover many e-mails on the subject).

I am reluctant to disagree with Lewis Sheiner (if this indeed was his

citation that was not taken out of context) but SEs are very useful (not

perfect but useful) as an indicator of identifiability of the problem

and some measure of precision. One cannot do bootstrap on each and every

step of the modeling (moreover, bootstrap CI are not that different from

asymptotic CI, especially for well-estimated parameters). Profiling is

also local, and unlikely to give something significantly different from

the Nonmem SE results unless RSEs are very large (and profiling is as

time-consuming as bootstrap). So SEs serve as a very useful indicator

whether data support the model.

Note that Nonmem SEs are local; they rely on approximation of the OF

surface, essentially, give the curvature of this surface, so they do not

take into account any constrains. In this sense, form of the model

(CL=THETA(1) or CL=log(THETA(2)) just define the rule of extrapolation

beyond the locality of the estimate. For practical purposes, SEs of

these two representations are related (approximately) as

SE(THETA2)=SE(THETA(1))/THETA(1)

(propagation of error rule). Given SE(THETA(1)) and THETA(1) you can

estimate SE(THETA(2)) and simulate accordingly.

Thanks

Leonid

--------------------------------------

Leonid Gibiansky, Ph.D.

President, QuantPharm LLC

web: www.quantpharm.com

e-mail: LGibiansky at quantpharm.com

tel: (301) 767 5566

On 2/12/2015 9:04 AM, Bob Leary wrote:

Received on Thu Feb 12 2015 - 10:09:08 EST