Dear Prof Holford, Ken, Alan, Jeroen & others,

Thanks for the engaging discussions.

In context of monitoring at the iteration level, I vaguely recall that in
NMUSERS or in one of ACOP conferences , there was a presentation &
demonstration with R scripts on looking at the convergence and other
parameters in real time.

The interpretations of SEs is interesting based on linear or non-linear
models, and also based on size of variance of parameters.

On a different note, I am also interested in hearing from you about SEs
when estimated based on transformed distribution space and their values &
interpretations in back-transformed space. Would the notion of precision
 still be valid  when viewing both transformed and untransformed space?
This is in context of dealing with untransformed space of non-normal or
non-lognormal distributions.

Best regards
Santosh


On Mon, Jul 29, 2024 at 8:52 AM Nick Holford <n.holf...@auckland.ac.nz>
wrote:

> Hi Jeroen,
>
> A small correction. Please re-read my email to nmusers on 12 Feb 2015
> which I quote here. Sorry I cannot show the original but the 1999 URL is
> not available to me anymore.
>
> =================  start quote ===================
> Nick Holford Thu, 12 Feb 2015 11:54:59 -0800
> Hi,
> The original quote about electrons comes from a remark I made in 1999 on
> nmusers.
> http://www.cognigencorp.com/nonmem/nm/99nov121999.html
> Lewis Sheiner agreed in the same thread. Thanks to the wonders of living
> on a sphere Lewis appears to agree with me the day before I made the
> comment :-)
> =================  end quote ===================
>
> I had been meaning to add to Ken's great email which confirms my original
> assertion about electrons.
>
> If Santosh really wanted to calculate SE's after every "iteration" (which
> I think was Ken's interpretation of  every "estimation") then this can be
> done by running a non-parametric bootstrap with the parameter estimates
> produced after every iteration.
>
> I wonder if Santosh would like to spend a few hours doing that and adding
> to the nmusers collection about standard errors by reporting the results to
> us?
>
>
> Best wishes,
> Nick
>
>
> --
> Nick Holford, Professor Emeritus Clinical Pharmacology, MBChB, FRACP
> mobile: NZ+64(21) 46 23 53 ; FR+33(6) 62 32 46 72
> email: n.holf...@auckland.ac.nz
> web: http://holford.fmhs.auckland.ac.nz/
>
> -----Original Message-----
> From: owner-nmus...@globomaxnm.com <owner-nmus...@globomaxnm.com> On
> Behalf Of Jeroen Elassaiss-Schaap (PD-value B.V.)
> Sent: Monday, July 29, 2024 3:37 PM
> To: kgkowalsk...@gmail.com; 'Santosh' <santosh2...@gmail.com>;
> nmusers@globomaxnm.com
> Cc: 'Alan Maloney' <al_in_swe...@hotmail.com>; Pyry Välitalo <
> pyry.valit...@gmail.com>
> Subject: Re: [NMusers] Obtaining RSE%
>
> [Some people who received this message don't often get email from
> jer...@pd-value.com. Learn why this is important at
> https://aka.ms/LearnAboutSenderIdentification ]
>
> Dear NMusers,
>
> This is a great reminder for us to consider the reliability of standard
> errors in our models, thanks Ken & Alan. The more non-linear the models
> become, the less reliable and the more important other perspectives on
> parameter values such as sensitivity analysis and prior knowledge.
>
> The nmusers archive has many great threads on the topic that are available
> to review such as
> https://www.mail-archive.com/nmusers@globomaxnm.com/msg05423.html and
> related https://www.mail-archive.com/nmusers@globomaxnm.com/msg05419.html
> . In summary, log-transformation only can get you so far but can perhaps be
> seen as a sort of minimal effort.
>
> To add to the Lewis's quote about SEs - "they are not worth the electrons
> used to compute them" (see the links), Pyry had some very interesting
> observations he shared with me about the SE of the CV of a log-normal
> omega: it inflates with higher values of omega compared to the SE of omega
> itself.
>
> Best regards,
>
> Jeroen
>
> http://pd-value.com
> jer...@pd-value.com
> @PD_value
> +31 6 23118438
> -- More value out of your data!
>
> On 29-07-2024 14:41, kgkowalsk...@gmail.com wrote:
> >
> > Dear NMusers,
> >
> > It was recently pointed out to me by a statistical colleague that my
> > recent NMusers post about the inverse Hessian (R matrix) evaluated at
> > the maximum likelihood estimates is a consistent estimator of the
> > covariance matrix (i.e., converges to the true value with large N) is
> > only true for linear models.  For nonlinear models, the standard
> > errors produced by NONMEM and other nonlinear estimation software are
> > not only asymptotic but also approximate.  Moreover, how well that
> > approximation works will also depend on the parameterization.  This I
> > believe is one of the motivations behind “mu referencing” in NONMEM
> > and the use of log transformations of the parameters to help improve
> > Wald-based approximations.  I thank Alan Maloney for pointing this out
> > to me.
> >
> > Kind regards,
> >
> > Ken
> >
> > *From:*kgkowalsk...@gmail.com <kgkowalsk...@gmail.com>
> > *Sent:* Saturday, July 27, 2024 12:36 PM
> > *To:* 'Santosh' <santosh2...@gmail.com>; nmusers@globomaxnm.com
> > *Subject:* RE: [NMusers] Obtaining RSE%
> >
> > Dear Santosh,
> >
> > There is a good reason for this.  Wald (1943) has shown that the
> > inverse of the Hessian (R matrix) evaluated at the maximum likelihood
> > estimates is a consistent estimator of the covariance matrix.  It is
> > based on Wald’s approximation that the likelihood surface locally near
> > the maximum likelihood estimates can be approximated by a quadratic
> > function in the parameters.  This theory does not hold for any set of
> > parameter estimates along the algorithm’s search path prior to
> > convergence to the maximum likelihood estimates. Moreover,  inverting
> > the Hessian evaluated at an interim step prior to convergence would
> > likely be a poor approximation especially early in the search path
> > where the gradients are large (i.e., large changes in OFV for a given
> > change in the parameters would probably have substantial curvature and
> > not be well approximated by a quadratic model in the parameters).
> >
> > Thus, the COV step in NONMEM is only applied once convergence is
> > obtained during the EST step.
> >
> > Wald, A. “Tests of statistical hypotheses concerning several
> > parameters when the number of observations is large.” /Trans. Amer.
> > Math. Soc./ 1943;54:426.
> >
> > Best,
> >
> > Ken
> >
> > Kenneth G. Kowalski
> >
> > President
> >
> > Kowalski PMetrics Consulting, LLC
> >
> > Email: kgkowalsk...@gmail.com <mailto:kgkowalsk...@gmail.com>
> >
> > Cell:  248-207-5082
> >
> > *From:*owner-nmus...@globomaxnm.com
> > <mailto:owner-nmus...@globomaxnm.com><owner-nmus...@globomaxnm.com
> > <mailto:owner-nmus...@globomaxnm.com>> *On Behalf Of *Santosh
> > *Sent:* Friday, July 26, 2024 3:38 AM
> > *To:* nmusers@globomaxnm.com <mailto:nmusers@globomaxnm.com>
> > *Subject:* [NMusers] Obtaining RSE%
> >
> >  Dear esteemed experts!
> >
> > When using one or more estimation methods & covariance step in a
> > NONMEM control stream, the resulting ext file contains final estimate
> > (for all estimation steps)  & standard error (only for the last
> > estimation step).
> >
> > Is there a way that standard error is generated for every estimation
> step?
> >
> > TIA
> >
> > Santosh
> >
>
>

Reply via email to