On Feb 9, 2014, at 9:28 PM, Jun Kudo <[email protected]> wrote:
> What’s the reason for this behavior?   From looking over Svanberg’s paper, 
> the trust region, rho parameter, and the form of the MMA approximations seem 
> to be independent of the magnitude of the function and gradients.  I’d like 
> to think that the initial steps should then be independent of the magnitude 
> of the function/gradient.
> 

Ideally, optimization algorithms would be invariant under rescaling of the 
coordinates or of the function values, but unfortunately this is almost never 
the case.

For example, equation (6.1a) of the Svanberg paper suggests initializing rho=1, 
and NLopt uses this initial value.   Unfortunately, rho is a dimensionful 
constant (it has the same units as f), so if you rescale the objective function 
then you are effectively changing the initial rho.  This changes the initial 
step sizes and hence the convergence rate: if rho is too large initially, then 
the first few steps will be smaller than necessary.

Unfortunately, it seems quite difficult to get rid of these "magic" 
dimensionful constants in most algorithms.   The algorithm still eventually 
converges, but they may affect the convergence rate, especially in the 
beginning.

The upshot is that it is typically a good idea to scale your functions and 
coordinates so that they have values of order unity, as that is the regime for 
which the magic constants were probably tuned.

--SGJ
_______________________________________________
NLopt-discuss mailing list
[email protected]
http://ab-initio.mit.edu/cgi-bin/mailman/listinfo/nlopt-discuss

Reply via email to