On 1 September 2014 12:34, Troels Emtekær Linnet <[email protected]> wrote:
> Anyway, before minfx can handle constraints in for example BFGS,
> this is just a waste of time.

Minfx can do this :)  The log-barrier constraint algorithm works with
all optimisation techniques in minfx, well, apart from the grid search
(https://en.wikipedia.org/wiki/Barrier_function#Logarithmic_barrier_function).
And if gradients are supplied, the more powerful
Methods-of-Multipliers algorithm can also be used in combination with
all optimisation techniques
(https://en.wikipedia.org/wiki/Augmented_Lagrangian_method).


> I think there will be a 10 x speed up, just for the Jacobian.

For the analytic models, you could have a 10x speed up if symbolic
gradients and Hessians are implemented.  I'm guessing that's what you
mean.


> And when you have the Jacobian, estimating the errors are trivial.
>
> std(q) = sqrt ( (dq/dx std(x))*2 + (dq/dz std(z))*2 )

:S  I'm not sure about this estimate.  It looks rather too linear.  I
wish errors would be so simple.


> where q is the function. x and z are R1 and R1rho_prime.
>
> So, until then, implementing the Jacobian is only for testing the
> error estimation compared to
> Monte-Carlo simulations.

If you do add the equations, the lib.dispersion.dpl94 module would be
the natural place to put them.  And the interface as dfunc_DPL94(),
d2func_DPL94(), and jacobian_DPL94().

Regards,

Edward

_______________________________________________
relax (http://www.nmr-relax.com)

This is the relax-devel mailing list
[email protected]

To unsubscribe from this list, get a password
reminder, or change your subscription options,
visit the list information page at
https://mail.gna.org/listinfo/relax-devel

Reply via email to