Regression.jl seems to have like a sweet implementation of gradient based 
optimization algorithms. How does this compare to the work in Optim.jl? 
Would it be useful to join these efforts?

Op donderdag 23 april 2015 11:12:58 UTC+2 schreef Dahua Lin:
>
> Hi, 
>
> I am happy to announce three packages related to empirical risk 
> minimization
>
> EmpiricalRisks <https://github.com/lindahua/EmpiricalRisks.jl>
>
> This Julia package provides a collection of predictors and loss functions, 
> as well as the efficient computation of gradients, mainly to support the 
> implementation of (regularized) empirical risk minimization methods.
>
> Predictors:
>
>    - linear prediction
>    - affine prediction
>    - multivariate linear prediction
>    - multivariate affine prediction
>    
> Loss functions:
>
>    - squared loss
>    - absolute loss
>    - quantile loss
>    - huber loss
>    - hinge loss
>    - smoothed hinge loss
>    - logistic loss
>    - sum squared loss (for multivariate prediction)
>    - multinomial logistic loss
>    
> Regularizers:
>
>    - squared L2 regularization
>    - L1 regularization
>    - elastic net (L1 + squared L2)
>    - evaluation of proximal operators, w.r.t. these regularizers.
>    
>
> Regression <https://github.com/lindahua/Regression.jl>
>
> This package was dead before, and I revived it recently. It is based on 
> EmpiricalRisks, and provides methods for regression analysis (for moderate 
> size problems, i.e. the data can be loaded entirely to memory). It supports 
> the following problems:
>
>    - Linear regression
>    - Ridge regression
>    - LASSO
>    - Logistic regression
>    - Multinomial Logistic regression
>    - Problems with customized loss and regularizers
>    
> It also provides a variety of solvers:
>
>    - Analytical solution (for linear & ridge regression)
>    - Gradient descent
>    - BFGS
>    - L-BFGS
>    - Proximal gradient descent (recommended for LASSO & sparse regression)
>    - Accelerated gradient descent (experimental)
>
>
> SGDOptim <https://github.com/lindahua/SGDOptim.jl>
>
> I announced this couple weeks ago. Now this package has been fundamentally 
> refactored, and now it is based on EmpiricalRisks. It aims to provide 
> stochastic algorithms (e.g. SGD) for solve large scale regression problems.
>
>
> Cheers,
> Dahua
>
>
>
>
>
>
>

Reply via email to