Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Modifying AdaProx for LASSO #23

Open
pythonometrist opened this issue Dec 9, 2020 · 2 comments
Open

Modifying AdaProx for LASSO #23

pythonometrist opened this issue Dec 9, 2020 · 2 comments

Comments

@pythonometrist
Copy link

This isnt an issue per se. I did want to figure out if I could use a similar approach for a simple LASSO regression in pytorch. Working with proximal operators with SGD is straightforward (but then SGD has step size issues). ADAM requires memory for past gradients - but isn't meant for non-differentiable convex problems (even though L1 regularization does improve results a fair bit). I wanted tos ee if AdaProx improves results.

@pmelchior
Copy link
Owner

Do you have an elementwise l1 penalty? If so, you can use operators.prox_soft. adaprox should work then.

@pythonometrist
Copy link
Author

Thanks- let me dig into it and revert. I am going to evaluate how this compares with a smooth Huber loss for linear regression.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants