In this work, we consider learning sparse models in large scale setting, where the number of samples and the feature dimension can grow as large as millions or billions. Two immediate issues occur under such challenging scenarios: (i) com- putational cost; (ii) memory overhead. In particular, the memory issue precludes a large volume of prior algorithms that are based on batch optimization technique. To remedy the problem, we propose to learn sparse models such as Lasso in online manner where in each iteration, only one randomly chosen sample is revealed to update a sparse iterate. Thereby, the memory cost is independent of the sample size and gradient evaluation for one sample is efficient. Perhaps amazing, we find that with the same parameter, sparsity promoted by batch methods is not preserved in online fashion. We analyze such interesting phenomenon and illustrate some effective variants including mini-batch methods and a hard thresholding based stochastic gradient algorithm. Extensive experiments are carried out on a public dataset which supports our findings and algorithms.
-
Notifications
You must be signed in to change notification settings - Fork 0
In this work, we consider learning sparse models in large scale setting, where the number of samples and the feature dimension can grow as large as millions or billions. Two immediate issues occur under such challenging scenarios: (i) com- putational cost; (ii) memory overhead.
License
dhingratul/Large-Scale-Sparse-Models
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
In this work, we consider learning sparse models in large scale setting, where the number of samples and the feature dimension can grow as large as millions or billions. Two immediate issues occur under such challenging scenarios: (i) com- putational cost; (ii) memory overhead.
Topics
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published