This website requires JavaScript.
Explore
Help
Sign In
tommy
/
libpostal
Watch
1
Star
0
Fork
0
You've already forked libpostal
Code
Issues
Pull Requests
Actions
Packages
Projects
Releases
Wiki
Activity
1,292
Commits
2
Branches
0
Tags
62017fd33d708a227373d4091929178b0fb069fe
Commit Graph
2 Commits
Author
SHA1
Message
Date
Al
62017fd33d
[optimization] Using sparse updates in stochastic gradient descent. Decomposing the updates into the gradient of the loss function (zero for features not observed in the current batch) and the gradient of the regularization term. The derivative of the regularization term in L2-regularized models is equivalent to an exponential decay function. Before computing the gradient for the current batch, we bring the weights up to date only for the features observed in that batch, and update only those values
2016-01-09 03:37:31 -05:00
Al
8b70529711
[optimization] Stochastic gradient descent with gain schedule a la Leon Bottou
2016-01-08 00:54:17 -05:00