CTJLSVM: Componentwise Triple Jump Acceleration for Training Linear SVM

author: Chun-Nan Hsu, AIIA Lab, Academia Sinica
published: Sept. 1, 2008,   recorded: July 2008,   views: 3211


Related Open Educational Resources

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.


The triple jump extrapolation method is an effective approximation of Aitken’s acceleration for accelerating the convergence of many machine learning algorithms that can be formulated as fixedpoint iteration. In the remainder of this abstract, we briefly review the general idea of the triple jump method and then describe how to apply it to accelerate stochastic gradient descent (SGD) for training linear support vector machines (SVM).

See Also:

Download slides icon Download slides: icml08_hsu_ctja_01.pdf (1009.9 KB)

Help icon Streaming Video Help

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: