Poster
Fast Iterative Hard Thresholding Methods with Pruning Gradient Computations
Yasutoshi Ida · Sekitoshi Kanai · Atsutoshi Kumagai · Tomoharu Iwata · Yasuhiro Fujiwara
West Ballroom A-D #5809
We accelerate the iterative hard thresholding (IHT) method, which finds (k) important elements from a parameter vector in a linear regression model. Although the plain IHT repeatedly updates the parameter vector during the optimization, computing gradients is the main bottleneck. Our method safely prunes unnecessary gradient computations to reduce the processing time.The main idea is to efficiently construct a candidate set, which contains (k) important elements in the parameter vector, for each iteration. Specifically, before computing the gradients, we prune unnecessary elements in the parameter vector for the candidate set by utilizing upper bounds on absolute values of the parameters. Our method guarantees the same optimization results as the plain IHT because our pruning is safe. Experiments show that our method is up to 73 times faster than the plain IHT without degrading accuracy.
Live content is unavailable. Log in and register to view live content