Skip to yearly menu bar Skip to main content


Poster

STk: A Scalable Module for Solving Top-k Problems

Hanchen Xia · Weidong Liu · Xiaojun Mao

East Exhibit Hall A-C #4708
[ ]
Wed 11 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract: The cost of ranking becomes significant in the new stage of deep learning. We propose STk, a fully differentiable module with a single trainable parameter, designed to solve the Top-k problem without requiring additional time or GPU memory. Due to its fully differentiable nature, STk can be embedded end-to-end into neural networks and optimize the Top-k problems within a unified computational graph. We apply STk to the Average Top-k Loss (ATk), which inherently faces a Top-k problem. The proposed STk Loss outperforms ATk Loss and achieves the best average performance on multiple benchmarks, with the lowest standard deviation. With the assistance of STk Loss, we surpass the state-of-the-art (SOTA) on both CIFAR-100-LT and Places-LT leaderboards.

Chat is not available.