Skip to yearly menu bar Skip to main content


Poster

Semi-Supervised Learning with Declaratively Specified Entropy Constraints

Haitian Sun · William Cohen · Lidong Bing

Room 517 AB #112

Keywords: [ Natural Language Processing ] [ Semi-Supervised Learning ]


Abstract:

We propose a technique for declaratively specifying strategies for semi-supervised learning (SSL). SSL methods based on different assumptions perform differently on different tasks, which leads to difficulties applying them in practice. In this paper, we propose to use entropy to unify many types of constraints. Our method can be used to easily specify ensembles of semi-supervised learners, as well as agreement constraints and entropic regularization constraints between these learners, and can be used to model both well-known heuristics such as co-training, and novel domain-specific heuristics. Besides, our model is flexible as to the underlying learning mechanism. Compared to prior frameworks for specifying SSL techniques, our technique achieves consistent improvements on a suite of well-studied SSL benchmarks, and obtains a new state-of-the-art result on a difficult relation extraction task.

Live content is unavailable. Log in and register to view live content