Skip to yearly menu bar Skip to main content


Lightning Talk
in
Workshop: Data Centric AI

LSH methods for data deduplication in a Wikipediaartificial dataset


Abstract:

This paper illustrates locality sensitive hasing (LSH) models for the identification and removal of nearly redundant data in a text dataset. To evaluate the different models, we create an artificial dataset for data deduplication using English Wikipedia articles. Area-Under-Curve (AUC) over 0.9 were observed for most models, with the best model reaching 0.96. Deduplication enables more effective model training by preventing the model from learning a distribution that differs from the real one as a result of the repeated data.