Efficient classification for metric data

Lee Ad Gottlieb, Aryeh Kontorovich, Robert Krauthgamer

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

19 Scopus citations

Abstract

Recent advances in large-margin classification of data residing in general metric spaces (rather than Hilbert spaces) enable classification under various natural metrics, such as edit and earthmover distance. The general framework developed for this purpose by von Luxburg and Bousquet [JMLR, 2004] left open the question of computational efficiency and providing direct bounds on classification error. We design a new algorithm for classification in general metric spaces, whose runtime and accuracy depend on the doubling dimension of the data points. It thus achieves superior classification performance in many common scenarios. The algorithmic core of our approach is an approximate (rather than exact) solution to the classical problems of Lipschitz extension and of Nearest Neighbor Search. The algorithm's generalization performance is established via the fat-shattering dimension of Lipschitz classifiers.

Original languageEnglish
Title of host publicationCOLT 2010 - The 23rd Conference on Learning Theory
Pages433-440
Number of pages8
StatePublished - 1 Dec 2010
Event23rd Conference on Learning Theory, COLT 2010 - Haifa, Israel
Duration: 27 Jun 201029 Jun 2010

Publication series

NameCOLT 2010 - The 23rd Conference on Learning Theory

Conference

Conference23rd Conference on Learning Theory, COLT 2010
Country/TerritoryIsrael
CityHaifa
Period27/06/1029/06/10

Fingerprint

Dive into the research topics of 'Efficient classification for metric data'. Together they form a unique fingerprint.

Cite this