TY - UNPB
T1 - Metric-valued regression
AU - Cohen, Dan Tsir
AU - Kontorovich, Aryeh
PY - 2022/2/7
Y1 - 2022/2/7
N2 - We propose an efficient algorithm for learning mappings between two metric spaces, X and Y. Our procedure is strongly Bayes-consistent whenever X and Y are topologically separable and Y is “bounded in expectation” (our term; the separability assumption can be somewhat weakened). At this level of generality, ours is the first such learnability result for unbounded loss in the agnostic setting. Our technique is based on metric medoids (a variant of Fréchet means) and presents a significant departure from existing methods, which, as we demonstrate, fail to achieve Bayes-consistency on general instance- and label-space metrics. Our proofs introduce the technique of semi-stable compression, which may be of independent interest.
AB - We propose an efficient algorithm for learning mappings between two metric spaces, X and Y. Our procedure is strongly Bayes-consistent whenever X and Y are topologically separable and Y is “bounded in expectation” (our term; the separability assumption can be somewhat weakened). At this level of generality, ours is the first such learnability result for unbounded loss in the agnostic setting. Our technique is based on metric medoids (a variant of Fréchet means) and presents a significant departure from existing methods, which, as we demonstrate, fail to achieve Bayes-consistency on general instance- and label-space metrics. Our proofs introduce the technique of semi-stable compression, which may be of independent interest.
KW - cs.LG
KW - stat.ML
U2 - 10.48550/arXiv.2202.03045
DO - 10.48550/arXiv.2202.03045
M3 - Preprint
BT - Metric-valued regression
ER -