Abstract
We consider the problem of learning a multi-class classifier from labels as well as simple explanations that we call discriminative features. We show that such explanations can be provided whenever the target concept is a decision tree, or can be expressed as a particular type of multi-class DNF formula. We present an efficient online algorithm for learning from such feedback and we give tight bounds on the number of mistakes made during the learning process. These bounds depend only on the representation size of the target concept and not on the overall number of available features, which could be infinite. We also demonstrate the learning procedure experimentally.
Original language | English |
---|---|
Pages (from-to) | 3955-3963 |
Number of pages | 9 |
Journal | Advances in Neural Information Processing Systems |
Volume | 2018-December |
State | Published - 1 Jan 2018 |
Event | 32nd Conference on Neural Information Processing Systems, NeurIPS 2018 - Montreal, Canada Duration: 2 Dec 2018 → 8 Dec 2018 |
ASJC Scopus subject areas
- Computer Networks and Communications
- Information Systems
- Signal Processing