TY - GEN
T1 - Learning from a mixture of labeled and unlabeled examples with parametric side information
AU - Ratsaby, Joel
AU - Venkatesh, Santosh S.
N1 - Publisher Copyright:
© 1995 ACM.
PY - 1995/7/5
Y1 - 1995/7/5
N2 - We investigate the tradeoff between labeled and unlabeled sample complexities in learning a classification rule for a parametric two-class problem. In the problem considered, a sample of m labeled examples and n unlabeled examples generated from a two-class, N-variate Gaussian mixture is provided together with side information specifying the parametric form of the probability densities. The class means and a priori class probabilities are, however, unknown parameters. In this framework we use the maximum likelihood estimation method to estimate the unknown parameters and utilize rates of convergence of uniform strong laws to determine the tradeoff between error rate and sample complexity. In particular, we show that for the algorithm used, the misclassification probability deviates from the minimal Bayes error rate by O(N3/5n-1/5) + O(e-cm) where N is the dimension of the feature space, m is the number of labeled examples, n is the number of unlabeled examples, and c is a positive constant.
AB - We investigate the tradeoff between labeled and unlabeled sample complexities in learning a classification rule for a parametric two-class problem. In the problem considered, a sample of m labeled examples and n unlabeled examples generated from a two-class, N-variate Gaussian mixture is provided together with side information specifying the parametric form of the probability densities. The class means and a priori class probabilities are, however, unknown parameters. In this framework we use the maximum likelihood estimation method to estimate the unknown parameters and utilize rates of convergence of uniform strong laws to determine the tradeoff between error rate and sample complexity. In particular, we show that for the algorithm used, the misclassification probability deviates from the minimal Bayes error rate by O(N3/5n-1/5) + O(e-cm) where N is the dimension of the feature space, m is the number of labeled examples, n is the number of unlabeled examples, and c is a positive constant.
UR - http://www.scopus.com/inward/record.url?scp=84947134568&partnerID=8YFLogxK
U2 - 10.1145/225298.225348
DO - 10.1145/225298.225348
M3 - Conference contribution
AN - SCOPUS:84947134568
T3 - Proceedings of the 8th Annual Conference on Computational Learning Theory, COLT 1995
SP - 412
EP - 417
BT - Proceedings of the 8th Annual Conference on Computational Learning Theory, COLT 1995
PB - Association for Computing Machinery, Inc
T2 - 8th Annual Conference on Computational Learning Theory, COLT 1995
Y2 - 5 July 1995 through 8 July 1995
ER -