Efficient Model Selection for Regularized Classification by Exploiting Unlabeled DataReportar como inadecuado




Efficient Model Selection for Regularized Classification by Exploiting Unlabeled Data - Descarga este documento en PDF. Documentación en PDF para descargar gratis. Disponible también para leer online.

1 AMA - Analyse de données, Modélisation et Apprentissage automatique Grenoble LIG - Laboratoire d-Informatique de Grenoble 2 VISEO 3 Max Plank Institute for Intelligent Systems

Abstract : Hyper-parameter tuning is a resource-intensive task when optimizing classification models. The commonly used k-fold cross validation can become intractable in large scale settings when a classifier has to learn billions of parameters. At the same time, in real-world, one often encounters multi-class classification scenarios with only a few labeled examples; model selection approaches often offer little improvement in such cases and the default values of learners are used. We propose bounds for classification on accuracy and macro measures precision, recall, F1 that motivate efficient schemes for model selection and can benefit from the existence of unlabeled data. We demonstrate the advantages of those schemes by comparing them with k-fold cross validation and hold-out estimation in the setting of large scale classification.





Autor: Georgios Balikas - Ioannis Partalas - Eric Gaussier - Rohit Babbar - Massih-Reza Amini -

Fuente: https://hal.archives-ouvertes.fr/



DESCARGAR PDF




Documentos relacionados