Mutual Information : an Adequate Tool for Feature Selection ?
Résumé : In the field of machine learning, mutual information (MI) has been widely used as a multivariate criterion to select relevant features. However, we have shown in several of our recent works that selecting the feature subset which maximises MI is not necessarily optimal. This talk first introduces MI from an information theory point of view. Then, the interest and use of MI in machine learning is discussed. We will study under which conditions this criterion can be safely used in classification and regression, showing that mutual information is in general a good choice, even if counterexamples can be obtained. Both theoretical and experimental results will support the discussion.
Cet exposé se tiendra en salle C20-13, 20ème étage, Université Paris 1, Centre Pierre Mendès-France, 90 rue de Tolbiac, 75013 Paris (métro : Olympiades).