Résumé
Linear data transformations constitute essential operations in various machine learning algorithms, ranging from linear regression up to adaptive metric transformation. Often, linear scalings are not only used to improve the model accuracy, rather feature coefficients as provided by the mapping are interpreted as an indicator for the relevance of the feature for the task at hand. This principle, however, can be misleading in particular for high-dimensional or correlated features, since it easily marks irrelevant features as relevant or vice versa. In this contribution, we propose a mathematical formalisation of the minimum and maximum feature relevance for a given linear transformation which can efficiently be solved by means of linear programming. We evaluate the method in several benchmarks, where it becomes apparent that the minimum and maximum relevance closely resembles what is often referred to as weak and strong relevance of the features; hence unlike the mere scaling provided by the linear mapping, it ensures valid interpretability.
langue originale | Anglais |
---|---|
titre | IEEE SSCI 2014 - 2014 IEEE Symposium Series on Computational Intelligence - CIDM 2014: 2014 IEEE Symposium on Computational Intelligence and Data Mining, Proceedings |
Editeur | Institute of Electrical and Electronics Engineers Inc. |
Pages | 149-156 |
Nombre de pages | 8 |
ISBN (imprimé) | 9781479945191 |
Les DOIs | |
Etat de la publication | Publié - 13 janv. 2015 |
Evénement | 5th IEEE Symposium on Computational Intelligence and Data Mining, CIDM 2014 - Orlando, États-Unis Durée: 9 déc. 2014 → 12 déc. 2014 |
Une conférence
Une conférence | 5th IEEE Symposium on Computational Intelligence and Data Mining, CIDM 2014 |
---|---|
Pays/Territoire | États-Unis |
La ville | Orlando |
période | 9/12/14 → 12/12/14 |