Abstract
One approach to interpreting multidimensional scaling (MDS)
embeddings is to estimate a linear relationship between the MDS dimen-
sions and a set of external features. However, because MDS only preserves
distances between instances, the MDS embedding is invariant to rotation.
As a result, the weights characterizing this linear relationship are arbitrary
and difficult to interpret. This paper proposes a procedure for selecting
the most pertinent rotation for interpreting a 2D MDS embedding.
embeddings is to estimate a linear relationship between the MDS dimen-
sions and a set of external features. However, because MDS only preserves
distances between instances, the MDS embedding is invariant to rotation.
As a result, the weights characterizing this linear relationship are arbitrary
and difficult to interpret. This paper proposes a procedure for selecting
the most pertinent rotation for interpreting a 2D MDS embedding.
Original language | English |
---|---|
Title of host publication | 26th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning |
Subtitle of host publication | ESANN 2018 : Bruges, Belgium, April 25, 26, 27, 2018 |
Editors | Michel Verleysen |
Place of Publication | Louvain-la-Neuve |
Publisher | CIACO |
Pages | 537-542 |
ISBN (Electronic) | 9782875870476 |
Publication status | Published - 1 Jan 2018 |
Event | 26th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2018) - Bruges, Bruges, Belgium Duration: 25 Apr 2018 → 27 Apr 2018 |
Conference
Conference | 26th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN 2018) |
---|---|
Country/Territory | Belgium |
City | Bruges |
Period | 25/04/18 → 27/04/18 |
Keywords
- Machine learning
- Interpretability
- Dimensionality reduction
- Multidimensional scaling
- Multi-view
- Sparsity
- Lasso regularization
Fingerprint
Dive into the research topics of 'Finding the Most Interpretable MDS Rotation for Sparse Linear Models based on External Features'. Together they form a unique fingerprint.Student theses
-
Interpretability and Explainability in Machine Learning and their Application to Nonlinear Dimensionality Reduction
Bibal, A. (Author)FRENAY, B. (Supervisor), VANHOOF, W. (President), Cleve, A. (Jury), Dumas, B. (Jury), Lee, J. A. (Jury) & Galarraga, L. (Jury), 16 Nov 2020Student thesis: Doc types › Doctor of Sciences
File