Abstract
We examine the learning behavior of a shallow and a deep convolutional neural network performing classification tasks on subsets of two databases. Our investigation focuses on the label, the input, and the prediction layer, and we compute the mutual information between these layers epoch-wise using Rényi’s matrix-based entropy functional. We evaluate the data processing inequality to interpret the learning behavior in a consistent information-theoretic framework. Our primary goals are to 1) clarify the relation between the two training objectives of minimizing the cross-entropy and maximizing the mutual information between the label and the prediction layer, 2) gradually switch from the first to the second training objective, and 3) interpret the impact of the latter transition. One of the main contributions is the proposed novel method for regularizing the cross-entropy objective and assessing the neural network’s learning activity.
Original language | English |
---|---|
Publication status | Published - 2024 |
Event | WIVACE 2024 XVIII International Workshop on Artificial Life and Evolutionary Computation - University of Namur, Namur, Belgium Duration: 11 Sept 2024 → 13 Sept 2024 https://events.info.unamur.be/wivace2024/ |
Conference
Conference | WIVACE 2024 XVIII International Workshop on Artificial Life and Evolutionary Computation |
---|---|
Abbreviated title | WIVACE 2024 |
Country/Territory | Belgium |
City | Namur |
Period | 11/09/24 → 13/09/24 |
Internet address |
Keywords
- Neural networks, Rényi’s entropy functional, Data processing inequality, InfoMax