Show simple item record

dc.contributor.authorHaasbroek, Daniël G.
dc.contributor.authorDavel, Marelie H.
dc.date.accessioned2021-02-26T14:01:08Z
dc.date.available2021-02-26T14:01:08Z
dc.date.issued2020
dc.identifier.isbn978-0-620-89373-2
dc.identifier.urihttp://hdl.handle.net/10394/36796
dc.description.abstractEach node in a neural network is trained to activate for a specific region in the input domain. Any training samples that fall within this domain are therefore implicitly clustered together. Recent work has highlighted the importance of these clusters during the training process but has not yet investigated their evolution during training. Towards this goal, we train several ReLU-activated MLPs on a simple classification task (MNIST) and show that a consistent training process emerges: (1) sample clusters initially increase in size and then decrease as training progresses, (2) the size of sample clusters in the first layer decreases more rapidly than in deeper layers, (3) binary node activations, especially of nodes in deeper layers, become more sensitive to class membership as training progresses, (4) individual nodes remain poor predictors of class membership, even if accurate when applied as a group. We report on the detail of these findings and interpret them from the perspective of a high-dimensional clustering process.en_US
dc.language.isoenen_US
dc.publisherSouthern African Conference for Artificial Intelligence Researchen_US
dc.subjectNeural networksen_US
dc.subjectGeneralizationen_US
dc.subjectClusteringen_US
dc.titleExploring neural network training dynamics through binary node activationsen_US
dc.typeArticleen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record