Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Regularity Normalization: Neuroscience-Inspired Unsupervised Attention across Neural Network Layers release_2m6wl4xsrbeixiysjgwgkld2xu [as of editgroup_uv4hewwicre4xep2oev4flg5cy]

by Baihan Lin

Published in Entropy by MDPI AG.

2021   Volume 24, p59

Abstract

Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (RN) as an unsupervised attention mechanism (UAM) which computes the statistical regularity in the implicit space of neural networks under the Minimum Description Length (MDL) principle. Treating the neural network optimization process as a partially observable model selection problem, the regularity normalization constrains the implicit space by a normalization factor, the universal code length. We compute this universal code incrementally across neural network layers and demonstrate the flexibility to include data priors such as top-down attention and other oracle information. Empirically, our approach outperforms existing normalization methods in tackling limited, imbalanced and non-stationary input distribution in image classification, classic control, procedurally-generated reinforcement learning, generative modeling, handwriting generation and question answering tasks with various neural network architectures. Lastly, the unsupervised attention mechanisms is a useful probing tool for neural networks by tracking the dependency and critical learning stages across layers and recurrent time steps of deep networks.
In application/xml+jats format

Type  article-journal
Stage   published
Date   2021-12-28
Language   en ?
Container Metadata
Open Access Publication
In DOAJ
In ISSN ROAD
In Keepers Registry
ISSN-L:  1099-4300
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Work In Progress

This entity has not been "accepted" into the official database yet.

Catalog Record
State: wip
Revision: 534a5432-c0e1-45e8-b545-657a6259b93b
API URL: JSON