Bayesian Learning of Neural Network Architectures

View Researcher's Other Codes

Disclaimer: The provided code links for this paper are external links. Science Nest has no responsibility for the accuracy, legality or content of these links. Also, by downloading this code(s), you agree to comply with the terms of use as set out by the author(s) of the code(s).

Please contact us in case of a broken link from here

Authors Georgi Dikov, Patrick van der Smagt, Justin Bayer
Journal/Conference Name AISTATS 2019 - 22nd International Conference on Artificial Intelligence and Statistics
Paper Category
Paper Abstract In this paper we propose a Bayesian method for estimating architectural parameters of neural networks, namely layer size and network depth. We do this by learning concrete distributions over these parameters. Our results show that regular networks with a learnt structure can generalise better on small datasets, while fully stochastic networks can be more robust to parameter initialisation. The proposed method relies on standard neural variational learning and, unlike randomised architecture search, does not require a retraining of the model, thus keeping the computational overhead at minimum.
Date of publication 2019
Code Programming Language Jupyter Notebook
Comment

Copyright Researcher 2022