This repository hosts a notebook that extends the research presented in "Training neural networks using Metropolis Monte Carlo and an adaptive variant" by Whitelam et al. (2022), accessible at arXiv:2205.07408. This notebook was developed by Giosuè Castellano, Alan Picucci and Fabio Pernisi (myself).
The referenced paper investigates the application of the zero-temperature Metropolis Monte Carlo method for neural network training. This approach aims to minimize a loss function through an algorithm that initially resembles Gradient Descent. In the original version, as described by Whitelam and colleagues, the algorithm updates each network weight
Here,
The paper introduces an adaptive version of the Monte Carlo algorithm (aMC), which proposes weight updates in a more nuanced manner:
In this adaptive approach, both the mean
Here, n_reset
in the code), the following updates occur:
Initially, the step-size
In this project, we delve into exploring a non-zero-temperature variant of the aMC algorithm. Our aim is to understand its implications and effectiveness in neural network training under different temperature settings.
Note: This README is intended to provide a brief overview of our work in this repository. For a comprehensive understanding, please refer to the original paper and the detailed documentation within the code.