Skip to content

Latest commit

 

History

History
11 lines (5 loc) · 730 Bytes

README.md

File metadata and controls

11 lines (5 loc) · 730 Bytes

Loss Landscapes are All You Need: Neural Network Generalization Can Be Explained Without the Implicit Bias of Gradient Descent

This repository trains large number of models in parallel with non-gradient based optimizers.

To set up the environment, you could use conda with conda env create -f environment.yml

All scripts for reproducing the tables in the paper "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explained Without the Implicit Bias of Gradient Descent") - ICLR 2023 can be found in ./scripts.

train_distributed.py trains models in parallel on different host and then save the resulting metrics in a single shared sqllite database.