Skip to content

Latest commit

 

History

History
86 lines (52 loc) · 3.12 KB

README.rst

File metadata and controls

86 lines (52 loc) · 3.12 KB

Build Status codecov

This package is still under development. If you have any trouble running this code, please open an issue on GitHub.

DiCoDiLe

Package to run the experiments for the preprint paper Distributed Convolutional Dictionary Learning (DiCoDiLe): Pattern Discovery in Large Images and Signals.

Installation

All the tests should work with python >=3.6. This package depends on the python library numpy, matplotlib, scipy, mpi4py, joblib. The package can be installed with the following command run from the root of the package.

pip install  -e .

Or using the conda environment:

conda env create -f dicodile_env.yml

To build the doc use:

pip install  -e .[doc]
cd docs
make html

To run the tests:

pip install  -e .[test]
pytest .

Usage

All experiments are with mpi4py and will try to spawned workers depending on the parameters set in the experiments. If you need to use an hostfile to configure indicate to MPI where to spawn the new workers, you can set the environment variable MPI_HOSTFILE=/path/to/the/hostfile and it will be automatically detected in all the experiments. Note that for each experiments you should provide enough workers to allow the script to run.

All figures can be generated using scripts in benchmarks. Each script will generate and save the data to reproduce the figure. The figure can then be plotted by re-running the same script with the argument --plot. The figures are saved in pdf in the benchmarks_results folder. The computation are cached with joblib to be robust to failures.

Note

Open MPI tries to use all up network interfaces. This might cause the program to hang due to virtual network interfaces which could not actually be used to communicate with MPI processes. For more info Open MPI FAQ.

In case your program hangs, you can launch computation with the mpirun command:

  • either spefifying usable interfaces using --mca btl_tcp_if_include parameter:
$ mpirun -np 1 \
         --mca btl_tcp_if_include wlp2s0 \
         --hostfile hostfile \
         python -m mpi4py examples/plot_mandrill.py
  • or by excluding the virtual interfaces using --mca btl_tcp_if_exclude parameter:
$ mpirun -np 1 \
         --mca btl_tcp_if_exclude docker0 \
         --hostfile hostfile \
         python -m mpi4py examples/plot_mandrill.py

Alternatively, you can also restrict the used interface by setting environment variables OMPI_MCA_btl_tcp_if_include or OMPI_MCA_btl_tcp_if_exclude

$ export OMPI_MCA_btl_tcp_if_include="wlp2s0"

$ export OMPI_MCA_btl_tcp_if_exclude="docker0"``