Skip to content

CVLab-TUDelft/pytorch-lightning-example

Repository files navigation

PyTorch Lightning Example

Robert-Jan Bruintjes

This repository serves as a starting point for any PyTorch-based Deep Computer Vision experiments. It uses PyTorch Lightning to power the training logic (including multi-GPU training), OmegaConf to provide a flexible and reproducible way to set the parameters of experiments, and Weights & Biases to log all experimental results and outputs.

Installation

CUDA: conda env create -f environment-cuda11.3.yml

CPU: conda env create -f environment.yml

Usage

Use command-line arguments to override the defaults given in config.yaml. For example:

python train.py wandb.log=True wandb.entity=<wandb-username> wandb.project=<wandb-project> wandb.experiment_name=<name-in-wandb> dataset.name=MNIST dataset.data_dir=./data dataset.channels=1 dataset.classes=10 model.name=LeNet

HPC: to run on the HPC, copy your code to the HPC, adapt the given run.sbatch to your HPC settings (see the top of the file) and use it by appending the Python call to the call to the sbatch file:

sbatch --partition general --qos short --time 4:00:00 -J name-in-slurm run.sbatch python train.py wandb.log=True wandb.entity=<wandb-username> wandb.project=<wandb-project> wandb.experiment_name=<name-in-wandb> dataset.name=MNIST dataset.data_dir=./data dataset.channels=1 dataset.classes=10 model.name=LeNet

GPUs: the code will automatically detect available GPUs and attempt to use them. When multiple GPUs are used each fits train.batch_size samples, so the total batch size is NUM_GPUs * train.batch_size. Consider this when tuning your hyperparameters!

Extending

If you end up extending this code to support generally useful functionality, or widely used datasets and/or models, consider a pull request to share your code!

Adding models

  • Add the code for the model in a new file in models;
  • Import & call the new model in model_factory.py

Adding datasets

  • Add the code for the dataset in a new file in datasets. Make sure to make methods for creating dataloaders for train and val/test.
  • Import & call the new methods in dataset_factory.py

Resuming training from a checkpoint

W&B saves checkpoints as "artifacts".

  • Use code like below to make W&B download the Runner checkpoint to disk:
artifact_name = f"{cfg.wandb.entity}/{project_name}/{artifact_name}"
print(artifact_name)
artifact = wandb_logger.experiment.use_artifact(artifact_name)
directory = artifact.download()
filename = os.path.join(directory, 'model.ckpt')
  • Add flag ckpt_path=filename to the call to Trainer.fit()
  • Consider generalizing this by making artifact_name given by a new config key cfg.resume.artifact

Known issues

Warnings related to num_workers: Depending on the amount of CPU threads available you may want to adapt cfg.dataset.num_workers for data loading to be more efficient.

Warnings related to sync_dist: PyTorch Lightning recommends to sync logging calls. As far as I know this doesn't affect the reported accuracies, but it may slow down training, so I chose to ignore these warnings.

Questions

If you have any questions about the use of this code, open an issue on the repository.

About

Example code for PyTorch experiments

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published