Skip to content

jerpint/bigearthnet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

BigEarthNet ๐ŸŒŽ

Welcome to the bigearthnet classification repo!
This project was built in the context of an applied deep learning workshop in computer vision. This repo showcases modern tools and libraries used for applied deep learning. Accompanying slides and explanations can be found here.

For more details about the workshop itself (slides, content, etc.), view the Workshop Information section

Features

Here are a few of the features baked-in to this repo:

  • Pytorch-Lightning: Implements all the training loops and boilerplate code
  • Hydra: Easily manage and configure experiment parameters
  • TIMM: A model zoo of SOTA pre-trained classification models
  • Tensorboard: Logger used to track experiment progress
  • Deep Lake / Activeloop Hub: An efficient dataset/dataloader manager (think HDF5 but deep-learning centric).

The focus of this repository is centered around model training and evaluation. Deployment is not considered in this project.

Getting Started

Here are the basic steps for getting setup on most mochines:

1. Clone this project

git clone https://github.com/jerpint/bigearthnet

2. Install the dependencies:

It is recommended to work in a virtual environment (e.g. conda):

conda create -n bigearth python=3.8
conda activate bigearth

Once activated, run:

cd ~/bigearthnet/
pip install -e .

3. Test your installation

To test your install, simply run:

cd ~/bigearthnet/bigearthnet
python train.py

This will run an experiment with all the default configurations on a tiny bigearthnet subset. It will automatically download a small dataset and train a shallow baseline model end-to-end for 3 epochs. This should run on a CPU (<1 minute) and will ensure everything is properly installed.

Dataset

This project uses BigEarthNet Sentinel-2 Image Patches from bigearthnet. For more in-depth information about the original dataset, you can read the release paper. For convenience, the raw data has already been converted into Hub datasets which are provided with this repo. In these hub datasets, we are only considering bands 2,3,4 of the original spectral data, which (roughly) correspond to the B,G and R channels. To view how the data was prepared, head to /bigearthnet/data/scripts/.

3 versions of the dataset have been constructed using the BGR bands of the original dataset:

Name Size
Bigearthnet-mini 9.3 MB
Bigeathnet-medium 2.5 GB
Bigeartnet-full 30 GB
  • bigearthnet-mini: A tiny subset of the original data meant for debugging code and functionality. Used for running end-to-end tests during github actions. It is composed of 90 train samples, 30 validation samples and 30 test samples.

  • bigearthnet-medium: Composed of ~10% of the bigearthnert data. It is meant to be used to train models on a smaller scale and run vast hyper-parameter searches in reasonable time/compute. It is composed of 25 000 train samples, 5000 validation samples and 5000 test samples.

  • bigearthnet-full: The full dataset. It is composed of 269 695 train samples, 123 723 validation samples and 125 866 test samples. Splits were obtained from here

Getting the data

All data is hosted on a shared google drive.

You do not need to download the files manually. By default, the bigearthnet-mini dataset will automatically be downloaded and extracted to the datasets/ folder when first running an experiment.

To use another dataset version, simply overwrite the datamodule.dataset_name parameter. For example, to run an experiment using the bigearthnet-medium dataset, simply run:

python train.py ++datamodule.dataset_name=bigearthnet-medium

You can specify a different download folder by overriding the datamodule.dataset_dir directory.

Training Models

To train a model, simply run:

python train.py 

This will launch an experiment with all of the default parameters.

Overriding Parameters

This project uses hydra to manage configuration files.

A default configuration can be found under the configs directory.

To specify different parameters, you can simply override the appropriate parameter from the command line. For example, to train with a TIMM resnet34 pretrained model, a learning rate of 0.001 and adam optimizer, run the following:

python train.py model=timm ++model.model_name=resnet34 ++model.pretrained=true ++config.optimizer.name='adam' ++config.optimizer.lr=0.001 ++max_epochs=100

Hyper-parameter search

To perform hyper-parameter search, we can run a grid-search over common parameters using the --multirun hydra flag. For example, we will sweep a bunch of different learning rates:

python train.py --multirun ++config.optimizer.name='adam','sgd' ++config.optimizer.lr=0.1,0.01,0.001,0.0001

There exist many great tools for doing more advanced hyper-parameter searching, and hydra plugins easy ways to extend this support.

Tensorboard

After training models, you can view logged stats in tensorboard. By default, all experiments get saved under the outputs/ folder. Simply run

tensorboard --logdir outputs/

Then, open a browser and head to http://localhost:6006/ to view experiments.

This repo supports hyper parameter view under hparams tab, evolution of confusion matrices under images tab and profiling of the code under pytorch profiler tab.

A tensorboard summary of the best models can be found here

Model Evaluation

By default, all outputs can be found under the outputs/ folder. Each run will be timestamped, and contain a checkpoints directory with a last-model.ckpt and best-model.ckpt.

To evaluate a model on the test set, run the eval.py script while specifying the checkpoint of the best trained model, e.g.:

python eval.py --ckpt-path /path/to/best-model.ckpt

This will automatically load the model checkpoint and produce a summary of all the metrics on a given test set. Results of the evaluation will be saved where the script was run. You can specify which test set to evaluate on with the --dataset-name flag (by default it evaluates on bigearthnet-mini). This is useful for e.g. training on bigearthnet-medium and evaluating on bigearthnet-full test set.

python eval.py --ckpt-path /path/to/best-model.ckpt --dataset-name bigearthnet-full

For additional parameters such as speciying to do the evaluation on a gpu, run:

python eval.py --help for additional parameters.

Sample Notebook

You can view a sample notebook here. You can also follow the setup in the notebook to run models from within colab.

Note that viewing results within tensorboard won't be possible from colab, but you can download the outputs/ folder locally to view them. You can also use pre-trained models from within colab.

Pretrained models

A tensorboard summary of the best models can be found here

You can download the pre-trained models here:

Pre-trained ConvNext Pre-trained ViT

Workshop Information

The workshop consists of 4 virtual sessions, that were presented live. You can find the accompanying slides and materials here:

Session 1 - Introduction to image classification - theory

Tuesday, October 4, 2022 Duration: 2 hours

link to slides

link to pdf

Additional Material on Colab

Content:

  • Review what images are and how computers interpret them (grayscale vs. RGB images)
  • Review of concepts behind image classification in a supervised learning setting
  • Common neural network architectures and their underlying operations (Multilayer perceptrons, Convolutional neural networks)
  • Gradient descent intuitions and the role of the learning rate and optimizer
  • How to evaluate models, performance metrics, confusion matrices, precision, recall, etc.
  • Data Augmentation

Session 2 - Introduction to image classification - practical

Friday, October 7, 2022 Duration: 2 hours

Link to notebook (To complete)

Link to notebook (Solved)

Content:

This will be a hands-on, live-coding session. The task will be to implement a rock, paper, scissors game using pytorch. Simple models will be implemented and trained from scratch (e.g. MLP, LeNet). Training loops and evaluation routines will be implemented. Data splits will be considered, as well as concepts behind data augmentation and data distribution shifts.

Session 3 - DL Model Architectures - theory

Tuesday, October 11, 2022 Duration: 2 hours

link to slides

link to pdf

Content:

  • Overview of the evolution of state-of-the-art models on the ImageNet benchmark (VGG, ResNet)
  • Introduction to self-attention, transformers and vision transformers (ViT)
  • Comparisons (pros, cons, differences) between CNNs and ViT (Tradeoffs, benchmarks, speed, inference, model size etc.)
  • Foundation models + zero shot learning (CLIP)
  • Introduction to object detection and image segmentation

Session 4 - Implementation of a DL project - practical

Duration: 2 hours Friday, October 14, 2022

Link to slides

link to pdf

link to notebook

Summary: In this session, we will be implementing a remote-sensing task on the bigearthnet dataset (multi-label classification). We will implement a baseline and compare it to state-of-the-art models (both pre-trained and fine-tuned models). Concepts of fine-tuning pretrained models and hyper parameter tuning will be explained and implemented. We will focus on what goes into implementing a codebase for larger models/datasets as well as best practices surrounding the experiments - debugging, monitoring results, testing, reporting, checkpointing etc.