Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
kwotsin authored Jul 1, 2017
1 parent f272b80 commit 946f81f
Showing 1 changed file with 4 additions and 3 deletions.
7 changes: 4 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
# TensorFlow-ENet
TensorFlow implementation of [**ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation**](https://arxiv.org/pdf/1606.02147.pdf).

This model was tested on the CamVid dataset with street scenes taken from Cambridge, UK. For more information on this dataset, please visit: http://mi.eng.cam.ac.uk/research/projects/VideoRec/CamVid/
This model was tested on the CamVid dataset with street scenes taken from Cambridge, UK. For more information on this dataset, please visit: http://mi.eng.cam.ac.uk/research/projects/VideoRec/CamVid/.


## Visualizations
Expand Down Expand Up @@ -36,7 +36,8 @@ Execute `tensorboard --logdir=log` on your root directory to monitor your traini

- **dataset**: Contains 6 folders that holds the original train-val-test images and their corresponding ground truth annotations.

- **checkpoint**: The checkpoint directory that could be used for predicting the segmentation output. The model was trained using the default parameters mentioned in the paper, except that it uses median frequency balancing to obtain the class weights.
- **checkpoint**: The checkpoint directory that could be used for predicting the segmentation output. The model was trained using the default parameters mentioned in the paper, except that it uses median frequency balancing to obtain the class weights. The final checkpoint model size is under 5MB.

- **visualizations**: Contains the gif files that were created from the output of `predict_segmentation.py`.


Expand All @@ -54,7 +55,7 @@ Execute `tensorboard --logdir=log` on your root directory to monitor your traini
6. Once you've fine-tuned to get your best hyperparameters, there's an option to combine the training and validation datasets together. However, if your training dataset is large enough, this won't make a lot of difference.

## Implementation and Architectural Changes
1. By default, skip connections are added to connect the corresponding encoder and decoder portions for better performance.
1. Skip connections can be added to connect the corresponding encoder and decoder portions for better performance.

2. The number of initial blocks and the depth of stage 2 residual bottlenecks are tunable hyperparameters. This allows you to build a deeper network if required, since ENet is rather lightweight.

Expand Down

0 comments on commit 946f81f

Please sign in to comment.