Skip to content

2g-XzenG/Food_caption

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Food_caption

In this project, I use the code from show attend and tell, which is implemented base on paper: show attend and tell

Since the goal is to predict the label of the image, and eventually hope one can generate the ingredient of the food image, so we think using the image caption generation technique is more appropriate than image classification. But due to the limit of the dataset, for now, I can only use the image label as the caption. Overall though, the result is not bad.

General Idea

  1. Extract Image feature by CNN (VGG-19)
  2. Use attention base LSTM model to generate the caption of the image, which in this case, is the label of the image

Result

alt tag

Note: The white part is the attention area that the machine focus on to make the caption/label prediction

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published