Skip to content

A NN that generates captions word by word using an ensemble model composed of a VGG16, and 2 trained LSTMs.

Notifications You must be signed in to change notification settings

unnamedplay-r/image-captioning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Helping The Blind See

This project is aimed to deliver a prototype of an assistive technology tool using image captioning to help the blind understand the visual environment around them. This is done using an encoder-decoder neural network that builds captions word by word from a given image and associated caption.

Take a peek into the notebook within the repo. It contains alllll the details for creating, training and testing the model.

About

A NN that generates captions word by word using an ensemble model composed of a VGG16, and 2 trained LSTMs.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published