IIIT-Delhi Institutional Repository

Exploiting independent visual and textual data sources to improve multi-modal methods for description and querying of visual data

Show simple item record

dc.contributor.author Pal, Ambar
dc.contributor.author Sharma, Gaurav (Advisor)
dc.contributor.author Arora, Chetan (Advisor)
dc.date.accessioned 2018-09-24T13:30:22Z
dc.date.available 2018-09-24T13:30:22Z
dc.date.issued 2016-07-18
dc.identifier.uri http://repository.iiitd.edu.in/xmlui/handle/123456789/673
dc.description.abstract Recent methods on combining textual and visual information using supervised (textual, visual) data have shown encouraging performance. However they are mostly limited to paired (textual, visual) data. We are interested in exploring methods which can leverage large, but independently annotated, datasets of visual and textual data. Applications include image and video captioning and, the induction of novel objects, wherein we try to describe objects that were not seen in the paired annotated data by harnessing knowledge from unpaired data . en_US
dc.language.iso en_US en_US
dc.publisher IIIT-Delhi en_US
dc.subject Image captioning en_US
dc.subject Recurrent neural nets en_US
dc.subject Supervised learning en_US
dc.subject Multi-modal methods en_US
dc.subject Image understanding en_US
dc.title Exploiting independent visual and textual data sources to improve multi-modal methods for description and querying of visual data en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository


Advanced Search

Browse

My Account