dc.contributor.author | Pal, Ambar | |
dc.contributor.author | Sharma, Gaurav (Advisor) | |
dc.contributor.author | Arora, Chetan (Advisor) | |
dc.date.accessioned | 2018-09-24T13:30:22Z | |
dc.date.available | 2018-09-24T13:30:22Z | |
dc.date.issued | 2016-07-18 | |
dc.identifier.uri | http://repository.iiitd.edu.in/xmlui/handle/123456789/673 | |
dc.description.abstract | Recent methods on combining textual and visual information using supervised (textual, visual) data have shown encouraging performance. However they are mostly limited to paired (textual, visual) data. We are interested in exploring methods which can leverage large, but independently annotated, datasets of visual and textual data. Applications include image and video captioning and, the induction of novel objects, wherein we try to describe objects that were not seen in the paired annotated data by harnessing knowledge from unpaired data . | en_US |
dc.language.iso | en_US | en_US |
dc.publisher | IIIT-Delhi | en_US |
dc.subject | Image captioning | en_US |
dc.subject | Recurrent neural nets | en_US |
dc.subject | Supervised learning | en_US |
dc.subject | Multi-modal methods | en_US |
dc.subject | Image understanding | en_US |
dc.title | Exploiting independent visual and textual data sources to improve multi-modal methods for description and querying of visual data | en_US |
dc.type | Other | en_US |