Blog(2)
VQA [1,2] is the field of research that aims to develop methods for answering natural language questions based on the information provided in corresponding images.
The widespread adoption of mobile devices has led to a rapid growth of video content that is captured, transmitted and shared on various social media platforms.
Research Areas(0)
Publications(3)
LEARNING TO JOINTLY SHARE AND PRUNE WEIGHTS FOR GROUNDING BASED VISION AND LANGUAGE MODELS
AuthorShangqian Gao,Burak Uzkent,Yilin Shen,Hongxia Jin
PublishedInternational Conference on Learning Representation (ICLR)
Date2023-05-01
Progressive Attention Memory Network for Movie Story Question Answering
AuthorJunyeong Kim, Minuk Ma, Kyungsu Kim, Sungjin Kim, Chang D. Yoo
PublishedComputer Vision and Pattern Recognition (CVPR)
Date2019-06-21
Gaining Extra Supervision via Multi-task learning
PublishedInternational Joint Conference on Neural Networks (IJCNN)
Date2019-02-08
News(1)
Others(0)