VQA Workshop 2021 (original) (raw)

logo

Visual Question Answering Workshop
at CVPR 2021, June 19

Zoom and Gatherly links on CVPR 2021 website: Link
Navigate to: Workshops -> Sat, June 19 -> Search for "Visual Question Answering" workshop entry

Ended Recording: [Video] Panel-1: Future Directions Vittorio Ferrari, Damien Teney, Raquel Fernández, Aida Nematzadeh, Olga RussakovskyHosted on Zoom, joining link on the internal CVPR website Starts June 19, 9 AM PT!
Ended Live QA-1 Individual live QA for challenge related talks and poster spotlight presenters.Hosted on Gatherly, joining link on the internal CVPR website Starts June 19, 12 PM PT!
Ended Recording: [Video] Panel-2: Future Directions Justin Johnson, He He, Mohit Bansal, Katerina Fragkiadaki, Anirudh KoulHosted on Zoom, joining link on the internal CVPR website Starts June 19, 3 PM PT!
Live QA-2 Individual live QA for challenge related talks and poster spotlight presenters.Hosted on Gatherly, joining link on the internal CVPR website Starts June 20, 12 AM PT!

Home Program Poster Spotlights


Introduction

The primary goal of this workshop is two-fold. First is to benchmark progress in Visual Question Answering.

There will be three tracks in the Visual Question Answering Challenge this year.

The second goal of this workshop is to continue to bring together researchers interested in visually-grounded question answering, dialog systems, and language in general to share state-of-the-art approaches, best practices, and future directions in multi-modal AI. In addition to invited talks from established researchers, we invite submissions of extended abstracts of at most 2 pages describing work in the relevant areas including: Visual Question Answering, Visual Dialog, (Textual) Question Answering, (Textual) Dialog Systems, Commonsense knowledge, Vision + Language, etc. The submissions are not specific to any challenge track. All accepted abstracts will be presented as posters at the workshop to disseminate ideas. The workshop is on June 19, 2021, at the IEEE Conference on Computer Vision and Pattern Recognition, 2021.


Invited Speakers

He He
New York University


Program

Live

June 19, 2021 9 AM - 10 AM PT Panel-1: Future Directions Vittorio Ferrari, Damien Teney, Raquel Fernández, Aida Nematzadeh, Olga Russakovsky Hosted on Zoom, joining link on the internal CVPR website: Link (Panel ended) Recording available now: [Video]
June 19, 2021 12 PM - 1 PM PT Live QA-1 Individual live QA for challenge related talks and poster spotlight presenters. Hosted on Gatherly, joining link on the internal CVPR website: Link
June 19, 2021 3 PM - 4 PM PT Panel-2: Future Directions Justin Johnson, He He, Mohit Bansal, Katerina Fragkiadaki, Anirudh Koul Hosted on Zoom, joining link on the internal CVPR website: Link (Panel ended) Recording available now: [Video]
June 20, 2021 12 AM - 1 AM PT Live QA-2 Individual live QA for challenge related talks and poster spotlight presenters. Hosted on Gatherly, joining link on the internal CVPR website: Link

Prerecorded

| | | Welcome Aishwarya Agrawal (University of Montreal / Mila / Deepmind)[Video] [Slides] | | | ------------------------------------------------------------------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | | To ask questions, join Panel-1 | | Invited Talk Title: Visual & Conversational Saliency Raquel Fernández (University of Amsterdam) [Video] [Slides] | | To ask questions, join Panel-1 | | Invited Talk Title: Connecting Vision and Language with Localized Narratives and Open Images Vittorio Ferrari (Google) [Video] [Slides] | | To ask questions, join Panel-2 | | Invited Talk Title: Towards Overcoming the Language Prior in VQA He He (New York University) [Video] [Slides] | | To ask questions, join Panel-2 | | Invited Talk Title: How AI Can Empower The Blind Community Anirudh Koul (Pinterest) [Video] [Slides] | | To ask questions, join Panel-1 | | Invited Talk Title: Models, metrics, tasks and fairness in vision and language Olga Russakovsky (Princeton University) [Video] [Slides] | | To ask questions, join Live QA-1 or Live QA-2 | | VQA Challenge Talk (Overview, Analysis and Winner Announcement) Ayush Shrivastava (Georgia Tech) [Video] [Slides] Poster ID: 12 | | To ask questions, join Live QA-1 or Live QA-2 | | VQA Challenge Winner Talk Team: AliceMindMembers: Ming Yan, Haiyang Xu, Chenliang Li , Junfeng Tian, Wei Wang, Bin Bi, Zheng Cao, Ji Zhang, Songfang Huang, Fei Huang, Luo SiAffiliation: Alibaba Group [Video] [Slides] Poster ID: 13 | | To ask questions, join Live QA-1 or Live QA-2 | | TextVQA Challenge Talk (Overview, Analysis and Winner Announcement) Yash Kant (Georgia Tech) [Video] [Slides] Poster ID: 14 | | To ask questions, join Live QA-1 or Live QA-2 | | TextVQA Challenge Winner Members: Yixuan Qiao, Hao Chen, Jun Wang, Xianbin Ye, Ziliang Li, Peng Gao, Guotong Xie [Paper] Poster ID: 15 | | To ask questions, join Live QA-1 or Live QA-2 | | Poster Spotlights To watch the poster spotlights, visit: https://visualqa.org/posters_2021. | | To ask questions, join Live QA-1 or Live QA-2 | | TextCaps Challenge Talk (Overview, Analysis and Winner Announcement) Amanpreet Singh (Facebook AI Research) [Video] [Slides] Poster ID: 16 | | To ask questions, join Live QA-1 or Live QA-2 | | TextCaps Challenge Winner Talk Members: Zhengyuan Yang, Jianfeng Wang, Xiaowei Hu, Zhe Gan, Lijuan Wang, Yijuan Lu, Dinei Florencio, Cha Zhang, Jiebo Luo, Zicheng LiuAffiliations: Microsoft, University of Rochester [Video] [Slides] Poster ID: 17 | | To ask questions, join Panel-1 | | Invited Talk Title: Towards Better Multimodal Pretraining Aida Nematzadeh (DeepMind) [Video] [Slides] | | To ask questions, join Panel-2 | | Invited Talk Title: Learning Visual Representations from Language Justin Johnson (University of Michigan) [Video] [Slides] | | To ask questions, join Panel-1 | | Invited Talk Title: Visual question answering and the limits of statistical learning. Are we building a ladder to the moon? Damien Teney (Idiap Research Institute) [Video] [Slides] | | To ask questions, join Panel-2 | | Invited Talk Title: Knowledgeable & Spatial-Temporal Vision+Language Mohit Bansal (UNC Chapel Hill) [Video] [Slides] | | To ask questions, join Panel-2 | | Invited Talk Title: Augment Machine Intelligence with Multimodal Information Katerina Fragkiadaki (Carnegie Mellon University) [Video] [Slides] | | | | Closing Remarks Aishwarya Agrawal (University of Montreal / Mila / Deepmind) [Video] [Slides] | |


Submission Instructions

We invite submissions of extended abstracts of at most 2 pages (excluding references) describing work in areas such as: Visual Question Answering, Visual Dialog, (Textual) Question Answering, (Textual) Dialog Systems, Commonsense knowledge, Video Question Answering, Video Dialog, Vision + Language, and Vision + Language + Action (Embodied Agents). Accepted submissions will be presented as posters at the workshop. The extended abstract should follow the CVPR formatting guidelines and be emailed as a single PDF to the email id mentioned below.

Dual Submissions
We encourage submissions of relevant work that has been previously published, or is to be presented at the main conference. The accepted abstracts will not appear in the official IEEE proceedings.

Where to Submit?
Please send your abstracts to visualqa.workshop@gmail.com


Dates

March 2021 Challenge Announcements May 21, 2021Extended Workshop Paper Submission mid-May 2021 Challenge Submission Deadlines May 28, 2021 Notification to Authors June 19, 2021 Workshop


Organizers

Ayush Shrivastava
Georgia Tech

Yash Kant
Georgia Tech

Sashank Gondala
Georgia Tech

Satwik Kottur
Facebook AI

Dhruv Batra
Georgia Tech / Facebook AI Research

Devi Parikh
Georgia Tech / Facebook AI Research

Aishwarya Agrawal
University of Montreal / Mila / Deepmind


Sponsors

This work is supported by grants awarded to Dhruv Batra and Devi Parikh.