-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathabstracts.html
231 lines (227 loc) · 12.4 KB
/
abstracts.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
<!DOCTYPE html>
<html lang="en">
<head>
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
<title>VQA: Visual Question Answering</title>
<link rel="stylesheet" href="./static/css/foundation.css">
<link rel="stylesheet" href="./static/css/main.css">
</head>
<body class="off-canvas hide-extras" style="min-width:1200px; min-height:750px;">
<header>
<div class="row">
<a href="http://visualqa.org/"><img style="height: 100px; position:absolute; top:4px; left:0px;" src="./static/img/main.png" alt="logo" /></a>
<h1><img style="height: 90px;" src="./static/img/logo.png" alt="logo" /><br></h1>
<br>
</div>
</header>
<div class="contain-to-grid">
<nav class="top-bar" data-topbar>
<section class="top-bar-section">
<!-- Right Nav Section -->
<ul class="right">
<li><a href="index.html">Home</a></li>
<li><a href="people.html">People</a></li>
<li><a href="code.html">Code</a></li>
<li><a href="http://vqa.cloudcv.org/" onClick="ga('send', 'event', { eventCategory: 'Outgoing Link', eventAction: 'Demo', eventLabel: 'Demo'});">Demo</a></li>
<li class="has-dropdown"><a href="download.html">Download</a>
<ul class="dropdown">
<li><a href="download.html">VQA v2</a></li>
<li><a href="vqa_v1_download.html">VQA v1</a></li>
</ul>
</li>
<li><a href="evaluation.html">Evaluation</a></li>
<li class="has-dropdown"><a href="challenge.html">Challenge</a>
<ul class="dropdown">
<li><a href="challenge.html">2021</a></li>
<li><a href="challenge_2020.html">2020</a></li>
<li><a href="challenge_2019.html">2019</a></li><li><a href="challenge_2018.html">2018</a></li>
<li><a href="challenge_2017.html">2017</a></li>
<li><a href="challenge_2016.html">2016</a></li>
</ul>
</li>
<li class="has-dropdown"><a href="http://visualqa.org/vqa_v2_teaser.html">Browse</a>
<ul class="dropdown">
<li><a href="http://visualqa.org/vqa_v2_teaser.html">VQA v2</a></li>
<li><a href="https://vqabrowser.cloudcv.org/">VQA v1</a></li>
</ul>
</li>
<li><a href="http://visualqa.org/visualize/">Visualize</a></li>
<!-- <li class="has-dropdown"><a href="http://visualqa.org/visualize/">Visualize</a>
<ul class="dropdown">
<li><a href="http://visualqa.org/visualize/">VQA v2</a></li>
<li><a href="http://visualqa.org/visualize/">VQA v1</a></li>
</ul>
</li> -->
<li class="has-dropdown"><a href="workshop.html">Workshop</a>
<ul class="dropdown">
<li><a href="workshop.html">2021</a></li>
<li><a href="workshop_2020.html">2020</a></li>
<li><a href="workshop_2019.html">2019</a></li>
<li><a href="workshop_2018.html">2018</a></li>
<li><a href="workshop_2017.html">2017</a></li>
<li><a href="workshop_2016.html">2016</a></li>
</ul>
</li>
<li><a href="sponsors.html">Sponsors</a></li>
<li><a href="terms.html">Terms</a></li>
<li><a href="external.html">External</a></li>
</ul>
</section>
</nav>
</div>
<section role="main" style="padding: 1em;">
<div class="row">
<p style="font-size:30px; color:black; font-weight: 50" align=center>VQA Challenge and Visual Dialog Workshop
<br>
<span style="font-size:20px; color:black; font-weight: 400" align=center>Location: <b>Room 155A, Calvin L. Rampton Salt Palace Convention Center</b></span>
<br>
<span style="font-size:18px; color:gray; font-weight: 50" align=center>at CVPR 2018, June 18, Salt Lake City, Utah, USA</span></p>
<p style="font-size:20px; color:black; font-weight: 50" align=center><a href="workshop.html" style="padding:13px">Home</a> <a href="workshop.html#program" style="padding:13px">Program</a> <a href="workshop.html#sub" style="padding:13px">Submission</a><a href="abstracts.html" style="padding:13px">Accepted Abstracts</a>
</p>
<hr>
</div>
<div class="row">
<h1 style="font-size:30px; color:grey; font-weight: 200">Accepted Abstracts</h1>
<div class="large-12 columns" style="text-align:left;">
<p style="font-size:15px; font-weight: 200; text-align:left">
<b>Bilinear Attention Networks for Visual Question Answering</b>
<br> Jin-Hwa Kim, Jaehyun Jun, Byoung-Tak Zhang
<br>
<br>
<b>Two can play this Game: Visual Dialog with Discriminative Question Generation and Answering</b>
<br> Unnat Jain, Svetlana Lazebnik, Alexander Schwing
<br>
<br>
<b>Learning to Count Objects in Natural Images for Visual Question Answering</b>
<br> Yan Zhang, Jonathon Hare, Adam Prugel-Bennett
<br>
<br>
<b>Joint Image Captioning and Question Answering</b>
<br> Jialin Wu, Zeyuan Hu, Raymond J. Mooney
<br>
<br>
<b>Robustness Analysis of Visual QA Models by Basic Questions</b>
<br> Jia-Hong Huang, Cuong Duc Dao*, Modar Alfadly*, C. Huck Yang, Bernard Ghanem
<br>
<br>
<b> Mind Your Language: Learning Visually Grounded Dialog in a Multi-Agent Setting</b>
<br> Akshat Agarwal*, Swaminathan Gurumurthy*, Vasu Sharma*, Katia Sycara
<br>
<br>
<b>Dialog-based Interactive Image Retrieval</b>
<br> Hui Wu*, Xiaoxiao Guo*, Yu Cheng, Steven Rennie, Gerald Tesauro, Rogerio Schmidt Feris
<br>
<br>
<b>Cross Domain Normalization for Natural Language Object Retrieval</b>
<br> Asi Sheffer, Michael Elhadad
<br>
<br>
<b>DVQA: Understanding Data Visualizations via Question Answering</b>
<br> Kushal Kafle, Brian Price, Scott Cohen, Christopher Kanan
<br>
<br>
<b>Block Superdiagonal Fusion</b>
<br> Remi Cadene, Hedi Ben-younes, Matthieu Cord, Nicolas Thome
<br>
<br>
<b>Convolutional Image Captioning</b>
<br> Jyoti Aneja, Aditya Deshpande, Alexander Schwing
<br>
<br>
<b>Visual Question Answering System as a Teaching Aid</b>
<br> Sourav Suresh, Varun Nagaraj Rao, Gowri Srinivasa
<br>
<br>
<b>VizWiz Grand Challenge: Answering Visual Questions from Blind People</b>
<br> Danna Gurari, Qing Li, Abigale J. Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, Jeffrey P. Bigham
<br>
<br>
<b>Women also Snowboard: Overcoming Bias in Captioning Models</b>
<br> Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach
<br>
<br>
<b>Visual Question Answering as a Meta Learning Task</b>
<br> Damien Teney, Anton van den Hengel
<br>
<br>
<b>Improving Visual Question Answering Using Dropout and Enhanced Question Encoder</b>
<br> Zhiwei Fang, Jing Liu, Qu Tang, Yanyuan Qiao, Fei Liu, Yong Li, Hanqing Lu
<br>
<br>
<b>Focal Visual-Text Attention for Visual Question Answering</b>
<br> Junwei Liang, Lu Jiang, Liangliang Cao, Li-Jia Li, Alexander Hauptmann
<br>
<br>
<b>Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments</b>
<br> Peter Anderson, Qi Wu, Damien Teney, Jake Bruce, Mark Johnson, Niko Sunderhauf, Ian Reid, Stephen Gould, Anton van den Hengel
<br>
<br>
<b>Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering</b>
<br> Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, Lei Zhang
<br>
<br>
<b>CS-VQA: Visual Question Answering with Compressively Sensed Images</b>
<br> Li-Chi Huang, Kuldeep Kulkarni, Anik Jha, Suhas Lohit, Suren Jayasuriya, Pavan Turaga
<br>
<br>
<b>Bottom Up and Top Down Attention on Steroids: Architecture Tweaks, Learning Rate Schedules, and Ensembling</b>
<br> Yu Jiang*, Vivek Natarajan*, Xinlei Chen*, Marcus Rohrbach, Dhruv Batra, Devi Parikh
<br>
<br>
<b>Audio Visual Scene-Aware Dialog (AVSD) Challenge at DSTC7</b>
<br> Huda Alamri, Vincent Cartillier, Raphael Gontijo Lopes, Abhishek Das, Jue Wang, Irfan Essa, Dhruv Batra, Devi Parikh, Anoop Cherian, Tim K. Marks, Chiori Hori
<br>
<br>
<b>Answer Again: Multi-Answering Model for Visual Question Answering</b>
<br> Liang Peng, Yang Yang, Heng Tao Shen
<br>
<br>
<b>Dual Recurrent Attention Units for Visual Question Answering</b>
<br> Ahmed Osman, Wojciech Samek
<br>
<br>
<b> Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answering</b>
<br> Medhini G. Narasimhan, Alexander Schwing
<br>
<br>
<b> Evaluating Visual-Semantic Explanations using a Collaborative Image Guessing Game</b>
<br> Arijit Ray, Yi Yao, Avi Ziskind, Rakesh Kumar, Giedrius Burachas
</p>
</div>
<hr>
</div>
</section>
<script>
(function(i, s, o, g, r, a, m) {
i['GoogleAnalyticsObject'] = r;
i[r] = i[r] || function() {
(i[r].q = i[r].q || []).push(arguments)
}, i[r].l = 1 * new Date();
a = s.createElement(o),
m = s.getElementsByTagName(o)[0];
a.async = 1;
a.src = g;
m.parentNode.insertBefore(a, m)
})(window, document, 'script', '//www.google-analytics.com/analytics.js', 'ga');
ga('create', 'UA-63638588-1', 'auto');
ga('send', 'pageview');
</script>
<!-- jquery smooth scroll to id's -->
<script>
$(function() {
$('a[href*=#]:not([href=#])').click(function() {
if (location.pathname.replace(/^\//, '') == this.pathname.replace(/^\//, '') && location.hostname == this.hostname) {
var target = $(this.hash);
target = target.length ? target : $('[name=' + this.hash.slice(1) + ']');
if (target.length) {
$('html,body').animate({
scrollTop: target.offset().top
}, 1000);
return false;
}
}
});
});
</script>
</body>
</html>