-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathindex.html
More file actions
333 lines (253 loc) · 18.4 KB
/
index.html
File metadata and controls
333 lines (253 loc) · 18.4 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
---
layout: default
---
<section style="text-align: justify">
<h1> Visual Domain Adaptation Challenge </h1>
</br>
[<a href = "#news">News</a>]
[<a href = "#overview">Overview</a>]
[<a href = "#winners">Winners</a>]
[<a href = "#browse">Browse Data</a>]
[<a href = "#download">Download</a>]
[<a href = "#evaluation">Evaluation</a>]
[<a href = "#rules">Rules</a>]
[<a href = "#faq">FAQ</a>]
[<a href = "#workshop">TASK-CV Workshop</a>]
[<a href = "#organizers">Organizers</a>]
[<a href = "#sponsors">Sponsors</a>]
</br> </br> </nr>
<a name = "news"></a>
<h2 class="section-title"> News </h2>
<p>
The 2018 VisDA challenge is live and open for registration! It features a larger, more diverse dataset and two new competition tracks--open-set object classification and object detection. Vist the challenge <a href='http://ai.bu.edu/visda-2018/'>website</a> for more details and to register! In the meantime, the 2017 <a href="#evaluation">test server</a> is still up and running if you wish to submit to it.</p>
<ul>
<li><b>November 1</b> TASK-CV workshop presentations available. </li>
<li><b>October 10</b> Winners announcement! </li>
<li><b>September 30</b> Thank you to all who participated in the challenge! Stay tuned for the winners announcement. </li>
<li><b>September 10</b> Prizes announced! See <a href = "#prizes">here</a> for details.</li>
<li><b>September 8</b> Testing data released</li>
<li><b>July 13</b> Our competition is featured on <a href = "https://competitions.codalab.org/"> CodaLab</a>! </li>
<li><b>June 23</b> Evaluation servers are open for submission</li>
<li><b>June 22</b> The deadline for abstract submission to the <a href = "http://adas.cvc.uab.es/task-cv2017/">TASK-CV</a> workshop has been extended</li>
<li><b>June 19</b> Training data, validation data and DevKits released</li>
</ul>
</br>
<a name = "winners"></a>
<h2 class="section-title"> Winners of VisDA-2017 challenge </h2>
<div class="summary">
<b style="font-size: 20px;"> Classification </b> <a href="https://competitions.codalab.org/competitions/17052#results"> [full leaderboard]</a>
</br> </br>
<table style="width:100%" class="table">
<tr class="active">
<th width='3%'>#</th>
<th width='25%'>Team Name</th>
<th>Affiliation</th>
<th>Score</th>
</tr>
<tr>
<td> 1 </td>
<td>GF_ColourLab_UEA</td>
<td>University of East Anglia, Colour Lab <a href="https://arxiv.org/abs/1706.05208">[arxiv]</a>
<a href="https://github.com/Britefury/self-ensemble-visual-domain-adapt-photo/"> [code]</a> </td>
<td>92.8</td>
</tr>
<tr>
<td> 2 </td>
<td>NLE_DA</td>
<td>NAVER LABS Europe</td>
<td>87.7</td>
</tr>
<tr>
<td> 3 </td>
<td>BUPT_OVERFIT</td>
<td>Beijing University of Posts and Telecommunications</td>
<td>85.4</td>
</tr>
</table>
</br>
<b style="font-size: 20px;"> Segmentation </b> <a href="https://competitions.codalab.org/competitions/17054#results"> [full leaderboard]</a>
</br> </br>
<table style="width:100%" class="table">
<tr class="active">
<th width='3%'>#</th>
<th width='25%'>Team Name</th>
<th>Affiliation</th>
<th>Score</th>
</tr>
<tr>
<td> 1 </td>
<td>RTZH <a href="https://openaccess.thecvf.com/content_cvpr_2018/papers/Zhang_Fully_Convolutional_Adaptation_CVPR_2018_paper.pdf">[pdf]</a></td>
<td>Microsoft Research Asia</td>
<td>47.5</td>
</tr>
<tr>
<td> 2 </td>
<td> _piotr_</td>
<td>University of Oxford, Active Vision Laboratory</td>
<td>44.7</td>
</tr>
<tr>
<td> 3 </td>
<td> whung </td>
<td>University of California, Merced, Vision and Learning Lab</td>
<td>42.4</td>
</tr>
</table>
</br>
<p>Thank you to those who attended TASK-CV and made it a success! Below please find the VisDA challenge presentations from workshop organizers and from the top performing teams in each track.</p>
<ul>
<li><a href="{{site.baseurl}}/assets/attachments/VisDA_2017.pdf">VisDA organizers presentation</li>
<li><a href="{{site.baseurl}}/assets/attachments/VisDA_NaverLabs.pdf">Classification track runner-up presentation</li>
<li><a href="{{site.baseurl}}/assets/attachments/VisDA_ColourLab.pdf">Classification track winner presentation</li>
<li><a href="{{site.baseurl}}/assets/attachments/VisDA_VLLab.pdf">Segmentation track honorable mention presentation</li>
<li><a href="{{site.baseurl}}/assets/attachments/VisDA_MSM_at_MSRA.pdf">Segmentation track winner presentation</li>
</ul>
</br>
<a name = "overview"></a>
<h2 class="section-title"> Overview </h2>
<div class="summary">
<p> We are pleased to announce the 2017 Visual Domain Adaptation (VisDA2017) Challenge! It is well known that the success of machine learning methods on visual recognition tasks is highly dependent on access to large labeled datasets. Unfortunately, performance often drops significantly when the model is presented with data from a new deployment domain which it did not see in training, a problem known as <i>dataset shift</i>. The VisDA challenge aims to test domain adaptation methods’ ability to transfer source knowledge and adapt it to novel target domains. </p>
</br>
<figure>
<img class="img-responsive" src="{{site.baseurl}}/assets/images/bad-example.png"/>
</figure>
<p><strong>Caption:</strong> An example of a deep learning model failing to properly segment the scene into semantic categories such as road (green), building (gray), etc., because the input image looks different from its training data. Unsupervised Domain Adaptation methods aim to use labeled samples from the train domain and large volumes of unlabeled samples from the test domain to reduce a prediction error on the test domain. [1]</p>
</br>
<p> The competition will take place during the months of June -- September 2017, and the top performing teams will be invited to present their results at the <a href="http://adas.cvc.uab.es/task-cv2017/">TASK-CV</a> workshop at <a href="http://iccv2017.thecvf.com/">ICCV 2017 in Venice, Italy</a>. This year’s challenge focuses on synthetic-to-real visual domain shifts and includes two tracks: </p>
<ul>
<li><a href = "#classification">image classification</a></li>
<li><a href = "#segmentation">image segmentation</a></li>
</ul>
<p> Participants are welcome to enter in one or both tracks. </p>
</br>
<a name = "classification"></a>
<h2 class="section-title"> Classification Track </h2>
<p> In this challenge, the goal is to develop a method of unsupervised domain adaptation for image classification. Participants will be given three datasets, each containing the same object categories: </p>
<ul>
<li> <b> training domain (source)</b>: synthetic 2D renderings of 3D models generated from different angles and with different lighting conditions </li>
<li> <b> validation domain (target)</b>: a photo-realistic or real-image validation domain that participants can use to evaluate performance of their domain adaptation methods </li>
<li> <b> test domain (target)</b>: a new real-image test domain, <b> different from the validation domain and without labels </b>. The test set will be released shortly before the end of the competition </li>
</ul>
<p> The reason for using different target domains for validation and test is to evaluate the performance of proposed models as an out-of-the-box domain adaptation tool. This setting more closely mimics realistic deployment scenarios where the target domain is unknown at training time and discourages algorithms that are designed to handle a particular target domain. </p>
<figure>
<img class="img-responsive" src="{{site.baseurl}}/assets/images/classification-shift.png"/>
</figure>
</br>
<a name = "segmentation"></a>
<h2 class="section-title"> Segmentation Track </h2>
<p> In this challenge, the goal is to develop an algorithm that can adapt between synthetic dashcam views and real dashcam footage for the semantic image segmentation task. The training data will include pixel-level semantic annotations for 19 classes. We will also provide validation and testing data, following same protocol: </p>
<ul>
<li> <b>training domain (source)</b>: synthetic dashcam renderings along with semantic segmentation labels for 19 classes</li>
<li> <b>validation domain (target)</b>: a real world collection of dashcam images along with semantic segmentation labels for the corresponding 19 classes to be used for validating the unsupervised semantic segmentation performance</li>
<li> <b>test domain (target)</b>: a collection of real world dashcam images, <b>different from the validation domain and without labels</b>. The test set will be released shortly before the end of the competition </li>
</ul>
<figure>
<img class="img-responsive" src="{{site.baseurl}}/assets/images/semantic-shift.png" style="display: block; margin: auto;"/>
</figure>
</br>
<a name = "prizes"></a>
<h2 class="section-title"> Prizes </h2>
<p> We are excited to announce prizes for the top performing teams! The winners in both the classification and segmentation tracks will receive: </p>
<ul>
<li> <b>1st place:</b> $2,000 </li>
<li> <b>2nd place:</b> $500 </li>
<li> <b>3rd place:</b> $250 </li>
</ul>
</br>
<a name = "browse"></a>
<h2 class="section-title"> Browse Data </h2>
<p > If you would like to browse sample images from each of the datasets currently available please use the links below: </p>
<br>
<p> Classification challenge: </p>
<ul>
<li><a href="pages/classification_train/aeroplane.html">training domain</a></li>
<li><a href="pages/classification_validation/aeroplane.html">validation domain</a></li>
</ul>
<p> Segmentation challenge: </p>
<ul>
<li><a href = "pages/segmentation_train/segmentation_train.html">training domain</a></li>
<li><a href = "pages/segmentation_validation/segmentation_validation.html">validation domain</a></li>
</ul>
</br>
<a name = "download"></a>
<h2 class="section-title"> Download </h2>
<p>Please follow the instructions outlined in the <a href = "https://github.com/VisionLearningGroup/taskcv-2017-public">VisDA GitHub repository</a> to download data and development kits for the classification and segmentation tracks. Training, validation and testing data are now available to download. We have also included baseline models and instructions on training several existing domain adaptation methods.</p>
<p>The DevKits are currently in beta-release. If you find any bugs, please open an issue in our GitHub repo rather than emailing the organizers directly.</p>
<p> You can get the <b><a href="https://arxiv.org/abs/1710.06924">tech report</a></b> describing dataset and baseline experiments from
<b><a href="https://arxiv.org/abs/1710.06924">arxiv</a></b>. If you use data, code or its derivatives, please consider citing:
</p>
<pre style="overflow: auto; word-wrap: normal; white-space: pre;"><code>@misc{visda2017,
Author = {Xingchao Peng and Ben Usman and Neela Kaushik and Judy Hoffman and Dequan Wang and Kate Saenko},
Title = {VisDA: The Visual Domain Adaptation Challenge},
Year = {2017},
Eprint = {arXiv:1710.06924},
}</code></pre>
</br>
<a name = "evaluation"></a>
<h2 class="section-title"> Evaluation </h2>
<p> We will use CodaLab to evaluate submissions and maintain a leaderboard. To register for the evaluation server, please create an account on <a href = "https://competitions.codalab.org/">CodaLab</a> and and enter as a participant in one of the following competitions: </p>
<ul>
<li><a href = "https://competitions.codalab.org/competitions/17052">Classification Challenge</a></li>
<li><a href = "https://competitions.codalab.org/competitions/17054">Segmentation Challenge</a></li>
</ul>
<p>If you are working as a team, please register using <b>one</b> account for your team and indicate the names of all of the members on your team. This can be modified in the “User Settings” tab. See the challenge <a href = "#rules">rules</a> for more details. </p>
<p>Please refer to the instructions in the <a href = "https://github.com/VisionLearningGroup/taskcv-2017-public">DevKit</a> ReadMe file for specific details on submission formatting and evaluation for the classification and segmentation challenges.</p>
</br>
<a name = "rules"></a>
<h2 class="section-title"> Rules </h2>
<p>The VisDA challenge tests adaptation and model transfer, so the rules are different than most challenges. Please read them carefully.</p>
<p><b>Supervised Training:</b> Teams may only submit test results of models trained on the source domain data and optionally pre-trained on ImageNet. Training on the validation dataset is not allowed for test submissions. Note, this may be different from other challenges. In this year’s VisDA challenge, the goal is to test how well models can adapt from synthetic to real data. Therefore, training on the validation domain is not allowed. To ensure equal comparison, we also do not allow any other external training data, modifying the provided training dataset, or any form of manual data labeling. </p>
<p><b>Pretraining on ImageNet:</b> If pre-training on the <a href = "http://image-net.org/challenges/LSVRC/2012/">ImageNet ILSVRC</a> classification training data, only the weights can be transferred, not the actual classifiers for specific objects, i.e. participants should not manually exploit correspondences between ImageNet output labels and labels in the data. Please indicate in your method description which pre-trained weights were used for initialization of the model. Teams who place in the top of the “no ImageNet pretraining” track will receive special recognition. </p>
<p><b>Unsupervised training:</b> Models can be adapted (trained) on the test data in an unsupervised way, i.e. without labels. Adaptation, even unsupervised, on the validation data is not allowed for test submissions. Note, we have released the validation labels to facilitate algorithm development. </p>
<p><b>Source Models:</b> The performance of a domain adaptation algorithm greatly depends on the baseline performance of the model trained only on source data. We ask that teams submit two sets of results: 1) predictions obtained only with the source-trained model, and 2) predictions obtained with the adapted model. See the development kit for submission formatting details. </p>
<p><b>Leaderboard:</b>The main leaderboard for each competition track will show results of adapted models and will be used to determine the final team ranks. The expanded leaderboard will additionally show the team's source-only models, i.e. those trained only on the source domain without any adaptation. These results are useful for estimating how much the method improves upon its source-only model, but will not be used to determine team ranks. </p>
</br>
<a name = "faq"></a>
<h2 class="section-title"> FAQ </h2>
<ol>
<li><b>Can we train models on data other than the source domain?</b></li>
<p> Participants may elect to pre-train their models only on ImageNet. Please refer to the challenge evaluation instructions found in the <a href = "#download">DevKit</a> for more details. </p>
<li><b>Do we have to use the provided baseline models?</b></li>
<p>No, these are provided for your convenience and are optional. </p>
<li><b>How many submissions can each team submit per competition track?</b></li>
<p>For the validation domain, the number of submissions per team is limited to 20 upload per day and there are no restrictions on total number of submissions. For the test domain, the number of submissions per team is limited to 1 upload per day and 20 uploads in total. Only one account per team must be used to submit results. Do not create multiple accounts for a single project to circumvent this limit, as this will result in disqualification. </p>
<li><b>Can multiple teams enter from the same research group?</b><li>
<p>Yes, so long as each team is comprised of different members.</p>
<li><b>Can external data be used?</b></li>
<p>The allowed training data consists of the VisDA 2017 Training set. The VisDA 2017 Validation set can be used to test adaptation to a target domain offline, but cannot be used to train the final submitted model (with or without labels). Optional initialization of models with weights pre-trained on ImageNet is allowed and must be declared in the submission. Please see the <a href = "#rules">challenge rules</a> for more details. </p>
<li><b>Are challenge participants required to reveal all details of their methods?</b></li>
<p>Participants are encouraged to include a brief write-up regarding their methods when submitting their results. However, this is not mandatory. </p>
<li><b>Do participants need to adhere to TASK-CV abstract submission deadlines to participate in the challenge?</b></li>
<p>Submission of a <a href = "http://adas.cvc.uab.es/task-cv2017/">TASK-CV</a> workshop abstract is not required to participate in the challenge. The top-performing teams will be invited to present their approaches at the workshop, even if they did not submit an abstract.
</p>
<li><b>When is the final submission deadline?</b></li>
<p>All results must be submitted by September 29th at 11:59pm ET. On CodaLab, this is equivalent to September 30th at 3:59am UTC.<p>
</ol>
</br>
<a name = "workshop"></a>
<h2 class="section-title"> Workshop </h2>
<p>The challenge is associated with the 4th annual <a href = "http://adas.cvc.uab.es/task-cv2017/">TASK-CV</a> workshop, being held at <a href = "http://iccv2017.thecvf.com/">ICCV 2017</a> in Venice, Italy. Challenge participants are invited to submit abstracts to the workshop, but this is not required for challenge participation. If you wish to submit an abstract, please see the <a href = "http://adas.cvc.uab.es/task-cv2017/">TASK-CV</a> website for deadlines. The top performing teams will be invited to give a talk about their results at the workshop during a special session. </p>
<hr>
</br>
<a name = "organizers"></a>
<h2 class="section-title"> Organizers </h2>
Kate Saenko (Boston University),
Ben Usman (Boston University),
Xingchao Peng (Boston University),
Neela Kaushik (Boston University),
Judy Hoffman (Stanford University),
Dequan Wang (UC Berkeley)
</br>
</br>
<a name = "sponsors"></a>
<h2 class="section-title"> Sponsors </h2>
<img src="{{site.baseurl}}/assets/images/nsf_logo.png" style="float: left; width: 20%; margin-right: 5%; margin-bottom: 0.5em;">
<img src="{{site.baseurl}}/assets/images/Nexar-Logo.jpg" style="float: left; width: 40%; margin-right: 5%; margin-bottom: 0.5em;">
<p style="clear: both;">
</br></br></br>
<h2 class="section-title"> References </h2>
[1] FCNs in the Wild: Pixel-level Adversarial and Constraint-based Adaptation, Judy Hoffman, Dequan Wang, Fisher Yu, Trevor Darrell
</br></br>
<div height=500px> </div>
</div><!--//summary-->
</section><!--//section-->