[visionlist] CFP: New Trends in Image Restoration and Enhancement workshop and challenges @ CVPR 2018
Radu Timofte
timofte.radu at gmail.com
Tue Jan 16 13:45:40 -05 2018
Apologies for cross-posting
*******************************
CALL FOR PAPERS & CALL FOR PARTICIPANTS IN 3 CHALLENGES
NTIRE: 3rd New Trends in Image Restoration and Enhancement workshop and
challenges on image super-resolution, dehazing, and spectral reconstruction
In conjunction with CVPR 2018, June 18, Salt Lake City, USA.
Website: http://www.vision.ee.ethz.ch/ntire18/
Contact: radu.timofte at vision.ee.ethz.ch
SCOPE
Image restoration and image enhancement are key computer vision tasks,
aiming at the restoration of degraded image content or the filling in of
missing information. Recent years have witnessed an increased interest from
the vision and graphics communities in these fundamental topics of
research. Not only has there been a constantly growing flow of related
papers, but also substantial progress has been achieved.
Each step forward eases the use of images by people or computers for the
fulfillment of further tasks, with image restoration or enhancement serving
as an important frontend. Not surprisingly then, there is an ever growing
range of applications in fields such as surveillance, the automotive
industry, electronics, remote sensing, or medical image analysis. The
emergence and ubiquitous use of mobile and wearable devices offer another
fertile ground for additional applications and faster methods.
This workshop aims to provide an overview of the new trends and advances in
those areas. Moreover, it will offer an opportunity for academic and
industrial attendees to interact and explore collaborations.
TOPICS
Papers addressing topics related to image/video restoration and enhancement
are invited. The topics include, but are not limited to:
● Image/video inpainting
● Image/video deblurring
● Image/video denoising
● Image/video upsampling and super-resolution
● Image/video filtering
● Image/video dehazing
● Demosaicing
● Image/video compression
● Artifact removal
● Image/video enhancement: brightening, color adjustment, sharpening, etc.
● Style transfer
● Image/video generation and image hallucination
● Image/video quality assessment
● Hyperspectral imaging
● Underwater imaging
● Aerial and satellite imaging
● Methods robust to changing weather conditions / adverse outdoor
conditions
● Studies and applications of the above.
SUBMISSION
A paper submission has to be in English, in pdf format, and at most 8 pages
(excluding references) in CVPR style. The paper format must follow the same
guidelines as for all CVPR submissions.
http://cvpr2018.thecvf.com/submission/main_conference/author_guidelines
The review process is double blind. Authors do not know the names of the
chair/reviewers of their papers. Reviewers do not know the names of the
authors.
Dual submission is allowed with CVPR main conference only. If a paper is
submitted also to CVPR and accepted, the paper cannot be published both at
the CVPR and the workshop.
For the paper submissions, please go to the online submission site (opens
February 1, 2018).
Accepted and presented papers will be published after the conference in the
CVPR Workshops Proceedings on by IEEE (http://www.ieee.org) and Computer
Vision Foundation (www.cv-foundation.org).
The author kit provides a LaTeX2e template for paper submissions. Please
refer to the example for detailed formatting instructions. If you use a
different document processing system then see the CVPR author instruction
page.
Author Kit: http://cvpr2018.thecvf.com/files/cvpr2018AuthorKit.zip
WORKSHOP DATES
● Submission Deadline: March 01, 2018
● Decisions: March 29, 2018
● Camera Ready Deadline: April 05, 2018
CHALLENGE on SUPER-RESOLUTION (started!)
The challenge has 4 tracks as follows:
1. *Track 1: classic bicubic*
<https://competitions.codalab.org/competitions/18015> uses the bicubic
downscaling (Matlab imresize), the most common setting from the recent
single-image super-resolution literature.
2. *Track 2: realistic mild adverse conditions *
<https://competitions.codalab.org/competitions/18024> assumes that the
degradation operators (emulating the image acquisition process from a
digital camera) are the same within an image space and for all the images.
3. *Track 3: realistic difficult adverse conditions *
<https://competitions.codalab.org/competitions/18025> assumes that the
degradation operators (emulating the image acquisition process from a
digital camera) are the same within an image space and for all the images.
4. *Track 4: realistic wild conditions*
<https://competitions.codalab.org/competitions/18026> assumes that the
degradation operators (emulating the image acquisition process from a
digital camera) are the same within an image space but DIFFERENT from one
image to another. This setting is the closest to real "wild" conditions.
To learn more about the challenge, to participate in the challenge, and to
access dataset everybody is invited to register at the above links
http://www.vision.ee.ethz.ch/ntire18/
CHALLENGE on IMAGE DEHAZING (started!)
* A novel datasets of real hazy images obtained in outdoor and indoor
environments with ground truth is introduced with the challenge. It is the
first image dehazing online challenge.*
1. * Track 1: Indoor* - the goal is to restore the visibility in images
with haze generated in a controlled indoor environment.
2. * Track 2: Outdoor* - the goal is to restore the visibility in
outdoor images with haze generated using a professional haze/fog generator.
For data and more details:
http://www.vision.ee.ethz.ch/ntire18/
CHALLENGE on SPECTRAL RECONSTRUCTION (started!)
*The largest dataset to date will be introduced with the challenge. It is
the first spectral reconstruction from RGB images online challenge. *
1. *Track 1: “Clean” * recovering hyperspectral data from uncompressed
8-bit RGB images created by applying a know response function to ground
truth hyperspectral information.
2. *Track 2: “Real World” * recovering hyperspectral data from
jpg-compressed 8-bit RGB images created by applying an unknown response
function to ground truth hyperspectral information.
For data and more details:
http://www.vision.ee.ethz.ch/ntire18/
CHALLENGES DATES
● Release of train data: January 10, 2018
● Competition ends: March 01, 2018
ORGANIZERS
● Radu Timofte, ETH Zurich, Switzerland (radu.timofte [at] vision.ee.ethz.ch)
● Ming-Hsuan Yang, University of California at Merced, US (mhyang [at]
ucmerced.edu)
● Jiqing Wu, ETH Zurich, Switzerland (Jiqing.wu [at] vision.ee.ethz.ch)
● Lei Zhang, The Hong Kong Polytechnic University (cslzhang [at]
polyu.edu.hk)
● Luc Van Gool, KU Leuven, Belgium and ETH Zurich, Switzerland (vangool
[at] vision.ee.ethz.ch)
● Cosmin Ancuti, Université catholique de Louvain (UCL), Belgium
● Codruta O. Ancuti, University Politehnica Timisoara, Romania
● Boaz Arad, Ben-Gurion University, Israel
● Ohad Ben-Shahar, Ben-Gurion University, Israel
PROGRAM COMMITTEE (to be updated)
Cosmin Ancuti, Université catholique de Louvain (UCL), Belgium
Michael S. Brown, York University, Canada
Subhasis Chaudhuri, IIT Bombay, India
Sunghyun Cho, Samsung
Oliver Cossairt, Northwestern University, US
Chao Dong, SenseTime
Weisheng Dong, Xidian University, China
Alexey Dosovitskiy, Intel Labs
Touradj Ebrahimi, EPFL, Switzerland
Michael Elad, Technion, Israel
Corneliu Florea, University Politehnica of Bucharest, Romania
Alessandro Foi, Tampere University of Technology, Finland
Bastian Goldluecke, University of Konstanz, Germany
Luc Van Gool, ETH Zürich and KU Leuven, Belgium
Peter Gehler, University of Tübingen and MPI Intelligent Systems,
Germany
Hiroto Honda, DeNA Co., Japan
Michal Irani, Weizmann Institute, Israel
Phillip Isola, UC Berkeley, US
Zhe Hu, Light.co
Sing Bing Kang, Microsoft Research, US
Vivek Kwatra, Google
Christian Ledig, Twitter, UK
Kyoung Mu Lee, Seoul National University, South Korea
Seungyong Lee, POSTECH, South Korea
Stephen Lin, Microsoft Research Asia
Chen Change Loy, Chinese University of Hong Kong
Vladimir Lukin, National Aerospace University, Ukraine
Kai-Kuang Ma, Nanyang Technological University, Singapore
Vasile Manta, Technical University of Iasi, Romania
Yasuyuki Matsushita, Osaka University, Japan
Peyman Milanfar, Google and UCSC, US
Rafael Molina Soriano, University of Granada, Spain
Yusuke Monno, Tokyo Institute of Technology, Japan
Hajime Nagahara, Kyushu University, Japan
Vinay P. Namboodiri, IIT Kanpur, India
Sebastian Nowozin, Microsoft Research Cambridge, UK
Aleksandra Pizurica, Ghent University, Belgium
Fatih Porikli, Australian National University, NICTA, Australia
Hayder Radha, Michigan State University, US
Stefan Roth, TU Darmstadt, Germany
Aline Roumy, INRIA, France
Jordi Salvador, Amazon, US
Yoichi Sato, University of Tokyo, Japan
Samuel Schulter, NEC Labs America
Nicu Sebe, University of Trento, Italy
Boxin Shi, National Institute of Advanced Industrial Science and
Technology (AIST), Japan
Wenzhe Shi, Twitter Inc.
Alexander Sorkine-Hornung, Disney Research
Sabine Süsstrunk, EPFL, Switzerland
Yu-Wing Tai, Tencent Youtu
Hugues Talbot, Université Paris Est, France
Robby T. Tan, Yale-NUS College, Singapore
Masayuki Tanaka, Tokyo Institute of Technology, Japan
Jean-Philippe Tarel, IFSTTAR, France
Radu Timofte, ETH Zürich, Switzerland
Ashok Veeraraghavan, Rice University, US
Jue Wang, Megvii Research, US
Chih-Yuan Yang, UC Merced, US
Ming-Hsuan Yang, University of California at Merced, US
Qingxiong Yang, Didi Chuxing, China
Lei Zhang, The Hong Kong Polytechnic University
Wangmeng Zuo, Harbin Institute of Technology, China
SPEAKERS (to be announced)
SPONSORS (to be updated)
NVIDIA
SenseTime
Twitter Inc
Google
Contact: radu.timofte at vision.ee.ethz.ch
Website: http://www.vision.ee.ethz.ch/ntire18/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://visionscience.com/pipermail/visionlist_visionscience.com/attachments/20180116/db95395b/attachment-0001.html>
More information about the visionlist
mailing list