[visionlist] [CFP] FOCUS: FOundation models Creators meet USers - ECCV24 Workshop

Francesca Pistilli francesca.pistilli at polito.it
Thu Jun 13 05:37:23 -04 2024


********************************

Call for Papers

FOCUS: FOundation models Creators meet USers

International Workshop at ECCV 2024
https://focus-workshop.github.io/

********************************


=== Submission Deadline: July 15, 2024 ====

(11:59 p.m. CET)

Apologies for multiple posting

Please distribute this call to interested parties



AIMS AND SCOPE
===============

Over the last few years, the field of Artificial Intelligence has witnessed significant growth, largely fueled by the development of large-scale machine learning models. The foundational models are characterized by extensive training on diverse datasets that encompass various input modalities (e.g. images, text, audio, 3D data), showing excellent flexibility and effectiveness across a wide range of standard NLP and Computer Vision tasks. Such general-purpose solutions often reveal potentials that go beyond what their creators originally envisioned, motivating users to adopt these models for a broad spectrum of applications.


Nevertheless, the knowledge embedded in these models may not be enough when the final goal goes beyond perception benchmarks. These considerations spark important questions that can only be answered with a collaborative dialogue between researchers developing these models (creators) and those employing them in downstream tasks (users). Each group brings a unique perspective that will be crucial in shaping the future of this technology.

The goal of this workshop is to identify and discuss strategies to assess both positive and negative (possibly unexpected) behaviors in the development and use of foundation models.
Particular attention will be given to applications that diverge significantly from the scenarios encountered during the training phase of foundational models. These include application-specific visual understanding, uncertainty evaluation, goal-conditioned reasoning, human habits learning, task and motion planning, scene navigation, vision-based manipulation, etc.
Our purpose is to foster an open discussion between foundation model creators and users, targeting the analysis of the most pressing open questions for the two communities and fostering new fruitful collaborations.


TOPICS
=======

The workshop calls for submissions addressing, but not limited to, the
following topics:

- New vision-and-language applications
- Supervised vs unsupervised based foundation model and downstream tasks
- Zero-shot, Few-shot, continual and life-long learning of foundation model
- Open set, out-of-distribution detection and uncertainty estimation
- Perceptual reasoning and decision making: alignment with human intents and modeling
- Prompt and Visual instruction tuning
- Novel evaluation schemes and benchmarks
- Task-specific vs general-purpose models
- Robustness and generalization
- Interpretability and explainability
- Ethics and bias in prompting


IMPORTANT DATES
=================

- Paper Submission Deadline: July 15th, 2024
- Decision to Authors: July 31th, 2024
- Camera ready papers due: August 15th, 2024


SUBMISSION GUIDELINES
======================

Papers should be submitted at:
https://cmt3.research.microsoft.com/FOCUS2024

At the time of submission, authors must indicate the type of the paper:

   - Full papers: must present original research, not published elsewhere, and follow the ECCV main conference policies and format with a maximum length of 14 pages (extra pages with references only are allowed). Accepted full papers will be included in the ECCV 2024 Workshop proceedings. Supplemental materials are not allowed.

   - Short papers: previously or concomitantly published works that could foster the workshop objectives. Short-papers will have a maximum length of 4 pages (extra pages with references only are allowed), they will be presented without inclusion in the ECCV 2024 Workshop proceedings. Supplemental materials are not allowed.

The review process is double-blind and there is no rebuttal.



SPEAKERS
======================

- Kira Zsolt, Professor, Georgia Tech
- Ishan Misra, Director, Facebook AI Research
- Hilde Kuehne, University of Bonn



ORGANIZING COMMITTEE
======================

- Antonio Alliegro, Politecnico di Torino
- Francesca Pistilli, Politecnico di Torino
- Songyou Peng, ETH Zurich
- Biplab Banerjee, IIT Bombay
- Gabriela Csurka, Naver Labs Europe
- Giuseppe Averta, Politecnico di Torino



-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://visionscience.com/pipermail/visionlist_visionscience.com/attachments/20240613/497f34d0/attachment-0001.html>


More information about the visionlist mailing list