The first Workshop on
Visual Object-oriented Learning meets Interaction (VOLI):
Discovery, Representations, and Applications


October 24 @ ECCV 2022 Virtual


  • Speakers
  • Schedule
  • Papers
  • Contact

News

  • [Nov 2, 2022] Here are the recordings for the workshop talks and panel discussion: [Morning Part] [Afternoon Part]
  • [Oct 24, 2022] We are live-streaming the workshop soon 9 AM - 5:30 PM Tel-Aviv Time (GMT+3) on Oct 24th! Here is the ECCV Platform (please log in using your ECCV credentials, click "Workshops", then "Oct 24", find our workshop and click "Join Now").
  • [Oct 24, 2022] Check out the full schedule and the accepted papers.
  • [Oct 24, 2022] We will host the poster session on GatherTown (check out the GatherTown instructions on how to join).
  • [Apr 6, 2022] Workshop website launched, with Call-for-Papers and tentative speakers announced.

Introduction

Objects, as the most basic and composable units in visual data, exist in specific visual appearances and geometrical forms, carrying rich semantic, functional, dynamic, and relational information. One may discover objects by watching passive videos or actively interacting with the world to find them. Once detected, it is also an open research problem how to interact with the objects to extract and represent such object-oriented semantics. Furthermore, such representations need to be designed easily useful for various downstream perception and robotic interaction tasks. It is a crucial research domain for studying how to define, discover, and represent objects in visual data from/for interaction, and use them for various downstream applications.

In this workshop, we will be focusing on discussing learning formulations, approaches, and methodologies that define/discover "objects" (e.g., objects, parts, concepts) from/for interaction (e.g., objects interact in the world, agents interact with objects) in unsupervised, weak-supervised, and/or self-supervised manners, design/learn task-agnostic or task-aware visual representations for various object-oriented properties (e.g., dynamics, functionality, affordance), and/or explores ways to apply the developed object-oriented methods to different downstream applications in different fields (e.g., machine learning, computer graphics, computer vision, robotics, cognition). Some concrete examples are listed below.

  • How to define/discover visual ”objects” from/for interaction?
    • How to define the concepts of objects from/for downstream tasks?
    • How to achieve unsupervised, weak-supervised, self-supervised interaction, or embodied learning for object discovery?
    • What are different approaches to discover objects?
  • What are good object-oriented representations and how to learn them from/for interaction?
    • What properties of objects need to be learned and can be learned from/for interactions?
    • How to design suitable learning representations for different desired object attributes? Is there a unified way to represent desired object properties for various downstream tasks?
    • Shall we learn task-agnostic or task-aware representations? How to extract, learn, represent, and use the learned task-aware information?
  • How to learn useful object-oriented representations for different downstream applications?
    • What object-oriented representations do different tasks from different fields need? How much they are similar or different?
    • Do we need to learn task-specific representations? Can we learn universal object representations for all tasks?
    • How to design learning approaches that can allow objects or agents to interact with the "objects" in the environment for better learning representations for downstream tasks?

Invited Speakers (Check out the Full Schedule)

Adam R. Kosiorek

Christopher Xie

Klaus Greff

Yixin Zhu

DeepMind
Meta Reality Labs
Google Research
Peking University

Jeannette Bohg

Hao Su

Sungjin Ahn

Judith Fan

Stanford
UCSD
KAIST
UCSD

Thomas Kipf

Stefano Soatto

Kostas Daniilidis

Niloy Mitra

Google Brain
UCLA & Amazon
UPenn
UCL & Adobe

Call for Papers (Check out the Accepted Papers)

We accept both archival and non-archival paper submissions. The accepted archival papers will be included in the ECCV22 conference proceedings, while the non-archival ones will just be presented in the workshop. All papers will be peer-reviewed by three experts in the field in a double-blind manner. We also welcome papers that are accepted to the ECCV main conference or other previous conferences to present your work in the non-archival paper track. There is no need for peer-review for such previously accepted papers, so please indicate clearly in the submission form. Every accepted paper will have the opportunity to give a 5-min spotlight presentation and host two 30-min poster sessions (12-hours separated).

Submission Site: https://cmt3.research.microsoft.com/VOLI2022


Submission Instructions:

Please use the official ECCV template (under "Submission Guidelines") for your submissions. Make sure to anonymize your submission.

Timeline Table (11:59 PM, Pacific Time)

  • Friday, Jul 15: Paper submission deadline
  • Friday, Aug 5: Review deadline
  • Friday, Aug 12: decision announced to authors
  • Monday, Aug 22: Camera ready deadline

Organizers

Kaichun Mo

Yanchao Yang

Jiayuan Gu

Stanford & NVIDIA
Stanford
UCSD

Shubham Tulsiani

Hongjing Lu

Leonidas Guibas

CMU
UCLA
Stanford

Contact Info

E-mail: [email protected]

Acknowledgements

Website template borrowed from: https://futurecv.github.io/ (Thanks to Deepak Pathak)

 


OSZAR »