11 June, 2025, 8 AM - 12 PM 📍 Room 109 @ Music City Center, Nashville, TN, US
Workshop on Computer Vision for Mixed Reality
In conjunction with CVPR 2025
Call for Papers

Overview

Virtual Reality (VR) technologies have the potential to transform the way we use computing to interact with our environment, do our work and connect with each other. VR devices provide users with immersive experiences at the cost of blocking the visibility of the surrounding environment. With the advent of passthrough techniques such as those in Quest-3 and Apple Vision Pro, now users can build deeply immersive experiences which mix the virtual and the real world into one, often also called Mixed Reality (MR). MR poses a set of very unique research problems in computer vision that are not covered by VR. Our focus is on capturing the real environment around the user using cameras which are placed away from the user's eyes, yet reconstruct the environment with high fidelity, augmented the environment with virtual objects and effects, and all in real-time. We aim to offer the research community to deeply understand the unique challenges of Mixed Reality and research on novel methods encompassing View Synthesis, Scene Understanding, efficient On-Device AI among other things.

Speakers

Richard Newcombe (Keynote)

Vice President, Research Science
Meta Reality Labs Research

Anjul Patney

NVIDIA

Rana Hanocka

University of Chicago

Laura Leal-Taixé

NVIDIA, TUM

Margarita Grinvald

Meta

Program

08:00-08:30 am Intro + Accepted papers spotlight
08:30-08:40 am Rakesh Ranjan Opening remarks
08:40-09:35 am Richard Newcombe [Keynote] Problems still to be solved on the path to the next computing platform -- extending reality with always-on contextual MR, AI and Social Teleportation
09:35-10:05 am Anjul Patney Pixels at Speed of Light: Lessons in Deploying CV in Graphics & Games
10:05-10:35 am Rana Hanocka Data-Driven Neural Mesh Editing – without 3D Data
Abstract: Much of the current success of deep learning has been driven by massive amounts of curated data, whether annotated or unannotated. Compared to image datasets, developing large-scale 3D datasets is either prohibitively expensive or impractical. In this talk, I will present several works that harness the power of data-driven deep learning for tasks in shape editing and processing, without any 3D datasets. I will discuss works that learn to synthesize and analyze 3D geometry using large image datasets.
10:35-11:00 am Poster Spotlight + Break
11:00-11:30 pm Margarita Grinvald Differentiable Passthrough - A learning-based approach to reduce perceptual artifacts
11:30-12:00 pm Laura Leal-Taixé Towards a Foundation Model for 4D (Lidar)
12:00-12:10 pm Best Poster Award + Town Hall

Accepted papers

MR.NAVI: Mixed-Reality Navigation Assistant for the Visually Impaired
Nicolas Pfitzer, Yifan Zhou, Marco Poggensee, Defne Kurtulus, Bessie Dominguez-Dager, Mihai Dusmanu, Marc Pollefeys, Zuria Bauer [PDF]
Spot-On: A Mixed Reality Interface for Multi-Robot Cooperation
Tim Engelbracht, Petar Lukovic, Tjark Behrens, Kai Lascheit, René Zurbrügg, Marc Pollefeys, Hermann Blum, Zuria Bauer [PDF]
GATE3D : Generalized Attention-based Task-synergized Estimation in 3D
Eunsoo Im [PDF] [Project Page]
BOP Challenge 2024 on Model-Based and Model-Free 6D Object Pose Estimation
Van Nguyen Nguyen, Stephen Tyree, Andrew Y Guo, Médéric Fourmy, Anas Gouda, Taeyeop Lee, Sungphill Moon, Hyeontae Son, Lukas Ranftl, Jonathan Tremblay, Eric Brachmann, Bertram Drost, Vincent Lepetit, Carsten Rother, Stan Birchfield, Jiri Matas, Yann Labbé, Martin Sundermeyer, Tomas Hodan [PDF]
GFreeDet: Exploiting Gaussian Splatting and Foundation Models for Model-free Unseen Object Detection in the BOP Challenge 2024
Xingyu Liu, Gu Wang, Chengxi Li, Yingyue Li, Chenyangguang Zhang, Ziqin Huang, Xiangyang Ji [PDF]
GoTrack: Generic 6DoF Object Pose Refinement and Tracking
Van Nguyen Nguyen, Christian Forster, Sindi Shkodrani, Vincent Lepetit, Bugra Tekin, Cem Keskin, Tomas Hodan [PDF]
Lost & Found: Tracking Changes from Egocentric Observations in 3D Dynamic Scene Graph
Tjark Behrens, René Zurbrügg, Marc Pollefeys, Zuria Bauer, Hermann Blum [PDF]
MDBNet 360Âş: 3D Audio-Visual Indoor Scene Reconstruction and Completion from a Single 360Âş RGB-D Image
Mona Alawadh, Atiyeh Alinaghi, Mahesan Niranjan, Hansung Kim [PDF]
Text2Stereo: Repurposing Stable Diffusion for Stereo Generation with Consistency Rewards
Aakash Garg, Libing Zeng, Andrii Tsarov, Nima Kalantari [PDF]
CroCoDL: Cross-device Collaborative Dataset for Localization
Hermann Blum, Alessandro Mercurio, Joshua O'Reilly, Tim Engelbracht, Mihai Dusmanu, Marc Pollefeys, Zuria Bauer [PDF]
SMARTe-VR: Student Monitoring and Adaptive Response Technology for e-Learning in Virtual Reality
Roberto Daza, LinShengkai, Aythami Morales, Julian Fierrez, Katashi Nagao [PDF]

Call for Papers

Important Dates

  • Paper submission deadline: March 28, 2025
  • Notification to authors: April 14, 2025
  • Camera-ready deadline: April 24, 2025 (after this date, no more updates to the paper will be accepted)

Papers submitted to the workshop will appear in the proceedings of the CVPR workshops in 2025.

Announcement

Only two weeks before our workshop submission deadline, CVPR Workshop Chairs unexpectedly communicated us a March 31 deadline for proceedings submission, leaving us only 48 hours for the review process. Despite our efforts, we were unable to extend this deadline with them. Rather than compromise on review quality, we've decided not to submit accepted papers to the CVPR Workshop proceedings. Our priority is to provide the authors with high-quality feedback, select the best papers, and ensure they are highlighted on the workshop website. Other workshop organizers we know are taking the same approach. Thank you for your understanding.


Topics of Interest

The CV4MR 2025 workshop will highlight frontiers of innovation in turning wearable computers, sensors and displays into augmentations of human capability for productivity, life improvement or recreation. Since this topic is inherently interdisciplinary, we encourage authors to submit works in AI, Computer Vision, Image Processing or Computational Photography that they think are applicable to advancing this field.

Authors are highly encouraged to motivate their applications for Mixed Reality in the submissions.

Here is a non-exhaustive list of topics we encourage submissions on:

  • Innovations in Real-Time passthrough
    • View Synthesis
    • Image and Audio Stylization
  • 3D Scene recovery from ego sensors
    • Depth Estimation
    • 3D capture, reconstruction and rendering for virtual objects
    • SLAM and tracking
  • Scene and Human understanding
  • AI Agents for human assistance
    • Scenegraph-modifying or task agents
    • Humanoid / conversational agents
  • AI for User Interaction and Spatial Design
  • Innovations in AI-based graphics for Mixed Reality
  • We also encourage submission on novel Applications of Mixed Reality in areas such as Healthcare, Manufacturing, etc.

Best Workshop Paper Award

We are pleased to announce a CV4MR Best Workshop Paper Award (with a Meta Quest 3S prize sponsored by Meta), to be selected from the accepted papers.


Submission Guidelines:

  • We invite submissions of max 8 pages (excluding references), and 4-page (excluding references) extended abstracts as well.
  • Submitted manuscript should follow the CVPR 2025 paper template.
  • The review process is double-blind and does not involve a rebuttal phase. If you have other media to attach (videos etc), please feel free to add anonymized links.
  • Submissions will be rejected without review if they:
      1. Contain more than 8 pages (excluding references).
      2. Violate the double-blind policy.
      3. Violate the dual-submission policy for papers with more than 4 pages excluding references (to reiterate, we will not accept full papers already accepted into CVPR's papers track).
  • Authors of all accepted submissions will be asked to present their work in a poster session (the guidelines for the posters are the same as at the main conference), in addition selected papers may be invited to deliver spotlight talks.
  • Authors submitting to the workshop will be asked to be added to the pool of reviewers. For submissions with more than 2 authors, we will ask that at least 2 authors agree to be added to the workshop reviewer pool. We will do our best to make the review workload reasonable, but for fairness, submissions by authors not completing reviews may be desk-rejected.


Reviewing for CV4MR 2025

Reviewers are the backbone for the integrity of knowledge in our workshop. For those interested in being added to the reviewer pool, please email cv4mr@googlegroups.com with the subject “Reviewer Pool Participation”, some information about you, and your resume attached.

Organizers

Rakesh Ranjan

Meta

Omer Shapira

NVIDIA

Margarita Grinvald

Meta

Ramya Akula

Meta

Vikas Chandra

Meta

Andrea Colaco

Google

Contact Us

cv4mr@googlegroups.com