EARTHVISION 2025

EARTHVISION 2025

June 11/12th, Nashville, TN, USA

Aims and Scope / Call for Papers

Earth Observation (EO) and remote sensing are ever-growing fields of investigation where computer vision, machine learning, and signal/image processing meet. The general objective of the domain is to provide large-scale and consistent information about processes occurring at the surface of the Earth by exploiting data collected by airborne and spaceborne sensors. Earth Observation covers a broad range of tasks, from detection to registration, data mining, and multi-sensor, multi-resolution, multi-temporal, and multi-modality fusion and regression, to name just a few. It is motivated by numerous  applications such as location-based services, online mapping services, large-scale surveillance, 3D urban modeling, navigation systems, natural hazard forecast and response, climate change monitoring, virtual habitat modeling, food security, etc. The sheer amount of data calls for highly automated scene interpretation workflows. 

Earth Observation and in particular the analysis of spaceborne data directly connects to 34 indicators out of 40 (29 targets and 11 goals) of the Sustainable Development Goals defined by the United Nations ( https://sdgs.un.org/goals  ). The aim of EarthVision to advance the state of the art in machine learning-based analysis of remote sensing data is thus of high relevance. It also connects to other immediate societal challenges such as monitoring of forest fires and other natural hazards, urban growth, deforestation, and climate change.

A non exhaustive list of topics of interest includes the following:

  • Super-resolution in the spectral and spatial domain

  • Hyperspectral and multispectral image processing

  • Reconstruction and segmentation of optical and LiDAR 3D point clouds

  • Feature extraction and learning from spatio-temporal data 

  • Analysis  of UAV / aerial and satellite images and videos

  • Deep learning tailored for large-scale Earth Observation

  • Domain adaptation, concept drift, and the detection of out-of-distribution data

  • Data-centric machine learning

  • Evaluating models using unlabeled data

  • Self-, weakly, and unsupervised approaches for learning with spatial data

  • Foundation models and representation learning in the context of EO

  • Human-in-the-loop and active learning

  • Multi-resolution, multi-temporal, multi-sensor, multi-modal processing

  • Fusion of machine learning and physical models

  • Explainable and interpretable machine learning in Earth Observation applications

  • Uncertainty quantification of machine-learning based prediction from EO data

  • Applications for climate change, sustainable development goals, and geoscience

  • Public benchmark datasets: training data standards, testing & evaluation metrics, as well as open source research and development.

All manuscripts will be subject to a double-blind review process. Accepted EarthVision papers will be included in the CVPR2024 workshop proceedings (published open access on the Computer Vision Foundation website) and submitted to IEEE for publication in IEEEXplore. Publication in IEEEXplore will be granted only if the paper meets IEEE publication policies and procedures.

Important Dates

All deadlines are considered end of day anywhere on Earth.

March 3, 2025Submission deadline 
March 31,  2025Notification to authors 
April 7, 2025Camera-ready deadline 
June 11/12, 2025Workshop 

Program

TBD

Keynote 1 –  Sherrie Wang, Massachusetts Institute of Technology

“Scaling Ground Truth to Advance Earth Observation”

Bio

Sherrie Wang is an Assistant Professor at MIT in the Department of Mechanical Engineering and Institute of Data, Systems, and Society. Her research uses novel data and computational algorithms to monitor our planet and enable sustainable development. Her focus is on improving agricultural management and mitigating climate change, especially in low- or middle-income regions of the world. To this end, she frequently uses satellite imagery, crowdsourced data, LiDAR, and other spatial data. Due to the scarcity of ground truth data in these regions and the noisiness of real-world data in general, her methodological work is geared toward developing machine learning methods that work well with these constraints.

TBD

Keynote 2 – Kristof Van Tricht, VITO

“From the Research Lab to a Global Map: Overcoming Scaling Barriers in EO Algorithm Development”

Bio

Kristof Van Tricht is a senior researcher at VITO Remote Sensing, where he specializes in agricultural applications. With a focus on both global and local agriculture monitoring, he leverages advanced remote sensing technologies to address food security challenges. His work spans the development and deployment of crop mapping algorithms, and the development of the CropSAR technology, which fuses Sentinel-1 and Sentinel-2 observations to provide cloud-free data streams. His research projects cover a wide spectrum, from regional farming studies to large-scale initiatives like the ESA WorldCereal project, always aiming to bridge the gap between cutting-edge machine learning technologies, the reality of large-scale operationalization, and the science of remote sensing.

TBD

Keynote 2 – Patrick Beukema, Allen.ai

“AGI can wait: Building scalable AI to solve urgent environmental problems”

Bio

Patrick Beukema, Ph.D. is Principal AI engineer at the Allen Institute for Artificial Intelligence (AI2), where he established and leads a team focused on AI for earth observation and environmental monitoring. His work applies state-of-the-art deep learning strategies, including remote sensing and GPS-based modeling, to develop globally deployed, real-time solutions for conservation and sustainability. With a focus on robust engineering, his team emphasizes deploying models at scale in real-world environments to ensure rigorous validation and to assess real-time impact and effectiveness. He holds a Ph.D. in Neuroscience from the University of Pittsburgh, an M.S. from Carnegie Mellon University, and a B.A. from McGill University.

Challenge

This year’s data challenge is organized by the Embed2Scale consortium and revolves around Lossy Neural Compression for Geospatial Analytics. Specifically, participants are asked to develop an encoder to compress SSL4EO-S12 data cubes down to 1024 features, a.k.a. embeddings, that are then evaluated on a hidden set of applications, a.k.a. downstream tasks as in the literature of foundation models.

You will have three weeks in March 2025 to develop your encoder and test the quality of your embeddings by interacting with this challenge portal Eval.AI submitting embeddings for data available at HuggingFace. The 1st week of April will allow you three submissions in three days based on a separate dataset made public on HuggingFace two days before. This phase determines the final leaderboard and challenge winner who will present their solution at the workshop in Nashville or remotely as per availability.

Important Dates:

  • Development phase: Mar 10-31

  • Testing phase: Apr 3-5

  • Winner presentation: At the EarthVision WS

Webpage: eval.ai/web/challenges/challenge-page/2465

Challenge Data in HuggingFace

Development Phase ended, Final Submission for Ranking ahead!

The development phase closed. Now get ready to compress approx. 90GB from huggingface.co/datasets/embed2scale/SSL4EO-S12-downstream/tree/main/data_eval (available for download Apr 1) for submission starting Apr 3 through Apr 5.
Your scores will determine the final ranking to find our winners (cf. below). We are excited about your solution, all the best luck!

About CVPR EARTHVISION winners

In addition to the winner according to github.com/DLR-MF-DAS/embed2scale-challenge-supplement?tab=readme-ov-file#leader… we decided to also invite the solution with the highest `q_mean` score to present their solution at the CVPR EarthVision workshop. Each of the winning teams will receive a cash prize of 1k EUR as support to come to Nashville, TN, USA for the presentation. On April 7, 2025 we’ll get in touch with the two winning teams through the email ID they provided to Eval.AI .

And even should you not win the thing …

… we are so grateful you participated! We are also very happy to link the code to your solution. If you’d like so, pls open a corresponding issue in github.com/DLR-MF-DAS/embed2scale-challenge-supplement/issues.

 

Organizers

  • Ronny Hänsch, German Aerospace Center, Germany,
  • Devis Tuia, EPFL, Switzerland,
  • Jan Dirk Wegner, University of Zurich & ETH Zurich, Switzerland,
  • Nathan Jacobs, Washington University in St. Louis, USA
  • Loïc Landrieu, ENPC ParisTech, France
  • Charlotte Pelletier, UBS Vannes, France
  • Hannah Kerner, Arizona State University, USA

Technical Committee

  • Aayush Dhakal, Washington University in St Louis
  • Aimi Okabayashi, Université Bretagne Sud
  • Alex Levering, VU Amsterdam
  • Alexandre Xavier Falcao, IC-UNICAMP
  • Amanda Bright, National Geospatial-Intelligence Agency
  • Anastasia Schlegel, DLR
  • Ankit Jha, The LNM Institute of Information Technology, Jaipur
  • Antoine Bralet, UBS/IRISA
  • Begum Demir, TU Berlin
  • Biplab Banerjee, Indian Institute of Technology, Bombay
  • Caleb Robinson, Microsoft
  • Camille Couprie, Facebook
  • Camille Kurtz, Université Paris Cité
  • Christian Heipke, Leibniz Universität Hannover
  • Christopher R Ratto, JHUAPL
  • Claudio Persello, University of Twente
  • Clement Mallet, IGN, France
  • Conrad M Albrecht, German Aerospace Center
  • Corentin Dufourg, Univ. Bretagne Sud / IRISA
  • Dalton Lunga, Oak Ridge National Laboratory
  • Damien Robert, University of Zurich
  • Daniel Iordache, VITO, Belgium
  • Diego Marcos, Inria
  • Dimitri Gominski, University of Copenhagen
  • Elliot Vincent, Ecole nationale des ponts et chaussées / Inria / IGN
  • Emanuele Dalsasso, EPFL
  • ESTHER ROLF, Harvard University
  • Ewelina Rupnik, Univ Gustave Eiffel, LASTIG, ENSG-IGN
  • Ferda Ofli, Qatar Computing Research Institute, HBKU
  • Franz Rottensteiner, Leibniz Universitat Hannover, Germany
  • Gabriel Tseng, NASA Harvest
  • Gabriele Moser, Università di Genova
  • Gedeon Muhawenayo, Arizona State University
  • Gemine Vivone, CNR-IMAA
  • Gencer Sumbul, EPFL
  • Georgios Voulgaris, University of Oxford
  • Guillaume Astruc, ENPC/IGN
  • Gülşen Taşkın, İstanbul Teknik Üniversitesi
  • Gustau Camps-Valls, Universitat de València
  • Helmut Mayer, Bundeswehr University Munich
  • Islam Mansour, German Aerospace Center (DLR) & ETH Zurich
  • Jacob Arndt, Oak Ridge National Laboratory
  • Jakob Gawlikowski, German Aerospace Center (DLR)
  • Javiera Castillo Navarro, Cnam
  • Joëlle Hanna, University of St. Gallen
  • Jonathan Giezendanner, MIT
  • Linus M. Scheibenreif, University of St. Gallen
  • Lukas Drees, University of Zurich
  • M. Usman Rafique, Kitware Inc.
  • Marc Rußwurm, Wageningen University
  • Mareike Dorozynski, Institute of Photogrammetry and Geoinformation
  • Martin Weinmann, Karlsruhe Institute of Technology
  • Matt Leotta, Kitware
  • Michael Mommert, Stuttgart University of Applied Sciences
  • Michael Schmitt, University of the Bundeswehr Munich
  • Miguel-Ángel Fernández-Torres, Universidad Carlos III de Madrid
  • Minh-Tan Pham, IRISA-UBS
  • Myron Z Brown, JHU
  • Nicolas Audebert, IGN
  • Nikolaos Dionelis, ESA
  • Ragini Bal Mahesh, German Aerospace Center – DLR
  • Ramesh S. Nair, Planet Labs PBC
  • Roberto Interdonato, CIRAD
  • Scott Workman, DZYNE Technologies
  • Sophie Giffard-Roisin, Univ. Grenoble Alpes
  • Srikumar Sastry, Washington University in St Louis
  • Subit Chakrabarti, Floodbase
  • Sylvain Lobry, Université Paris Cité
  • Tanya Nair, Floodbase
  • Tatsumi Uezato, Hitachi, Ltd
  • Teng Wu, Univ Gustave Eiffel, ENSG, IGN, LASTIG, F-94160 Saint-Mandé
  • Thibaud Ehret, AMIAD, Pôle recherche
  • Utkarsh Mall, Columbia University
  • Valerio Marsocci, KU Leuven
  • Yohann PERRON, LIGM ENPC
  • Zhijie Zhang, University of Arizona
  • Zhuangfang Yi, Regrow

Challenge Sponsors

 

Affiliations

         .

Submissions

1. Prepare the anonymous, 8-page (references excluded) submission using the ev2025-template and following the paper guidelines. 

2. Submit at cmt3.research.microsoft.com/EarthVision2025. Please submit via this link only.

Policies

A complete paper should be submitted using the EarthVision templates provided above.

Reviewing is double blind, i.e. authors do not know the names of the reviewers and reviewers do not know the names of the authors. Please read Section 1.7 of the example paper EarthVision2025.pdf for detailed instructions on how to preserve anonymity. Avoid providing acknowledgments or links that may identify the authors.

Papers are to be submitted using the dedicated submission platform: cmt3.research.microsoft.com/EarthVision2025.

The submission deadline is strict.

By submitting a manuscript, the authors guarantee that it has not been previously published or accepted for publication in a substantially similar form. CVPR rules regarding plagiarism, double submission, etc. apply.

The Microsoft CMT service was used for managing the peer-reviewing process for this conference. This service was provided for free by Microsoft and they bore all expenses, including costs for Azure cloud services as well as for software development and support.

CVPR 2025

CVPR is the premier annual computer vision event comprising the main conference and several co-located workshops and short courses. With its high quality and low cost, it provides an exceptional value for students, academics and industry researchers.

Learn More: CVPR 2025