1st CASTLE Grand Challenge

Archived

This challenge has been completed.

Important

The overview paper of the first CASTLE Grand Challenge at ACM Multimedia 2025 has been published in the proceedings of ACM Multimedia 2025: Overview of the First CASTLE Grand Challenge at ACM Multimedia 2025.

Congratulations to the multiXview team multiXview team for winning the interactive track!

1st CASTLE Challenge at ACM Multimedia 2025

The first CASTLE multimodal analytics challenge will be held at ACM Multimedia 2025 in Dublin, Ireland.

To express your interest in participating to the challenge, please fill this form.

To submit your paper and your results (for participants of the automatic track), please use submission form on OpenReview.

Timeline

Guidelines for Participants

Participants will be required to register and agree to the dataset usage policy. For details on the submission format, please see below.

Tasks

The inaugural edition of the CASTLE Challenge features a diverse set of tasks, including event detection, retrieval, and question answering. Future editions will expand the scope, but for this edition, the tasks include:

Given a textual description (in English), participants must identify all timeframes where a specific event occurs. Events should be reported with both a time range and a video ID.

Given a textual (in English) or visual (i.e., using an image) example of a physical object, participants must find all occurrences of that object across any of the video streams.

💬 Question Answering

Given a question in natural language (in English), participants must provide an answer. The response should be formulated in natural language and include references to relevant sensor streams and time intervals as supporting evidence.


Evaluation

The challenge will operate across two tracks: fully-automatic and interactive.

⚙️ Fully-Automatic Track

Participants receive queries in advance and generate results using any method they choose. These results are then submitted to the challenge organizers for evaluation. Please see the list of queries below.

🎮 Interactive Track

This track will be evaluated live during the conference. Participants must solve tasks synchronously and interactively within a limited timeframe. This format follows established competitions such as the Video Browser Showdown and the Lifelog Search Challenge.


Queries

🔍 Event Instance Search

📦 Object Instance Search

💬 Question Answering


Submission Format

⚙️ Fully-Automatic Track

For the fully-automatic track, please submit your results together with your paper on OpenReview. Results should be submitted in CSV format with one file per task type (i.e., three files, one each for the 🔍 Event Instance Search, 📦 Object Instance Search, and 💬 Question Answering tasks) in a ZIP file. The name of each file should indicate the task type. The columns for the files are listed below.

🔍 Event Instance Search

📦 Object Instance Search

💬 Question Answering

🎮 Interactive Track

The interactive track will be evaluated during a dedicated session at ACM Multimedia 2025 in Dublin. Participants are expected to be on-site for the interactive evaluation. Systems are supposed to submit their task solutions to the Distributed Retrieval Evaluation Server via its API.