MindCube Logo MindCube
CVinW @ CVPR 2026

MindCube Challenge

Spatial Question Answering from Limited Multi-View Observations

🏛️
Hosted by: The 5th Workshop on Computer Vision in the Wild (CVinW), CVPR 2026.

We are organizing the MindCube Challenge, a spatial question answering benchmark designed to evaluate spatial mental modeling from limited multi-view observations. Participants will be ranked by accuracy on a held-out test set.

Challenge Overview

Goal

Given a multi-view observation and a question, predict the correct answer for each example.

What You Do

  1. Train / fine-tune on the MindCube training set.
  2. Develop and validate on MindCube_tinybench.
  3. Run inference on the held-out test set (to be released) and submit predictions.

Data Splits

📚

Train

MindCube_train.jsonl

Public
🔬

Validation

MindCube_tinybench.jsonl

Public
🏆

Test (Held-out)

Final evaluation set

Coming Soon

Dataset: Data can be found at huggingface.co/datasets/MLL-Lab/MindCube

Format & loading: Please refer to the official instructions in the MindCube repository.

Evaluation

  • Metric: Accuracy (exact match) on the held-out test set.
  • Ranking: Teams are ranked by overall accuracy.
  • (Optional) We may additionally report accuracy by setting/sub-category for analysis.
  • Tie-break: Higher accuracy on a specific subset, then earlier submission time.

Challenge Leaderboard

Performance of submitted methods on the held-out test set.

Click on column headers to sort the results

Baseline Participants
Rank Team / Method Overall Rotation Among Around
- Random (chance) 32.35 36.36 32.29 30.66
- Random (frequency) 33.02 38.30 32.66 35.79
Challenge submissions coming soon...
Leaderboard will be updated after the test set is released and submissions are evaluated.

Submission

Submission File Format (JSONL)

Submit a single .jsonl file with one JSON object per line, containing:

  • question_id (string)
  • prediction (string or integer — choose one convention and keep it consistent)
Example submission format
{"question_id":"mc_000001","prediction":"B"}
{"question_id":"mc_000002","prediction":"A"}

Requirements

  • Provide exactly one prediction for each question_id in the test set.
  • Duplicate IDs: Keep last / invalid submission
  • Missing IDs: Count as incorrect / invalid submission
  • (Recommended) You may gzip the file for size: predictions.jsonl.gz

How to Submit

1

Download the held-out test set (coming soon)

2

Generate your predictions.jsonl following the required format.

3

Name the file as: TeamName_MethodName.jsonl (or .jsonl.gz)

4

Submit via upload form (coming soon) or email to qinengw@u.northwestern.edu with subject: MindCube Challenge Submission: TeamName

Submission Limit: Up to 3 submissions per team; last submission before the deadline counts
Deadline: TBD
Results Announcement: TBD

Rules

  • External data / models / APIs: Allowed or disallowed — TBD; if allowed, require disclosure
  • Human-in-the-loop labeling on test: Disallowed
  • Participants must not attempt to obtain test labels or manipulate evaluation.
  • Verification: Top teams may be asked to provide a brief method description and reproducibility details.

Baselines & Starter Kit

Baselines, data loaders, and evaluation scripts are available in the official MindCube repository:

github.com/mll-lab-nu/MindCube

Contact

For questions, please reach out via: