DCVLR: Data Curation for Vision Language Reasoning

A NeurIPS 2025 Competition advancing the science of data curation for vision-language reasoning.

Join the Research Challenge!

Share your email address by clicking above and we'll send you competition updates, tips, and important announcements.

Sponsored by:

Overview

Reasoning models have become one of the most exciting areas of AI research. But most existing models are trained on data that is not publicly available, making it difficult to reproduce results and build on them. Furthermore, the modality of most frontier reasoning models remains language-only.

But with your help, that is about to change! DCVLR, to be hosted at NeurIPS 2025, is the first open-data, open-models, open-source competition for data curation in vision-language reasoning.

Your challenge? Curate 1K or 10K high-quality instruction-tuning examples using any strategyβ€”synthetic generation, smart filtering, or novel approaches. We handle the training and evaluation.

🌟

OPEN-DATA

The winning datasets from the competition will be made publicly available.

πŸ€–

OPEN-MODELS

Unlike previous competitions, we evaluate utility for frontier open-source models such as AI2's Molmo.

πŸ’»

OPEN-SOURCE

All of the code in the competition will be publicly released to enable future research.

Eligibility & Requirements

Who can participate and what you need to know

🌍 Who Can Participate

  • Up to 500 teams of size 1-20 can compete
  • Anyone can join a team; students, researchers, and industry professionals are particularly welcome
  • No model training experience required; just curate and submit a dataset!
  • Participants from all backgrounds encouraged; it's free to enter and free to compete.

πŸ“‹ Competition Rules

  • BYOD (Bring your own data) - use any dataset for training data curation
  • Teams must submit a write-up documenting their reproducible, scalable curation strategy
  • Teams must open-source any code used for data curation and submit a link to the GitHub repo
  • Detailed rules will be provided during the team signup period

🚫 Restrictions

  • Each individual can join only one team
  • (Free) OpenReview account signup required
  • Each team can submit a maximum of three datasets (1k or 10k tracks) to the leaderboard
  • Teams must submit datasets for content moderation and filtering (hateful and toxic content, decontamination with test sets, etc.) to compete

Key Dates

Key dates and milestones for the DCVLR competition

June 18, 2025

Release of Competition Materials

Competition website and starter kit (data, code, baselines) available for download.

July 1, 2025

Submission Portal Opens

Participants can begin submitting their curated datasets for evaluation.

October 1, 2025

Final Submission Deadline

Last day to submit curated datasets for the competition.

November 1, 2025

Results Announced

Leaderboard released with final results.

December 2025

NeurIPS 2025

Presentation of results and awards at the NeurIPS conference.

Competition Resources

The only thing you need in order to participate in DCVLR is (1) a reasoning dataset generated using (2) a reproducible curation strategy. That said, we provide all the resources you need to conduct your own experiments on your own compute. We expect the most successful teams will vet the quality of their datasets with experiments on small models before submitting to the leaderboard.

πŸ“Š

Starter Kit

Comprehensive starter kit with example datasets, training scripts, and best practices to help participants get started.

Access Starter Kit
πŸ’»

Training Scripts

Starting scripts for fine-tuning multiple vision-language models on your curated datasets with Oumi.

View Scripts
πŸ§ͺ

Evaluation Code

Scripts to evaluate model outputs on diverse benchmark development sets for local testing using VLMEvalKit.

Get Code
πŸ“

Baseline Submissions

Reference implementations and baseline approaches for data curation and model training (coming soon).

View Baselines
☁️

Compute Resources

GPU credits from our compute sponsor Lambda Labs for early student participants (coming soon).

Apply for Credits
πŸ“š

Documentation

Complete guides and tutorials to help you get started.

View Documentation

Prizes

Recognizing excellence in data curation for vision-language reasoning. These awards will be given for each track.

πŸ₯‡

First Place

$1,500 USD

Awarded to:

The best benchmarking team with a reproducible method

  • Podium presentation at NeurIPS 2025
  • Co-authorship on competition paper
  • Featured team spotlight
πŸ₯ˆ

IDC (Innovation in Data Curation) Prize

$1,000 USD

Awarded to:

The team with strong benchmark performance and an innovative method

  • Podium presentation at NeurIPS 2025
  • Co-authorship on competition paper
  • Featured team spotlight
πŸ₯‰

Honorable Mention

$250 USD

Awarded to:

The second-best benchmarking team with a reproducible method

  • Podium presentation at NeurIPS 2025
  • Co-authorship on competition paper
  • Featured team spotlight

Additional Recognition

Top 10 Teams

Certificate of Excellence and invitation to virtual presentation session

Student Teams

Special recognition for best all-student team submission

πŸš€ Quick Start

Get started with training in minutes

# install oumi
uv pip install "oumi[gpu]"

# checkout examples
git clone https://github.com/oumi-ai/oumi.git
ls configs/projects/dcvlr/starter-kit/

# train with your own config
oumi train -c $MY_CONFIG

Technical Specifications

Detailed requirements and specifications for the competition

πŸ€– Model Architecture

Evaluation Model: Undisclosed 8B-class model

  • Final evaluation uses an undisclosed model
  • Participants get scripts for multiple models
  • Good curation strategies should generalize
  • Multi-modal transformer architectures
  • Various parameter sizes available for development

πŸ“Š Dataset Specifications

Dataset Flexibility

  • Use any datasets for curation
  • Starter kit provides examples and guidance
  • Support for various image formats and resolutions
  • Flexible text formats and lengths
  • Focus on instruction-tuning quality over quantity

πŸ’» Hardware Requirements

Minimum Specifications

  • GPU: 1x NVIDIA A100-40GB or equivalent (4x A100-80GB GPUs recommended)
  • Students can apply for GPU credits from Lambda Labs
  • RAM: 64GB system memory
  • Storage: 500GB available space

Submission Format

How to structure and submit your curated datasets

1

Dataset Structure

Your submission must include:

submission/
β”œβ”€β”€ train_1k.jsonl|parquet  # or train_10k.jsonl|parquet
β”œβ”€β”€ metadata.json           # Team and method info
β”œβ”€β”€ technical_report.pdf    # 2-4 page report (optional)
└── README.md               # Instructions
2

Data Format

Each training example must follow this format:

[
    {
        # binary image data for parquet, base64 for jsonl
        "image": "encoded_image", 
        "prompt": "What is shown in this image?",
        "response": "The image shows a cat."
    },
    ...
]
        
3

Technical Report

Your report must include:

  • Data curation methodology
  • Selection criteria and rationale
  • Conversation generation approach
  • Computational resources used
  • Ablation studies (optional)
  • References and acknowledgments

Organizing Team

Professional headshot of Benjamin Feuer, Lead Organizer of the DCVLR competition from New York University (NYU).

Benjamin Feuer

Lead Organizer

NYU

Professional headshot of Rohun Tripathi, Lead Organizer of the DCVLR competition from the Allen Institute for AI.

Rohun Tripathi

Lead Organizer

Allen Institute for AI

Professional headshot of Oussama Elachqar, Lead Organizer of the DCVLR competition from Oumi.

Oussama Elachqar

Lead Organizer

Oumi

Professional headshot of Yuhui Zhang, Lead Organizer of the DCVLR competition from Stanford University.

Yuhui Zhang

Lead Organizer

Stanford University

Professional headshot of Nimrod Shabtay, Organizer of the DCVLR competition from IBM Research.

Nimrod Shabtay

Organizer

Tel Aviv University & IBM Research

Professional headshot of Neha Hulkund, Lead Organizer of the DCVLR competition from Massachusetts Institute of Technology (MIT).

Neha Hulkund

Lead Organizer

MIT

Professional headshot of Stefan Webb, Organizer of the DCVLR competition from Oumi.

Stefan Webb

Organizer

Oumi

Professional headshot of Thao Nguyen, Organizer of the DCVLR competition from the University of Washington.

Thao Nguyen

Organizer

University of Washington

Professional headshot of Vishaal Udandarao, Organizer of the DCVLR competition from The University of Tuebingen.

Vishaal Udandarao

Organizer

University of TΓΌbingen

Professional headshot of Xiaohan Wang, Organizer of the DCVLR competition from Stanford University.

Xiaohan Wang

Organizer

Stanford University

Professional headshot of Ludwig Schmidt, Organizer of the DCVLR competition from Stanford University.

Ludwig Schmidt

Organizer

Stanford University

Professional headshot of Saining Xie, Organizer of the DCVLR competition from NYU.

Saining Xie

Organizer

NYU

Professional headshot of Serena Yeung-Levy, Organizer of the DCVLR competition from Stanford University.

Serena Yeung-Levy

Organizer

Stanford University

Professional headshot of Paul Liang, Organizer of the DCVLR competition from MIT.

Paul Liang

Organizer

MIT

Professional headshot of Sara Beery, Organizer of the DCVLR competition from MIT.

Sara Beery

Organizer

MIT

Professional headshot of Georgia Gkioxari, Organizer of the DCVLR competition from Caltech.

Georgia Gkioxari

Organizer

Caltech

Leaderboard

Baseline performance on VMC-Bench reasoning subset. The private competition benchmark will use different evaluation metrics and datasets to ensure fair evaluation.

Rank Team Score Submissions Last Update
1 VisionCrafters 92.4% 2/3 2 hours ago
2 DataAlchemists 91.8% 3/3 5 hours ago
3 NeuralNavigators 91.2% 1/3 1 day ago
4 CurateAI 90.5% 2/3 1 day ago
5 VLM Masters 89.9% 1/3 2 days ago

The submission portal will open on July 1, 2025.

Sign up for updates

Frequently Asked Questions

Common questions about the DCVLR competition

What is the goal of this competition?

Who can participate?

What computational resources do I need?

How will submissions be evaluated?

Can I use external data or models?

Are there any prizes?

How many submissions can I make?

Contact Us

Have questions? Get in touch with the DCVLR team