Document VQA
Answer natural language questions about document images, requiring understanding of text, layout, and visual elements.
Multimodal Document Understanding
Building on OmniDocBench and its upcoming extension OmniDocBench-Pro, the challenge provides a unified evaluation framework for document-centric multimodal tasks involving charts, tables, figures, layouts, and natural text.
The DataMFM Challenge focuses on multimodal document understanding, a core challenge at the intersection of vision, language, and structured reasoning. Building on OmniDocBench and its upcoming extension OmniDocBench-Pro, the challenge provides a unified evaluation framework for document-centric multimodal tasks involving charts, tables, figures, layouts, and natural text.
This challenge is part of the Emerging Directions in Data for Multimodal Foundation Models (DataMFM) workshop at CVPR 2026, which examines research directions including web-scale to world-scale data, agentic and self-generated data pipelines, and principled data selection and mixture design.
Coming Soon: Detailed challenge tracks and evaluation metrics will be announced. The challenge will cover document-centric multimodal tasks involving charts, tables, figures, layouts, and natural text.
Answer natural language questions about document images, requiring understanding of text, layout, and visual elements.
Extract data and answer questions from charts and graphs, including bar charts, line plots, and pie charts.
Parse table structures and extract cell contents from complex document tables.
Note: Challenge timeline will be announced soon. See the workshop dates for paper submission deadlines.
Coming Soon: The leaderboard will be available once the challenge officially launches. Stay tuned!
| Rank | Team | Score | Date |
|---|---|---|---|
| 1 | — | — | — |
| 2 | — | — | — |
| 3 | — | — | — |
The challenge data is based on OmniDocBench and its upcoming extension OmniDocBench-Pro. Download links and detailed documentation will be released when the challenge launches.
Submit your results as a JSON file:
Open to all researchers worldwide. No team size limit.
Max 3 per day. Final evaluation allows 2 submissions.
Top teams must submit technical report. External data must be disclosed.