Document Understanding Dataset and Evaluation (DUDE) (2305.08455v3)
Abstract: We call on the Document AI (DocAI) community to reevaluate current methodologies and embrace the challenge of creating more practically-oriented benchmarks. Document Understanding Dataset and Evaluation (DUDE) seeks to remediate the halted research progress in understanding visually-rich documents (VRDs). We present a new dataset with novelties related to types of questions, answers, and document layouts based on multi-industry, multi-domain, and multi-page VRDs of various origins, and dates. Moreover, we are pushing the boundaries of current methods by creating multi-task and multi-domain evaluation setups that more accurately simulate real-world situations where powerful generalization and adaptation under low-resource settings are desired. DUDE aims to set a new standard as a more practical, long-standing benchmark for the community, and we hope that it will lead to future extensions and contributions that address real-world challenges. Finally, our work illustrates the importance of finding more efficient ways to model language, images, and layout in DocAI.
- Jordy Van Landeghem (6 papers)
- Łukasz Borchmann (17 papers)
- Michał Pietruszka (9 papers)
- Paweł Józiak (7 papers)
- Rafał Powalski (5 papers)
- Dawid Jurkiewicz (7 papers)
- Mickaël Coustaty (15 papers)
- Bertrand Ackaert (1 paper)
- Ernest Valveny (28 papers)
- Matthew Blaschko (26 papers)
- Sien Moens (1 paper)
- Tomasz Stanisławek (7 papers)
- Rubén Tito (1 paper)