ICDAR 2023 Competition and CLEF 2023 Lab on Document Information Localization and Extraction

DocILE is a large-scale research benchmark for cross-evaluation of machine learning methods for Key Information Localization and Extraction (KILE) and Line Item Recognition (LIR) from semi-structured business documents such as invoices, orders etc. Such large-scale benchmark was previously missing (Skalický et al., 2022), hindering comparative evaluation.

Quick Links:
Join DocILE announcements google group to receive updates and news about the challenge
Form to get access to the dataset
Github repository
Dataset paper (Šimsa et al., 2023) with Supplementary Material
Make submissions to the benchmark via Robust Reading Competition website

CLEF 2023 DocILE Lab Workshop

DocILE workshop will be at CLEF 2023 with the following program:
Tuesday - September 19 | ROOM 3
All times are in local time (EEST)
16:00-16:05 Welcome
16:05-16:35 Invited talk: "Document Visual Question Answering – Challenges and Opportunities"
Dimosthenis Karatzas, Computer Vision Center, Universitat Autónoma de Barcelona
16:35-16:50 Overview of DocILE Lab Submissions
16:50-17:05 USTC-iFLYTEK at DocILE: A Multi-modal Approach Using Domain-specific GraphDoc
Yan Wang, Jun Du, Jiefeng Ma, Pengfei Hu, Zhenrong Zhang, Jianshu Zhang
17:05-17:20 Object Detection Pipeline Using YOLOv8 for Document Information Extraction
Jakub Straka, Ivan Gruber
17:20-17:30 Awards Ceremony

Note: Since this is the last session of the day and the schedule is quite tight, expect the session can go slightly overtime.

Competition and Prizes

Official Results

Congratulations to all participating teams. Links to the competition overview paper and participants working notes will appear here once the CEUR Workshop Proceedings are ready.

Method Team/Affiliation
KILE: 1st place GraphDoc (RRC) USTC-iFLYTEK
KILE: 2nd place YOLOv8 (RRC) UWB
KILE: 3rd place Union-RoBERTa (RRC) UIT@AICLUB_TAB
LIR: 1st place GraphDoc (RRC) USTC-iFLYTEK
LIR: 2nd place - -
LIR: 3rd place YOLOv8 (RRC) UWB
Best Paper GraphDoc (RRC) USTC-iFLYTEK

Note that the 2nd place for the LIR task was not awarded as only one method managed to beat the baselines in the RRC benchmark.

Competition Information

The DocILE'23 competition is running as a CLEF 2023 lab and an ICDAR 2023 competition with a single leaderboard. The deadline for submissions is on May 10 May 24, 2023. The competition comes with a prize pool of $9000:

  • Top three eligible teams on the KILE leaderboard will receive $2000, $1000 and $500 respectively.
  • Top three eligible teams on the LIR leaderboard will receive $2000, $1000 and $500 respectively.
  • A $2000 best-paper award, selected by the lab organizers and the steering committee.
In order to participate in the competition (and be eligible for prizes), you need to follow the competition rules. Most notably:
  • You need to register using the CLEF Labs Registration Form.
  • It is prohibited to use external document datasets and models trained on these datasets.
The competition/benchmark submissions will be handled via the DocILE page at the Robust Reading Competition website. The benchmark will stay live after the competition ends.


DocILE is the largest dataset of business documents with KILE and LIR labels for 106680 documents (6680 annotated and 100k synthetic) and almost 1M unlabeled documents for unsupervised pre-training.

Fill out the Dataset Access Request form to get access to the data for research purposes.

We provide rossumai/docile repository to easily load the data, visualize the annotations and to run the evaluation.

[Legal information on the processing of personal data for the purpose of scientific research.]

Dataset and Benchmark Paper

The dataset, the benchmark tasks and the evaluation criteria are described in detail in the dataset paper (Šimsa et al., 2023) which was accepted to ICDAR 2023. The provided link is to arXiv version that includes Supplementary Material. To cite the dataset, please use the following BibTeX entry:

    title={{DocILE} Benchmark for Document Information Localization and Extraction},
    author={{\v{S}}imsa, {\v{S}}t{\v{e}}p{\'a}n and {\v{S}}ulc, Milan and U{\v{r}}i{\v{c}}{\'a}{\v{r}}, Michal and Patel, Yash and Hamdi, Ahmed and Koci{\'a}n, Mat{\v{e}}j and Skalick{\`y}, Maty{\'a}{\v{s}} and Matas, Ji{\v{r}}{\'\i} and Doucet, Antoine and Coustaty, Micka{\"e}l and Karatzas, Dimosthenis},
    url = {https://arxiv.org/abs/2302.05658},
    journal={arXiv preprint arXiv:2302.05658},


Contact the organizers at: docile-2023-organizers@googlegroups.com

Task Chairs:
Milan Šulc, Head of Rossum AI Labs, Rossum.ai, Czech Republic.
Štěpán Šimsa, Researcher at Rossum AI Labs, Rossum.ai, Czech Republic.

Ahmed Hamdi, Associate Professor at the University of La Rochelle, France.
Yash Patel, PhD. candidate at the Visual Recognition Group, Czech Technical University in Prague, Czech Republic.
Matyáš Skalický, Research Engineer at Rossum.ai, Czech Republic.

Steering Committee:
Michal Uřičář, Researcher at Rossum AI Labs, Rossum.ai, Czech Republic.
Antoine Doucet, Full Professor of Computer Science at the University of La Rochelle, France.
Mickael Coustaty, Associate Professor at the University of La Rochelle, France.
Dimosthenis Karatzas, Associate Director of the Computer Vision Center, Barcelona, Spain.