Datasets:
metadata
license: mit
language:
- en
tags:
- table-structure-recognition
- table-detection
- ocr
- document-ai
- icdar
ICDAR-2013-Logical: A Line-Level Logical Conversion of the ICDAR 2013 Table Dataset
Dataset Description
This dataset is a converted and enhanced version of the ICDAR 2013 Table Competition dataset, specifically reformatted for modern Table Structure Recognition (TSR) and OCR tasks. 📜
The primary contribution of this version is the creation of a direct link between low-level OCR output and the table's logical structure. For each table, the dataset provides:
- A high-resolution cropped PNG image of the table region (rendered at 144 DPI).
- A detailed JSON file that maps each detected text line's physical bounding box to its logical grid coordinates (
[row_start, row_end, col_start, col_end]).
This format is ideal for training and evaluating Document AI models that perform OCR and table understanding concurrently.
How to Use
You can load an example by pairing the images from the cropped_images directory with the JSON annotations in logical_gt.
import json
from PIL import Image
from pathlib import Path
# Assume dataset is loaded or cloned locally
base_path = Path("./") # Path to the dataset directory
# Get a list of all examples
gt_files = list((base_path / "logical_gt").glob("*.json"))
example_file = gt_files[0]
# Load the annotation data
with open(example_file, 'r') as f:
annotations = json.load(f)
# Load the corresponding image
image_path = base_path / "cropped_images" / (example_file.stem + ".png")
image = Image.open(image_path)
# Display the first annotation for the first line of text
first_line = annotations[0]
print(f"Text: {first_line['text']}")
print(f"Bounding Box: {first_line['box']}")
print(f"Logical Coordinates: {first_line['logical_coords']}")
# image.show()