Dataset Viewer
Auto-converted to Parquet Duplicate
id
stringlengths
3
6
image
imagewidth (px)
443
794
annotations
listlengths
2
28
300825
[ { "category_id": 1, "bbox": [ 304.7200012207031, 616.52001953125, 233.8800048828125, 118.91000366210938 ], "area": 27804.8828125, "iscrowd": 0, "id": 2931091, "image_id": 300825, "segmentation": [ [ 304.7200012207031, 616.52001953125, ...
177003
[ { "category_id": 1, "bbox": [ 63.209999084472656, 252.77000427246094, 467.30999755859375, 167.10000610351562 ], "area": 70943.359375, "iscrowd": 0, "id": 1716020, "image_id": 177003, "segmentation": [ [ 63.209999084472656, 252.77000427246...
338740
[ { "category_id": 1, "bbox": [ 172.13999938964844, 121.72000122070312, 8.579999923706055, 7.130000114440918 ], "area": 61.16270065307617, "iscrowd": 0, "id": 3295876, "image_id": 338740, "segmentation": [ [ 172.13999938964844, 121.72000122...
338741
[ { "category_id": 1, "bbox": [ 56.689998626708984, 88.54000091552734, 233.9499969482422, 214.89999389648438 ], "area": 49239.28125, "iscrowd": 0, "id": 3295879, "image_id": 338741, "segmentation": [ [ 56.689998626708984, 88.54000091552734,...
166403
[ { "category_id": 1, "bbox": [ 63.209999084472656, 264.6400146484375, 459.04998779296875, 47.22999954223633 ], "area": 19393.13671875, "iscrowd": 0, "id": 1617695, "image_id": 166403, "segmentation": [ [ 63.209999084472656, 264.64001464843...
166548
[ { "category_id": 1, "bbox": [ 56.689998626708984, 88.54000091552734, 233.8800048828125, 190.9199981689453 ], "area": 44416.4296875, "iscrowd": 0, "id": 1619095, "image_id": 166548, "segmentation": [ [ 56.689998626708984, 88.54000091552734...
166549
[ { "category_id": 1, "bbox": [ 56.689998626708984, 88.54000091552734, 233.9199981689453, 274.8699951171875 ], "area": 63534.3984375, "iscrowd": 0, "id": 1619112, "image_id": 166549, "segmentation": [ [ 56.689998626708984, 88.54000091552734...
166519
[ { "category_id": 1, "bbox": [ 62.70000076293945, 716.5999755859375, 297.19000244140625, 9.109999656677246 ], "area": 2707.620849609375, "iscrowd": 0, "id": 1618847, "image_id": 166519, "segmentation": [ [ 62.70000076293945, 716.5999755859...
166478
[ { "category_id": 1, "bbox": [ 56.689998626708984, 244.38999938964844, 233.9600067138672, 178.9499969482422 ], "area": 40948.79296875, "iscrowd": 0, "id": 1618452, "image_id": 166478, "segmentation": [ [ 56.689998626708984, 244.38999938964...
166479
[ { "category_id": 1, "bbox": [ 56.689998626708984, 400.4599914550781, 233.9199981689453, 46.91999816894531 ], "area": 9840.064453125, "iscrowd": 0, "id": 1618466, "image_id": 166479, "segmentation": [ [ 56.689998626708984, 400.459991455078...
166520
[ { "category_id": 1, "bbox": [ 62.70000076293945, 708.5599975585938, 455.30999755859375, 19.079999923706055 ], "area": 6517.734375, "iscrowd": 0, "id": 1618849, "image_id": 166520, "segmentation": [ [ 62.70000076293945, 708.5599975585938, ...
166480
[ { "category_id": 1, "bbox": [ 56.689998626708984, 244.44000244140625, 233.9199981689453, 178.9499969482422 ], "area": 40774.40234375, "iscrowd": 0, "id": 1618478, "image_id": 166480, "segmentation": [ [ 56.689998626708984, 244.44000244140...
127663
[ { "category_id": 1, "bbox": [ 63.209999084472656, 238.67999267578125, 462.1000061035156, 23.190000534057617 ], "area": 7716.37744140625, "iscrowd": 0, "id": 1240052, "image_id": 127663, "segmentation": [ [ 63.209999084472656, 238.67999267...
131382
[ { "category_id": 1, "bbox": [ 58.04999923706055, 61.58000183105469, 239.1199951171875, 136.11000061035156 ], "area": 31039.119140625, "iscrowd": 0, "id": 1276792, "image_id": 131382, "segmentation": [ [ 58.04999923706055, 61.5800018310546...
131383
[ { "category_id": 1, "bbox": [ 58.04999923706055, 401, 239.1300048828125, 18.81999969482422 ], "area": 3731.926025390625, "iscrowd": 0, "id": 1276813, "image_id": 131383, "segmentation": [ [ 58.04999923706055, 401, 297.179992675781...
131384
[ { "category_id": 1, "bbox": [ 58.04999923706055, 288.17999267578125, 239.1300048828125, 136.05999755859375 ], "area": 30460.646484375, "iscrowd": 0, "id": 1276826, "image_id": 131384, "segmentation": [ [ 67.01000213623047, 288.17999267578...
131385
[ { "category_id": 1, "bbox": [ 58.04999923706055, 671.77001953125, 239.07000732421875, 18.8700008392334 ], "area": 2479.2685546875, "iscrowd": 0, "id": 1276840, "image_id": 131385, "segmentation": [ [ 67.01000213623047, 671.77001953125, ...
131386
[ { "category_id": 1, "bbox": [ 58.060001373291016, 314.2699890136719, 239.13999938964844, 113.44000244140625 ], "area": 25646.84375, "iscrowd": 0, "id": 1276845, "image_id": 131386, "segmentation": [ [ 67.01000213623047, 314.2699890136719,...
305969
[ { "category_id": 1, "bbox": [ 56.689998626708984, 256.739990234375, 512.780029296875, 119.81999969482422 ], "area": 58849.859375, "iscrowd": 0, "id": 2982088, "image_id": 305969, "segmentation": [ [ 56.689998626708984, 256.739990234375, ...
306552
[ { "category_id": 1, "bbox": [ 303.05999755859375, 476.32000732421875, 246.63999938964844, 21.489999771118164 ], "area": 3006.80908203125, "iscrowd": 0, "id": 2987649, "image_id": 306552, "segmentation": [ [ 303.05999755859375, 476.3200073...
End of preview. Expand in Data Studio

YAML Metadata Warning:The task_ids "object-detection" is not in the official list: acceptability-classification, entity-linking-classification, fact-checking, intent-classification, language-identification, multi-class-classification, multi-label-classification, multi-input-text-classification, natural-language-inference, semantic-similarity-classification, sentiment-classification, topic-classification, semantic-similarity-scoring, sentiment-scoring, sentiment-analysis, hate-speech-detection, text-scoring, named-entity-recognition, part-of-speech, parsing, lemmatization, word-sense-disambiguation, coreference-resolution, extractive-qa, open-domain-qa, closed-domain-qa, news-articles-summarization, news-articles-headline-generation, dialogue-modeling, dialogue-generation, conversational, language-modeling, text-simplification, explanation-generation, abstractive-qa, open-domain-abstractive-qa, closed-domain-qa, open-book-qa, closed-book-qa, text2text-generation, slot-filling, masked-language-modeling, keyword-spotting, speaker-identification, audio-intent-classification, audio-emotion-recognition, audio-language-identification, multi-label-image-classification, multi-class-image-classification, face-detection, vehicle-detection, instance-segmentation, semantic-segmentation, panoptic-segmentation, image-captioning, image-inpainting, image-colorization, super-resolution, grasping, task-planning, tabular-multi-class-classification, tabular-multi-label-classification, tabular-single-column-regression, rdf-to-text, multiple-choice-qa, multiple-choice-coreference-resolution, document-retrieval, utterance-retrieval, entity-linking-retrieval, fact-checking-retrieval, univariate-time-series-forecasting, multivariate-time-series-forecasting, visual-question-answering, document-question-answering, pose-estimation

PubLayNet_mini Dataset

A diverse mini subset of the PubLayNet dataset with 500 samples for document layout analysis evaluation.

Dataset Details

  • Total Samples: 500 document images
  • Source: PubLayNet training set (146,874 total samples)
  • Task: Document Layout Analysis
  • Format: Parquet with embedded images and annotations
  • Image Size: 612×792 pixels (RGB)
  • Categories: 5 layout element types

Categories

The dataset contains annotations for 5 categories of document layout elements:

  1. Text (1): Regular text blocks and paragraphs
  2. Title (2): Document titles and headings
  3. List (3): Bulleted or numbered lists
  4. Table (4): Tabular data structures
  5. Figure (5): Images, charts, and diagrams

Features

Each sample contains:

  • id: Unique document identifier
  • image: Document image (PIL Image) - automatically loaded from embedded bytes
  • annotations: List of layout element annotations with:
    • category_id: Element type (1-5)
    • bbox: Bounding box coordinates [x, y, width, height]
    • area: Area of the bounding box
    • iscrowd: Whether the annotation is for a crowd of objects
    • id: Unique annotation identifier
    • image_id: Reference to the document image
    • segmentation: Polygon segmentation mask

Data Storage

Images are stored as embedded bytes in the parquet file and automatically converted to PIL Images when loaded. This ensures:

  • Self-contained dataset (no external image dependencies)
  • Fast loading and processing
  • Compatibility with HuggingFace datasets library

Category Distribution

This subset maintains diverse representation across categories:

  • Text: ~3,676 annotations
  • Title: ~1,000 annotations
  • List: ~73 annotations
  • Table: ~128 annotations
  • Figure: ~172 annotations

Usage

from datasets import load_dataset

# Load the dataset
dataset = load_dataset("kenza-ily/publaynet-mini")

# Each sample contains:
for sample in dataset['train']:
    print(f"Document ID: {sample['id']}")
    print(f"Number of layout elements: {len(sample['annotations'])}")
    
    # Access the image (automatically converted to PIL Image)
    image = sample['image']  # PIL Image object
    print(f"Image size: {image.size}")
    
    # Access annotations
    for ann in sample['annotations']:
        category = ann['category_id']
        bbox = ann['bbox']
        segmentation = ann['segmentation']
        print(f"Element {category}: bbox={bbox}")

Loading from Parquet

You can also load the data directly from the parquet file:

import pyarrow.parquet as pq
import pandas as pd
from PIL import Image as PILImage
import io

# Read parquet file
table = pq.read_table("publaynet_mini.parquet")
df = table.to_pandas()

# Convert images from bytes to PIL Images
def convert_image(img_data):
    if isinstance(img_data, dict) and 'bytes' in img_data:
        img_bytes = img_data['bytes']
        return PILImage.open(io.BytesIO(img_bytes))
    return img_data

df['image'] = df['image'].apply(convert_image)

# Access data
for idx, row in df.iterrows():
    image = row['image']  # PIL Image
    annotations = row['annotations']  # List of dicts

Citation

If you use this dataset, please cite both the original PubLayNet paper and the DISCO paper, which introduces this evaluation subset.


@inproceedings{zhong2019publaynet,
  title={PubLayNet: Largest Dataset Ever for Document Layout Analysis},
  author={Zhong, Xu and Tang, Jianbin and Yepes, Antonio Jimeno},
  booktitle={2019 International Conference on Document Analysis and Recognition (ICDAR)},
  pages={1015--1022},
  year={2019},
  organization={IEEE}
}

@inproceedings{benkirane2026disco,
  title={{DISCO}: Document Intelligence Suite for Comparative Evaluation},
  author={Benkirane, Kenza and Asenov, Martin and Goldwater, Daniel and Ghodsi, Aneiss},
  booktitle={ICLR 2026 Workshop on Multimodal Intelligence},
  year={2026},
  url={https://openreview.net/forum?id=Bb9vBASVzX}
}

License

This subset follows the original PubLayNet dataset license.

Downloads last month
18

Collection including kenza-ily/publaynet_disco