Datasets:
id stringlengths 3 6 | image imagewidth (px) 443 794 | annotations listlengths 2 28 |
|---|---|---|
300825 | [
{
"category_id": 1,
"bbox": [
304.7200012207031,
616.52001953125,
233.8800048828125,
118.91000366210938
],
"area": 27804.8828125,
"iscrowd": 0,
"id": 2931091,
"image_id": 300825,
"segmentation": [
[
304.7200012207031,
616.52001953125,
... | |
177003 | [
{
"category_id": 1,
"bbox": [
63.209999084472656,
252.77000427246094,
467.30999755859375,
167.10000610351562
],
"area": 70943.359375,
"iscrowd": 0,
"id": 1716020,
"image_id": 177003,
"segmentation": [
[
63.209999084472656,
252.77000427246... | |
338740 | [
{
"category_id": 1,
"bbox": [
172.13999938964844,
121.72000122070312,
8.579999923706055,
7.130000114440918
],
"area": 61.16270065307617,
"iscrowd": 0,
"id": 3295876,
"image_id": 338740,
"segmentation": [
[
172.13999938964844,
121.72000122... | |
338741 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
88.54000091552734,
233.9499969482422,
214.89999389648438
],
"area": 49239.28125,
"iscrowd": 0,
"id": 3295879,
"image_id": 338741,
"segmentation": [
[
56.689998626708984,
88.54000091552734,... | |
166403 | [
{
"category_id": 1,
"bbox": [
63.209999084472656,
264.6400146484375,
459.04998779296875,
47.22999954223633
],
"area": 19393.13671875,
"iscrowd": 0,
"id": 1617695,
"image_id": 166403,
"segmentation": [
[
63.209999084472656,
264.64001464843... | |
166548 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
88.54000091552734,
233.8800048828125,
190.9199981689453
],
"area": 44416.4296875,
"iscrowd": 0,
"id": 1619095,
"image_id": 166548,
"segmentation": [
[
56.689998626708984,
88.54000091552734... | |
166549 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
88.54000091552734,
233.9199981689453,
274.8699951171875
],
"area": 63534.3984375,
"iscrowd": 0,
"id": 1619112,
"image_id": 166549,
"segmentation": [
[
56.689998626708984,
88.54000091552734... | |
166519 | [
{
"category_id": 1,
"bbox": [
62.70000076293945,
716.5999755859375,
297.19000244140625,
9.109999656677246
],
"area": 2707.620849609375,
"iscrowd": 0,
"id": 1618847,
"image_id": 166519,
"segmentation": [
[
62.70000076293945,
716.5999755859... | |
166478 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
244.38999938964844,
233.9600067138672,
178.9499969482422
],
"area": 40948.79296875,
"iscrowd": 0,
"id": 1618452,
"image_id": 166478,
"segmentation": [
[
56.689998626708984,
244.38999938964... | |
166479 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
400.4599914550781,
233.9199981689453,
46.91999816894531
],
"area": 9840.064453125,
"iscrowd": 0,
"id": 1618466,
"image_id": 166479,
"segmentation": [
[
56.689998626708984,
400.459991455078... | |
166520 | [
{
"category_id": 1,
"bbox": [
62.70000076293945,
708.5599975585938,
455.30999755859375,
19.079999923706055
],
"area": 6517.734375,
"iscrowd": 0,
"id": 1618849,
"image_id": 166520,
"segmentation": [
[
62.70000076293945,
708.5599975585938,
... | |
166480 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
244.44000244140625,
233.9199981689453,
178.9499969482422
],
"area": 40774.40234375,
"iscrowd": 0,
"id": 1618478,
"image_id": 166480,
"segmentation": [
[
56.689998626708984,
244.44000244140... | |
127663 | [
{
"category_id": 1,
"bbox": [
63.209999084472656,
238.67999267578125,
462.1000061035156,
23.190000534057617
],
"area": 7716.37744140625,
"iscrowd": 0,
"id": 1240052,
"image_id": 127663,
"segmentation": [
[
63.209999084472656,
238.67999267... | |
131382 | [
{
"category_id": 1,
"bbox": [
58.04999923706055,
61.58000183105469,
239.1199951171875,
136.11000061035156
],
"area": 31039.119140625,
"iscrowd": 0,
"id": 1276792,
"image_id": 131382,
"segmentation": [
[
58.04999923706055,
61.5800018310546... | |
131383 | [
{
"category_id": 1,
"bbox": [
58.04999923706055,
401,
239.1300048828125,
18.81999969482422
],
"area": 3731.926025390625,
"iscrowd": 0,
"id": 1276813,
"image_id": 131383,
"segmentation": [
[
58.04999923706055,
401,
297.179992675781... | |
131384 | [
{
"category_id": 1,
"bbox": [
58.04999923706055,
288.17999267578125,
239.1300048828125,
136.05999755859375
],
"area": 30460.646484375,
"iscrowd": 0,
"id": 1276826,
"image_id": 131384,
"segmentation": [
[
67.01000213623047,
288.17999267578... | |
131385 | [
{
"category_id": 1,
"bbox": [
58.04999923706055,
671.77001953125,
239.07000732421875,
18.8700008392334
],
"area": 2479.2685546875,
"iscrowd": 0,
"id": 1276840,
"image_id": 131385,
"segmentation": [
[
67.01000213623047,
671.77001953125,
... | |
131386 | [
{
"category_id": 1,
"bbox": [
58.060001373291016,
314.2699890136719,
239.13999938964844,
113.44000244140625
],
"area": 25646.84375,
"iscrowd": 0,
"id": 1276845,
"image_id": 131386,
"segmentation": [
[
67.01000213623047,
314.2699890136719,... | |
305969 | [
{
"category_id": 1,
"bbox": [
56.689998626708984,
256.739990234375,
512.780029296875,
119.81999969482422
],
"area": 58849.859375,
"iscrowd": 0,
"id": 2982088,
"image_id": 305969,
"segmentation": [
[
56.689998626708984,
256.739990234375,
... | |
306552 | [
{
"category_id": 1,
"bbox": [
303.05999755859375,
476.32000732421875,
246.63999938964844,
21.489999771118164
],
"area": 3006.80908203125,
"iscrowd": 0,
"id": 2987649,
"image_id": 306552,
"segmentation": [
[
303.05999755859375,
476.3200073... |
YAML Metadata Warning:The task_ids "object-detection" is not in the official list: acceptability-classification, entity-linking-classification, fact-checking, intent-classification, language-identification, multi-class-classification, multi-label-classification, multi-input-text-classification, natural-language-inference, semantic-similarity-classification, sentiment-classification, topic-classification, semantic-similarity-scoring, sentiment-scoring, sentiment-analysis, hate-speech-detection, text-scoring, named-entity-recognition, part-of-speech, parsing, lemmatization, word-sense-disambiguation, coreference-resolution, extractive-qa, open-domain-qa, closed-domain-qa, news-articles-summarization, news-articles-headline-generation, dialogue-modeling, dialogue-generation, conversational, language-modeling, text-simplification, explanation-generation, abstractive-qa, open-domain-abstractive-qa, closed-domain-qa, open-book-qa, closed-book-qa, text2text-generation, slot-filling, masked-language-modeling, keyword-spotting, speaker-identification, audio-intent-classification, audio-emotion-recognition, audio-language-identification, multi-label-image-classification, multi-class-image-classification, face-detection, vehicle-detection, instance-segmentation, semantic-segmentation, panoptic-segmentation, image-captioning, image-inpainting, image-colorization, super-resolution, grasping, task-planning, tabular-multi-class-classification, tabular-multi-label-classification, tabular-single-column-regression, rdf-to-text, multiple-choice-qa, multiple-choice-coreference-resolution, document-retrieval, utterance-retrieval, entity-linking-retrieval, fact-checking-retrieval, univariate-time-series-forecasting, multivariate-time-series-forecasting, visual-question-answering, document-question-answering, pose-estimation
PubLayNet_mini Dataset
A diverse mini subset of the PubLayNet dataset with 500 samples for document layout analysis evaluation.
Dataset Details
- Total Samples: 500 document images
- Source: PubLayNet training set (146,874 total samples)
- Task: Document Layout Analysis
- Format: Parquet with embedded images and annotations
- Image Size: 612×792 pixels (RGB)
- Categories: 5 layout element types
Categories
The dataset contains annotations for 5 categories of document layout elements:
- Text (1): Regular text blocks and paragraphs
- Title (2): Document titles and headings
- List (3): Bulleted or numbered lists
- Table (4): Tabular data structures
- Figure (5): Images, charts, and diagrams
Features
Each sample contains:
id: Unique document identifierimage: Document image (PIL Image) - automatically loaded from embedded bytesannotations: List of layout element annotations with:category_id: Element type (1-5)bbox: Bounding box coordinates [x, y, width, height]area: Area of the bounding boxiscrowd: Whether the annotation is for a crowd of objectsid: Unique annotation identifierimage_id: Reference to the document imagesegmentation: Polygon segmentation mask
Data Storage
Images are stored as embedded bytes in the parquet file and automatically converted to PIL Images when loaded. This ensures:
- Self-contained dataset (no external image dependencies)
- Fast loading and processing
- Compatibility with HuggingFace datasets library
Category Distribution
This subset maintains diverse representation across categories:
- Text: ~3,676 annotations
- Title: ~1,000 annotations
- List: ~73 annotations
- Table: ~128 annotations
- Figure: ~172 annotations
Usage
from datasets import load_dataset
# Load the dataset
dataset = load_dataset("kenza-ily/publaynet-mini")
# Each sample contains:
for sample in dataset['train']:
print(f"Document ID: {sample['id']}")
print(f"Number of layout elements: {len(sample['annotations'])}")
# Access the image (automatically converted to PIL Image)
image = sample['image'] # PIL Image object
print(f"Image size: {image.size}")
# Access annotations
for ann in sample['annotations']:
category = ann['category_id']
bbox = ann['bbox']
segmentation = ann['segmentation']
print(f"Element {category}: bbox={bbox}")
Loading from Parquet
You can also load the data directly from the parquet file:
import pyarrow.parquet as pq
import pandas as pd
from PIL import Image as PILImage
import io
# Read parquet file
table = pq.read_table("publaynet_mini.parquet")
df = table.to_pandas()
# Convert images from bytes to PIL Images
def convert_image(img_data):
if isinstance(img_data, dict) and 'bytes' in img_data:
img_bytes = img_data['bytes']
return PILImage.open(io.BytesIO(img_bytes))
return img_data
df['image'] = df['image'].apply(convert_image)
# Access data
for idx, row in df.iterrows():
image = row['image'] # PIL Image
annotations = row['annotations'] # List of dicts
Citation
If you use this dataset, please cite both the original PubLayNet paper and the DISCO paper, which introduces this evaluation subset.
@inproceedings{zhong2019publaynet,
title={PubLayNet: Largest Dataset Ever for Document Layout Analysis},
author={Zhong, Xu and Tang, Jianbin and Yepes, Antonio Jimeno},
booktitle={2019 International Conference on Document Analysis and Recognition (ICDAR)},
pages={1015--1022},
year={2019},
organization={IEEE}
}
@inproceedings{benkirane2026disco,
title={{DISCO}: Document Intelligence Suite for Comparative Evaluation},
author={Benkirane, Kenza and Asenov, Martin and Goldwater, Daniel and Ghodsi, Aneiss},
booktitle={ICLR 2026 Workshop on Multimodal Intelligence},
year={2026},
url={https://openreview.net/forum?id=Bb9vBASVzX}
}
License
This subset follows the original PubLayNet dataset license.
- Downloads last month
- 18