Skip to content

ICText dataset. The largest dataset for text detection and recognition on integrated circuits with a total of 10,000 images and 100,152 characters annotated.

License

Notifications You must be signed in to change notification settings

chunchet-ng/ICText-AGCL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 

Repository files navigation

ICText - Integrated Circuit Text Dataset (Official Site)

Released on September 01, 2023

Official repository of the paper: "When IC Meets Text: Towards a Rich Annotated Integrated Circuit Text Dataset"

Authored By: Chun Chet Ng*, Che-Tsung Lin*, Zhi Qin Tan, Xinyu Wang, Jie Long Kew, Chee Seng Chan, Christopher Zach

*Equal Contribution

Project Page | Paper | ICText Dataset | AGCL | Bibtex

Abstract:

Automated Optical Inspection (AOI) is a process that uses cameras to autonomously scan printed circuit boards for quality control. Text is often printed on chip components, and it is crucial that this text is correctly recognized during AOI, as it contains valuable information. In this paper, we introduce ICText, the largest dataset for text detection and recognition on integrated circuits. Uniquely, it includes labels for character quality attributes such as low contrast, blurry, and broken. While loss-reweighting and Curriculum Learning (CL) have been proposed to improve object detector performance by balancing positive and negative samples and gradually training the model from easy to hard samples, these methods have had limited success with one-stage object detectors commonly used in industry. To address this, we propose Attribute-Guided Curriculum Learning (AGCL), which leverages the labeled character quality attributes in ICText. Our extensive experiments demonstrate that AGCL can be applied to different detectors in a plug-and-play fashion to achieve higher Average Precision (AP), significantly outperforming existing methods on ICText without any additional computational overhead during inference. Furthermore, we show that AGCL is also effective on the generic object detection dataset Pascal VOC.

Integrated Circuit Text (ICText) Dataset

The ViTrox-ORCR, ICText, and ICText-LT datasets can be accessed by following the steps outlined below.

Access to Dataset

To access the dataset, you must first download and sign a Non-Disclosure Agreement (NDA) form. A sample is also provided in the shared folder. Please refer to the following section for more details. Once you have signed the NDA, email it to us at ictext at vitrox.com. After verifying the signed NDA, we will grant you access to download the dataset.

Signing of NDA

You must sign the NDA electronically using Adobe Reader (online or offline) and send us the signed NDA in pdf format. The NDA should not be in the form of pictures or scanned handwritten documents. For participants without any affiliation, you must fill in "{YOUR NAME} - Individual Participant" in the affiliation field. Please fill in the relevant details on page 2, under the line '[Representative, Affiliation, Address of LICENSEE]' and sign at the 'Signature LICENSEE' section using black ink. If you have an affiliation, rename the NDA as '{AFFILIATION}_ICTEXT_LICENSE_AGREEMENT.pdf'. Otherwise, rename it as '{YOUR_NAME}_ICTEXT_LICENSE_AGREEMENT.pdf' before emailing it to us. NDAs that are not filled in according to the stated rules will be rejected.

ViTrox-OCR Dataset

The ViTrox-OCR dataset was introduced in the previous ICDAR 2021 Competition on Integrated Circuit Text Spotting and Aesthetic Assessment competition:

  1. Paper
  2. Competition Page
  3. RRC-ICText Eval GitHub Repo
  4. RRC-ICText Online Evaluation Platform

Please note that some annotations are outdated. Use this dataset only if you want to evaluate your trained model on the private validation and testing sets on the aforementioned eval.ai platform.

ICText Dataset

For the latest version of the ICText dataset, images and annotations are sourced entirely from the training set of ViTrox-OCR and split into training and validation sets at a 7:3 ratio. You can refer to our upcoming paper for more information.

ICText Annotation JSON Format

{
    "images": [
        {
        "id": int,
        "file_name" : str,
        "height": int,
        "width": int
        },
        ...
    ],
    "annotations": [
        {
        "id": int,
        "image_id": int,
        "bbox": [x1,y1,x2,y2,x3,y3,x4,y4],
        "category_id": int,
        "aesthetic": [one_hot_encoded],
        "area": float,
        "ignore": 0 or 1
        },
        ...
    ],
    "categories": [
        {
        "id": int, 
        "name": str
        }
        ...
    ]
}

Please note that the aesthetic label follows the sequence of low contrast, blurry, and broken. Given that the annotation for ICText follow MSCOCO's format closely, you can evaluate the predictions on ICText accordingly using this COCO Results Analysis tool.

ICText Statistics

Subset Images Perfect Text Difficult Text Total Text
Train 7,000 19,153 50,597 69,750
Test 3,000 10,288 20,114 30,402

ICText-LT Dataset

ICText-LT is a long-tailed character classification dataset introduced in the work of FFDS-Loss. Images are cropped from ICText and resampled to form this dataset.

ICText-LT Annotation TXT Format

file_name, char_class

ICText-LT Statistics

Subset Imbalance Factor Images
Train 18 68,307
Train 100 25,711
Test N/A 6,300

Attribute-Guided Curriculum Learning (AGCL)

In this work, we utilize the publicly available repositories for several object detection architectures:

  1. YOLOv4
  2. YOLOv5
  3. YOLOX
  4. EfficientDet

The pre-trained weights are not released to protect the intellectual property of ViTrox Corporation Berhad. However, a Jupyter Notebook is provided to detail the implementation of AGCL, which can be easily integrated into the aforementioned object detectors. The code is written in PyTorch.

Citation

If you wish to cite the ICDAR 2021 Competition on Integrated Circuit Text Spotting and Aesthetic Assessment paper and its participants' methods:

@inproceedings{icdar2021_ictext,
  author= {Ng, Chun Chet
  and Nazaruddin, Akmalul Khairi Bin
  and Lee, Yeong Khang
  and Wang, Xinyu
  and Liu, Yuliang
  and Chan, Chee Seng
  and Jin, Lianwen
  and Sun, Yipeng
  and Fan, Lixin},
  title= {ICDAR 2021 Competition on Integrated Circuit Text Spotting and Aesthetic Assessment},
  booktitle= {Document Analysis and Recognition -- ICDAR 2021},
  year= {2021},
  publisher= {Springer International Publishing},
  pages= {663--677}}

If you wish to cite the latest version of the ICText dataset and AGCL, published at Pattern Recognition, Elsevier:

@article{ictext_agcl,
  author= {Ng, Chun Chet
  and Lin, Che-Tsung
  and Tan, Zhi Qin
  and Wang, Xinyu
  and Kew, Jie Long
  and Chan, Chee Seng
  and Zach, Christopher},
  title= {When IC meets text: Towards a rich annotated integrated circuit text dataset},
  journal = {Pattern Recognition},
  year= {2024},
  volume = {147},
  pages = {110124},
  issn = {0031-3203},
  publisher={Elsevier},
  doi = {https://doi.org/10.1016/j.patcog.2023.110124}}

Feedback

We welcome all suggestions and opinions (both positive and negative) on this work. Please contact us by sending an email to ngchunchet95 at gmail.com or cs.chan at um.edu.my.

Acknowledgement

We sincerely thank ViTrox Corporation Berhad for their support in annotating the ICText dataset. Our appreciation also goes to NVIDIA Corporation and ViTrox Corporation Berhad for providing essential computing resources.

License and Copyright

This project is open source under the BSD-3 license (refer to the LICENSE file for more information).

© 2023 Universiti Malaya.

About

ICText dataset. The largest dataset for text detection and recognition on integrated circuits with a total of 10,000 images and 100,152 characters annotated.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published