Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset comes from Devashish Prasad, Ayan Gadpal, Kshitij Kapadni, Manish Visave, and Kavita Sultanpure - creators of CascadeTabNet.
Depending on the dataset version downloaded, the images will include annotations for 'borderless' tables, 'bordered' tables', and 'cells'. Borderless tables are those in which every cell in the table does not have a border. Bordered tables are those in which every cell in the table has a border, and the table is bordered. Cells are the individual data points within the table.
A subset of the full dataset, the ICDAR Table Cells Dataset, was extracted and imported to Roboflow to create this hosted version of the Cascade TabNet project. All the additional dataset components used in the full project are available here: All Files.
For the versions below: Preprocessing step of Resize (416by416 Fit within-white edges) was added along with more augmentations to increase the size of the training set and to make our images more uniform. Preprocessing applies to all images whereas augmentations only apply to training set images. 3. Version 3, augmented-FAST-model : 818 raw images of tables. Trained from Scratch (no transfer learning) with the "Fast" model from Roboflow Train. 3X augmentation (generated images). 4. Version 4, augmented-ACCURATE-model : 818 raw images of tables. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation. 5. Version 5, tableBordersOnly-augmented-FAST-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Fast" model from Roboflow Train. 3X augmentation. 6. Version 6, tableBordersOnly-augmented-ACCURATE-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation.
Example Image from the Datasethttps://i.imgur.com/ruizSQN.png" alt="Example Image from the Dataset">
Cascade TabNet in Actionhttps://i.imgur.com/nyn98Ue.png" alt="Cascade TabNet in Action">
CascadeTabNet is an automatic table recognition method for interpretation of tabular data in document images. We present an improved deep learning-based end to end approach for solving both problems of table detection and structure recognition using a single Convolution Neural Network (CNN) model. CascadeTabNet is a Cascade mask Region-based CNN High-Resolution Network (Cascade mask R-CNN HRNet) based model that detects the regions of tables and recognizes the structural body cells from the detected tables at the same time. We evaluate our results on ICDAR 2013, ICDAR 2019 and TableBank public datasets. We achieved 3rd rank in ICDAR 2019 post-competition results for table detection while attaining the best accuracy results for the ICDAR 2013 and TableBank dataset. We also attain the highest accuracy results on the ICDAR 2019 table structure recognition dataset.
If you find this work useful for your research, please cite our paper: @misc{ cascadetabnet2020, title={CascadeTabNet: An approach for end to end table detection and structure recognition from image-based documents}, author={Devashish Prasad and Ayan Gadpal and Kshitij Kapadni and Manish Visave and Kavita Sultanpure}, year={2020}, eprint={2004.12629}, archivePrefix={arXiv}, primaryClass={cs.CV} }
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is designed to annotate the structural elements of academic papers. It aims to train models to recognize different parts of a paper. Each class corresponds to a text or graphical element commonly found in papers.
Text indicating the name(s) of the author(s), typically found near the beginning of a document.
Identify the text block containing the author names. It usually follows the title and may include affiliations. Do not include titles, affiliations or titles of sections adjacent to author names.
Indicates a major division of the document, often labeled with a number and title.
Locate text labeled with "Chapter" followed by a number and title. Capture the entire heading, ensuring no unrelated text is included.
Symbols and numbers arranged to represent a mathematical concept.
Draw boxes around all mathematical expressions, excluding any accompanying text or numbers identifying the equations.
Numerals used to uniquely identify equations.
Identify numbers in parentheses next to equations. Do not include equation text or variables.
Visual content such as graphs, diagrams, code or images.
Outline the entire graphical representation. Do not include captions or any surrounding text.
Text providing a description or explanation above or below a figure.
Identify the text directly associated with a figure. Ensure no unrelated figures or text are included.
Clarifications or additional details located at the bottom of a page.
Locate text at the page's bottom that refers back to a mark or reference in the main text. Exclude any unrelated content.
Headings at the list of context text, identifying its purpose or content. This may also be called a list of figures.
Identify and label only the heading for lists in content sections. Do not include subsequent list items.
The detailed entries or points in a list. These often summarize all figures in the paper.
Identify each item in a content list. Exclude list headings and any non-list content.
Numerical indication of the current page.
Locate numbers typically positioned at the top or bottom margins. Do not include text or symbols beside the numbers.
Blocks of text separated by spacing or indentation.
Enclose individual text blocks that form coherent sections. Ensure each paragraph is distinguished separately.
Bibliographic information found typically in a reference sect
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is designed to annotate the structural elements of academic papers. It aims to train models to recognize different parts of a paper. Each class corresponds to a text or graphical element commonly found in papers.
Text indicating the name(s) of the author(s), typically found near the beginning of a document.
Identify the text block containing the author names. It usually follows the title and may include affiliations. Do not include titles or titles of sections adjacent to author names.
Indicates a major division of the document, often labeled with a number and title.
Locate text labeled with "Chapter" followed by a number and title. Capture the entire heading, ensuring no unrelated text is included.
Symbols and numbers arranged to represent a mathematical concept.
Draw boxes around all mathematical expressions, excluding any accompanying text or numbers identifying the equations.
Numerals used to uniquely identify equations.
Identify numbers in parentheses next to equations. Do not include equation text or variables.
Visual content such as graphs, diagrams, or images.
Outline the entire graphical representation. Do not include captions or any surrounding text.
Text providing a description or explanation of a figure.
Identify the text directly associated with a figure below it. Ensure no unrelated figures or text are included.
Clarifications or additional details located at the bottom of a page.
Locate text at the page's bottom that refers back to a mark or reference in the main text. Exclude any unrelated content.
Headings at the start of a list, identifying its purpose or content.
Identify and label only the heading for lists in content sections. Do not include subsequent list items.
The detailed entries or points in a list.
Identify each item in a content list. Exclude list headings and any non-list content.
Numerical indication of the current page.
Locate numbers typically positioned at the top or bottom margins. Do not include text or symbols beside the numbers.
Blocks of text separated by spacing or indentation.
Enclose individual text blocks that form coherent sections. Ensure each paragraph is distinguished separately.
Bibliographic information found typically in a reference section.
Identify the full reference entries. Ensure each citation is clearly distinguished without over
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset comes from Devashish Prasad, Ayan Gadpal, Kshitij Kapadni, Manish Visave, and Kavita Sultanpure - creators of CascadeTabNet.
Depending on the dataset version downloaded, the images will include annotations for 'borderless' tables, 'bordered' tables', and 'cells'. Borderless tables are those in which every cell in the table does not have a border. Bordered tables are those in which every cell in the table has a border, and the table is bordered. Cells are the individual data points within the table.
A subset of the full dataset, the ICDAR Table Cells Dataset, was extracted and imported to Roboflow to create this hosted version of the Cascade TabNet project. All the additional dataset components used in the full project are available here: All Files.
For the versions below: Preprocessing step of Resize (416by416 Fit within-white edges) was added along with more augmentations to increase the size of the training set and to make our images more uniform. Preprocessing applies to all images whereas augmentations only apply to training set images. 3. Version 3, augmented-FAST-model : 818 raw images of tables. Trained from Scratch (no transfer learning) with the "Fast" model from Roboflow Train. 3X augmentation (generated images). 4. Version 4, augmented-ACCURATE-model : 818 raw images of tables. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation. 5. Version 5, tableBordersOnly-augmented-FAST-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Fast" model from Roboflow Train. 3X augmentation. 6. Version 6, tableBordersOnly-augmented-ACCURATE-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation.
Example Image from the Datasethttps://i.imgur.com/ruizSQN.png" alt="Example Image from the Dataset">
Cascade TabNet in Actionhttps://i.imgur.com/nyn98Ue.png" alt="Cascade TabNet in Action">
CascadeTabNet is an automatic table recognition method for interpretation of tabular data in document images. We present an improved deep learning-based end to end approach for solving both problems of table detection and structure recognition using a single Convolution Neural Network (CNN) model. CascadeTabNet is a Cascade mask Region-based CNN High-Resolution Network (Cascade mask R-CNN HRNet) based model that detects the regions of tables and recognizes the structural body cells from the detected tables at the same time. We evaluate our results on ICDAR 2013, ICDAR 2019 and TableBank public datasets. We achieved 3rd rank in ICDAR 2019 post-competition results for table detection while attaining the best accuracy results for the ICDAR 2013 and TableBank dataset. We also attain the highest accuracy results on the ICDAR 2019 table structure recognition dataset.
If you find this work useful for your research, please cite our paper: @misc{ cascadetabnet2020, title={CascadeTabNet: An approach for end to end table detection and structure recognition from image-based documents}, author={Devashish Prasad and Ayan Gadpal and Kshitij Kapadni and Manish Visave and Kavita Sultanpure}, year={2020}, eprint={2004.12629}, archivePrefix={arXiv}, primaryClass={cs.CV} }