<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://iapr-tc11.org/mediawiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Liwicki</id>
	<title>TC11 - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="http://iapr-tc11.org/mediawiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Liwicki"/>
	<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php/Special:Contributions/Liwicki"/>
	<updated>2026-04-21T15:43:37Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.16</generator>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets&amp;diff=2274</id>
		<title>Datasets</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets&amp;diff=2274"/>
		<updated>2016-05-16T09:24:01Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: revision for new site&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Important Notice=&lt;br /&gt;
[[Image:ThisWayToTheDatasets.png|250px|right|link=http://datasets.iapr-tc11.org| Datasets List]]&lt;br /&gt;
&lt;br /&gt;
The datasets are maintained at http://datasets.iapr-tc11.org The old dataset repository will remain accessible during [[Datasets List | here]]&lt;br /&gt;
&lt;br /&gt;
=Overview – Message from TC-11=&lt;br /&gt;
[[Image:ThisWayToTheDatasets_OldRepo.png|200px|right|link=http://datasets.iapr-tc11.org| Datasets List]]&lt;br /&gt;
&lt;br /&gt;
It is extremely important for the Document Image Analysis and Recognition community to be able to cross check and reproduce results described in published papers in the field. In order to achieve this, any datasets used as the basis for publications should be publicly available, as is the norm in many other disciplines.&lt;br /&gt;
&lt;br /&gt;
Authors are actively encouraged to submit the datasets they used to train and / or evaluate their algorithms to the TC-11 in order for them to be published on the TC-11 Web site.&lt;br /&gt;
&lt;br /&gt;
This initiative is not restricted to datasets. At TC-11 we are interested in archiving online any piece of data (ground-truth data, software, etc) which would allow to easily reproduce results, set new targets, foster healthy competition, encourage collaboration and generally advance the DIAR field as a whole.&lt;br /&gt;
&lt;br /&gt;
A wealth of datasets and corresponding ground truth data are already available through the TC-11 [http://datasets.iapr-tc11.org Web portal].&lt;br /&gt;
&lt;br /&gt;
If you wish to contribute, please read below about the procedure to submit material to the TC-11 web-portal. The dataset curators will be notified as soon as the dataset is uploaded.&lt;br /&gt;
For any comments or suggestions, please contact Marcus Liwicki, the dataset curator at marcus.liwicki (at) unifr.ch&lt;br /&gt;
&lt;br /&gt;
=Submission Protocol=&lt;br /&gt;
In order to submit a protocol please create an account on the TC11 datasets portal (http://datasets.iapr-tc11.org) and follow the online submission instructions. For any problems, please contact the dataset curators.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Copyright Note=&lt;br /&gt;
TC-11 provides dataset hosting services as a benefit to the international research community. If it is determined that copyrighted material is improperly included in a dataset submitted to inclusion on the TC-11 website, we will immediately remove the offending material upon notification of the copyright holder.&lt;br /&gt;
&lt;br /&gt;
By submitting a dataset for inclusion to the TC-11 Web site, the author certifies that he/she has the right to publish the dataset and any associated data in the public domain and the act of doing so does not violate intellectual property rights or copyrights of some third party.&lt;br /&gt;
&lt;br /&gt;
The TC-11 will provide a service through which the submitted dataset and any associated data will be made public to the Document Analysis community worldwide. In case any legal dispute arises in the future in relation to the publishing of this dataset and associated data in the public domain, the author will hold TC-11 free from any wrongdoing and accept responsibility for the publication of these data.&lt;br /&gt;
&lt;br /&gt;
By submitting a dataset and associated data to the TC-11, you explicitly accept that any third party can independently submit additional information that relates to the original dataset (e.g. additional ground-truth data, software, etc).&lt;br /&gt;
&lt;br /&gt;
We strongly encourage the authors, where they own the copyrights of the submitted information, to consider offering it to the community under a [http://creativecommons.org/choose/ creative commons license]. See [http://wiki.creativecommons.org/Before_Licensing/ this link] for guidelines about how to choose a proper Creative Commons license.&lt;br /&gt;
&lt;br /&gt;
=Useful Definitions=&lt;br /&gt;
'''Dataset''': A collection of data along with metadata information, as required to use these data.&lt;br /&gt;
&lt;br /&gt;
'''Metadata''': Metadata is information specific to a particular dataset. Metadata are usually tightly structured within the dataset itself (e.g. information encoded within the filenames of submitted images). Metadata can only be submitted at the time of submission of the dataset.&lt;br /&gt;
&lt;br /&gt;
'''Ground Truth Specification''': The definition of the required information that accurately describes a particular aspect of the data at a high level where agreement between different human observers can be established, as well as the definition of an appropriate structure (format) for storing this information.&lt;br /&gt;
&lt;br /&gt;
'''Ground Truth Data''': A set of data conforming to a particular ground truth specification and relating to a specific dataset. Ground Truth Data can be submitted at any time, while different Ground Truth Data (corresponding to different aspects of the data) can be associated with the same dataset.&lt;br /&gt;
&lt;br /&gt;
'''Task''': A well defined process to evaluate algorithms in the context of a specific scientific problem. A task would typically provide a specific evaluation protocol, and link to specific resources as required (a dataset, and usually related ground truth data). Tasks should correspond to open challenges in the field. If you undertake any of the tasks defined and you have published results or code available, we would really like to know!&lt;br /&gt;
&lt;br /&gt;
'''Resources''': Any other type of related resources that are not specifically covered by the above definitions. Examples would include software to browse and visualise a dataset, software to create ground truth data, algorithms to do performance evaluation, codecs, reports, publications, etc.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Systems_that_improve_with_use_(2016)&amp;diff=2273</id>
		<title>DAS-Discussion:Systems that improve with use (2016)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Systems_that_improve_with_use_(2016)&amp;diff=2273"/>
		<updated>2016-05-16T09:14:11Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Report by Ido&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Systems that improve with use ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Ido Kissos, Tel Aviv Israel - Improving printed Arabic OCR with ML methods.&lt;br /&gt;
Participants:&lt;br /&gt;
* Marc-Peter Schambach, Siemens, Germany: Handwriting recognition, address recognition.&lt;br /&gt;
* Abdel Belaid, University of Lorraine, France. Administrative docs, info extraction, table detection, entity recognition, OCR and evaluation.&lt;br /&gt;
* Nicolas Ragaut, University of Tours, France. Medical image analysis, doc image analysis. Handwriting and OCR.&lt;br /&gt;
* Brian Davis, Utah, USA. Computer assited transcription (geneological docs).&lt;br /&gt;
* Anh Le, TUAT, Japan - Handwiritng recognition.&lt;br /&gt;
&lt;br /&gt;
=== Problem Definition ===&lt;br /&gt;
* Data centric systems break with usage environment change&lt;br /&gt;
* Goal: Handling changes over time in data centric systems&lt;br /&gt;
** Data changes&lt;br /&gt;
** User need or preferences&lt;br /&gt;
** Training data is not representative&lt;br /&gt;
** Problem domain adaptation&lt;br /&gt;
* Method: Exploit usage-data to improve systems&lt;br /&gt;
** Explicit: explicit data labeling in workflow, dedicated configuration modules&lt;br /&gt;
** Implicit: behavior change, continuous evaluation, negative feedbacks&lt;br /&gt;
&lt;br /&gt;
=== Challenges ===&lt;br /&gt;
# Algorithmic – online training with the new data&lt;br /&gt;
# Architecture – how to close the loop&lt;br /&gt;
# Labeling and data correctness – interpret all kinds of user behavior&lt;br /&gt;
# Privacy – getting feedback is non-trivial&lt;br /&gt;
# Psychologic – we do not want to look at systems that get worse with new data&lt;br /&gt;
# Economic – it may have no business case&lt;br /&gt;
&lt;br /&gt;
=== Solutions - Things to keep in mind ===&lt;br /&gt;
* Prior analysis of possible drifts&lt;br /&gt;
* Implication of drifts: know the robustness of your model to new data, the gains of more training data &lt;br /&gt;
* “One Click” training&lt;br /&gt;
* Online evaluation&lt;br /&gt;
** Measure performance over time&lt;br /&gt;
** Update ground truth&lt;br /&gt;
** Boosting&lt;br /&gt;
** Independent evaluation model - having an adversary classifier to online evaluate performances?&lt;br /&gt;
&lt;br /&gt;
=== Full Version ===&lt;br /&gt;
Until recently data changes would occur in intervals longer than the system’s life-span, as well as they did not mainly rely on vast amount and diverse data. Possible data or user need drifts would be handled by new versions of a software, and that would also be the software companies business model - sell new versions of the same product. Nowadays many software become more data-centric, relying on classifier modules as their core ability. In the “data-era” these changes happen fast, features for classification are sometimes hidden not, hence users do not realize their data drifts until the systems looses its liability. Therefore - the systems should adapt!&lt;br /&gt;
&lt;br /&gt;
Why don’t we develop such systems - because our labs don’t update with real data. Drifts happen within large times. You’re not looking for it - it is expensive, time consuming. Maybe there are psychological issues with this, we don’t want to see this going worst. In academy we are publication-oriented, not production-oriented.&lt;br /&gt;
&lt;br /&gt;
Do we need architecturally to support automatic retrain? If systems get better by adding new data, we must support it. &lt;br /&gt;
Maybe it is difficult to a full automatic system, just semi automatic. Users will not be willing to give a full new labeling as part of their daily use, but one can expect to have partial or implicit labeling. Systems have to learn from its errors so it will be able to correct it in the future, and modelling errors can be a difficult task. A system has to be self confident about its claim - a one class classifier may do it better to predict its false positive errors. &lt;br /&gt;
&lt;br /&gt;
What about online evaluation - evaluating after benchmark end delivery? Building an self-online adversary system that evaluates the main system without ground truth, after delivery time. Sometimes the evaluation can be measured explicitly/implicitly by users activity. The feedback on the data should not necessarily come from the user but the processing chain. &lt;br /&gt;
&lt;br /&gt;
Business problem: if I build a system that adapts to all changes there will be no incentive for the next project. People are paying only once, the economics are against the system. Then you move to subscription models, not purchase. &lt;br /&gt;
&lt;br /&gt;
Big data: from the day of delivery it is not adapted to your problem. No real “ground truth”.&lt;br /&gt;
&lt;br /&gt;
Privacy: we are not allowed to get back the data from the user. governments are strict about it, and it is getting worst. On the other hand big corporations have all data, or a big enough representation of it. Trying to anonymize methods data to allow research on top of them, or ease the regulation to make the data market more fair.&lt;br /&gt;
&lt;br /&gt;
Robustness of the training: what will be the impact of the improvement in the future. Try to make up a score to tell how much more training should we need for how much improvement. What are the gains. Maybe they should be standardized.&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Systems_that_improve_with_use_(2016)&amp;diff=2272</id>
		<title>DAS-Discussion:Systems that improve with use (2016)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Systems_that_improve_with_use_(2016)&amp;diff=2272"/>
		<updated>2016-05-16T09:13:58Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Report by Ido&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Systems that improve with use ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Ido Kissos, Tel Aviv Israel - Improving printed Arabic OCR with ML methods.&lt;br /&gt;
Participants:&lt;br /&gt;
* Marc-Peter Schambach, Siemens, Germany: Handwriting recognition, address recognition.&lt;br /&gt;
* Abdel Belaid, University of Lorraine, France. Administrative docs, info extraction, table detection, entity recognition, OCR and evaluation.&lt;br /&gt;
* Nicolas Ragaut, University of Tours, France. Medical image analysis, doc image analysis. Handwriting and OCR.&lt;br /&gt;
* Brian Davis, Utah, USA. Computer assited transcription (geneological docs).&lt;br /&gt;
* Anh Le, TUAT, Japan - Handwiritng recognition.&lt;br /&gt;
&lt;br /&gt;
=== Problem Definition ===&lt;br /&gt;
* Data centric systems break with usage environment change&lt;br /&gt;
* Goal: Handling changes over time in data centric systems&lt;br /&gt;
** Data changes&lt;br /&gt;
** User need or preferences&lt;br /&gt;
** Training data is not representative&lt;br /&gt;
** Problem domain adaptation&lt;br /&gt;
* Method: Exploit usage-data to improve systems&lt;br /&gt;
** Explicit: explicit data labeling in workflow, dedicated configuration modules&lt;br /&gt;
** Implicit: behavior change, continuous evaluation, negative feedbacks&lt;br /&gt;
&lt;br /&gt;
=== Challenges ===&lt;br /&gt;
# Algorithmic – online training with the new data&lt;br /&gt;
# Architecture – how to close the loop&lt;br /&gt;
# Labeling and data correctness – interpret all kinds of user behavior&lt;br /&gt;
# Privacy – getting feedback is non-trivial&lt;br /&gt;
# Psychologic – we do not want to look at systems that get worse with new data&lt;br /&gt;
# Economic – it may have no business case&lt;br /&gt;
&lt;br /&gt;
=== Solutions - Things to keep in mind ===&lt;br /&gt;
* Prior analysis of possible drifts&lt;br /&gt;
* Implication of drifts: know the robustness of your model to new data, the gains of more training data &lt;br /&gt;
* “One Click” training&lt;br /&gt;
* Online evaluation&lt;br /&gt;
** Measure performance over time&lt;br /&gt;
** Update ground truth&lt;br /&gt;
** Boosting&lt;br /&gt;
** Independent evaluation model - having an adversary classifier to online evaluate performances?&lt;br /&gt;
&lt;br /&gt;
=== Full Version ===&lt;br /&gt;
Until recently data changes would occur in intervals longer than the system’s life-span, as well as they did not mainly rely on vast amount and diverse data. Possible data or user need drifts would be handled by new versions of a software, and that would also be the software companies business model - sell new versions of the same product. Nowadays many software become more data-centric, relying on classifier modules as their core ability. In the “data-era” these changes happen fast, features for classification are sometimes hidden not, hence users do not realize their data drifts until the systems looses its liability. Therefore - the systems should adapt!&lt;br /&gt;
&lt;br /&gt;
Why don’t we develop such systems - because our labs don’t update with real data. Drifts happen within large times. You’re not looking for it - it is expensive, time consuming. Maybe there are psychological issues with this, we don’t want to see this going worst. In academy we are publication-oriented, not production-oriented.&lt;br /&gt;
&lt;br /&gt;
Do we need architecturally to support automatic retrain? If systems get better by adding new data, we must support it. &lt;br /&gt;
Maybe it is difficult to a full automatic system, just semi automatic. Users will not be willing to give a full new labeling as part of their daily use, but one can expect to have partial or implicit labeling. Systems have to learn from its errors so it will be able to correct it in the future, and modelling errors can be a difficult task. A system has to be self confident about its claim - a one class classifier may do it better to predict its false positive errors. &lt;br /&gt;
&lt;br /&gt;
What about online evaluation - evaluating after benchmark end delivery? Building an self-online adversary system that evaluates the main system without ground truth, after delivery time. Sometimes the evaluation can be measured explicitly/implicitly by users activity. The feedback on the data should not necessarily come from the user but the processing chain. &lt;br /&gt;
&lt;br /&gt;
Business problem: if i build a system that adapts to all changes there will be no incentive for the next project. People are paying only once, the economics are against the system. Then you move to subscription models, not purchase. &lt;br /&gt;
&lt;br /&gt;
Big data: from the day of delivery it is not adapted to your problem. No real “ground truth”.&lt;br /&gt;
&lt;br /&gt;
Privacy: we are not allowed to get back the data from the user. governments are strict about it, and it is getting worst. On the other hand big corporations have all data, or a big enough representation of it. Trying to anonymize methods data to allow research on top of them, or ease the regulation to make the data market more fair.&lt;br /&gt;
&lt;br /&gt;
Robustness of the training: what will be the impact of the improvement in the future. Try to make up a score to tell how much more training should we need for how much improvement. What are the gains. Maybe they should be standardized.&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2271</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2271"/>
		<updated>2016-05-16T07:19:05Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;br /&gt;
=== 2014 ===&lt;br /&gt;
* [[DAS-Discussion:DAS 2024]] by Faisal Shafait&lt;br /&gt;
* [[DAS-Discussion:Information Extraction (2014)]] by Nibal Nayef&lt;br /&gt;
=== 2016 ===&lt;br /&gt;
* [[DAS-Discussion:Systems that improve with use (2016)]] by Ido Kissos&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Information_Extraction_(2014)&amp;diff=2054</id>
		<title>DAS-Discussion:Information Extraction (2014)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Information_Extraction_(2014)&amp;diff=2054"/>
		<updated>2015-01-02T13:54:26Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* DAS Working Subgroup Meeting: Information Extraction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Information Extraction ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Nibal Nayef&lt;br /&gt;
Participants:&lt;br /&gt;
* Yoshinori AKAO (Japanese police)&lt;br /&gt;
* Saddok KEBAIRI (Itesoft)&lt;br /&gt;
* Manaba OHTA&lt;br /&gt;
* Xin TAO&lt;br /&gt;
* Ronaldo MESSINA (a2ia)&lt;br /&gt;
* Nibal NAYEF (France)&lt;br /&gt;
* Bao&lt;br /&gt;
&lt;br /&gt;
=== Introduction ===&lt;br /&gt;
We have totally different views of information extraction&lt;br /&gt;
Different tasks:&lt;br /&gt;
* Entity spotting (numbers, words, ….)&lt;br /&gt;
* Graphics spotting (logos, symbols, tables etc.)&lt;br /&gt;
* Semantics after text recognition&lt;br /&gt;
* Logical structure&lt;br /&gt;
&lt;br /&gt;
=== What is a document ??!! ===&lt;br /&gt;
We have many types of documents [and increasing]:&lt;br /&gt;
* Digitally born documents&lt;br /&gt;
* Camera / mobile captured&lt;br /&gt;
* Scanned&lt;br /&gt;
..&lt;br /&gt;
&lt;br /&gt;
To extract any kind of information from any type of document, we need a sort of “prerequisite” module, so that IE modules can work on all document types&lt;br /&gt;
&lt;br /&gt;
=== Problems of IE ===&lt;br /&gt;
* What kind of semantic information should we extract?: Technical terms, ….&lt;br /&gt;
* Define the logical structure of a document&lt;br /&gt;
* Same information in different representations: Same name in different languages&lt;br /&gt;
* What are the ground truth data, size of training data?: Use human voting to build GT&lt;br /&gt;
* Ultimate goal: Automatic and complete understanding of document contents.&lt;br /&gt;
* Application: Enrich Data Mining&lt;br /&gt;
&lt;br /&gt;
=== Approaches ===&lt;br /&gt;
CRF, NLP, and all methods for word/graphic spotting&lt;br /&gt;
&lt;br /&gt;
=== Future Directions ===&lt;br /&gt;
Combine methods from different fields:&lt;br /&gt;
* Image processing&lt;br /&gt;
* Natural language processing&lt;br /&gt;
&lt;br /&gt;
Take into account that documents are drastically changing&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Information_Extraction_(2014)&amp;diff=2053</id>
		<title>DAS-Discussion:Information Extraction (2014)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Information_Extraction_(2014)&amp;diff=2053"/>
		<updated>2015-01-02T13:54:02Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Back to DAS-Discussion:Index {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |  {| |- | {{Last updated}} |}  |}  == DAS Working Subgroup Meeting: Information Extraction == Autho...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Information Extraction ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Nibal Nayef&lt;br /&gt;
Participants:&lt;br /&gt;
* Yoshinori AKAO (Japanese police)&lt;br /&gt;
* Saddok KEBAIRI (Itesoft)&lt;br /&gt;
* Manaba OHTA&lt;br /&gt;
* Xin TAO&lt;br /&gt;
* Ronaldo MESSINA (a2ia)&lt;br /&gt;
* Nibal NAYEF (me !)&lt;br /&gt;
* Bao&lt;br /&gt;
&lt;br /&gt;
=== Introduction ===&lt;br /&gt;
We have totally different views of information extraction&lt;br /&gt;
Different tasks:&lt;br /&gt;
* Entity spotting (numbers, words, ….)&lt;br /&gt;
* Graphics spotting (logos, symbols, tables etc.)&lt;br /&gt;
* Semantics after text recognition&lt;br /&gt;
* Logical structure&lt;br /&gt;
&lt;br /&gt;
=== What is a document ??!! ===&lt;br /&gt;
We have many types of documents [and increasing]:&lt;br /&gt;
* Digitally born documents&lt;br /&gt;
* Camera / mobile captured&lt;br /&gt;
* Scanned&lt;br /&gt;
..&lt;br /&gt;
&lt;br /&gt;
To extract any kind of information from any type of document, we need a sort of “prerequisite” module, so that IE modules can work on all document types&lt;br /&gt;
&lt;br /&gt;
=== Problems of IE ===&lt;br /&gt;
* What kind of semantic information should we extract?: Technical terms, ….&lt;br /&gt;
* Define the logical structure of a document&lt;br /&gt;
* Same information in different representations: Same name in different languages&lt;br /&gt;
* What are the ground truth data, size of training data?: Use human voting to build GT&lt;br /&gt;
* Ultimate goal: Automatic and complete understanding of document contents.&lt;br /&gt;
* Application: Enrich Data Mining&lt;br /&gt;
&lt;br /&gt;
=== Approaches ===&lt;br /&gt;
CRF, NLP, and all methods for word/graphic spotting&lt;br /&gt;
&lt;br /&gt;
=== Future Directions ===&lt;br /&gt;
Combine methods from different fields:&lt;br /&gt;
* Image processing&lt;br /&gt;
* Natural language processing&lt;br /&gt;
&lt;br /&gt;
Take into account that documents are drastically changing&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2052</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2052"/>
		<updated>2015-01-02T13:49:17Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* 2014 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;br /&gt;
=== 2014 ===&lt;br /&gt;
* [[DAS-Discussion:DAS 2024]] by Faisal Shafait&lt;br /&gt;
* [[DAS-Discussion:Information Extraction (2014)]] by Nibal Nayef&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:DAS_2024&amp;diff=2051</id>
		<title>DAS-Discussion:DAS 2024</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:DAS_2024&amp;diff=2051"/>
		<updated>2015-01-02T13:48:20Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: DAS 2024 ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Faisal Shafait&lt;br /&gt;
&lt;br /&gt;
=== Progress in the last 10 years (DAS 2004) ===&lt;br /&gt;
* Open source OCR – gOCR&lt;br /&gt;
* Should we start using cameras?&lt;br /&gt;
* Handwriting recognition – only in the hand of experts&lt;br /&gt;
* Keynote talks&lt;br /&gt;
* Thomas M. Breuel The future of document imaging in the era of electronic documents [http://www.dsi.unifi.it/DAS04/Breuel-DAS04.pdf]&lt;br /&gt;
* Franco Lotti Image quality issues in digitization projects of historical documents [http://www.dsi.unifi.it/DAS04/Lotti-DAS04.pdf]&lt;br /&gt;
&lt;br /&gt;
=== Progress in the next 10 years ===&lt;br /&gt;
* Paper will be used but less and less over time&lt;br /&gt;
* Reusable ink&lt;br /&gt;
* Scanners will disappear&lt;br /&gt;
* Mobile based document recognition will mature to a state that it matches the performance of scanned documents under (almost) all capture conditions&lt;br /&gt;
* Majority of “pre-processing” problems will be solved for historical document processing&lt;br /&gt;
* Will there be a DAS 2024? Yes :)&lt;br /&gt;
* What will be the topics we will be addressing?&lt;br /&gt;
&lt;br /&gt;
=== DAS 2014 keynote talks ===&lt;br /&gt;
* An Inside Look into ABBYY OCR Technology&lt;br /&gt;
* From Academia to Industry, the knowledge transfert in Document Analysis&lt;br /&gt;
* Document Evolution drives Document Analysis&lt;br /&gt;
&lt;br /&gt;
=== The Next Big Thing? ===&lt;br /&gt;
* Understanding “Born-Digital” documents&lt;br /&gt;
* Something “Magical” for knowledge extraction from documents – similar to LSTM for handwriting recognition&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2050</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2050"/>
		<updated>2015-01-02T13:47:52Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* 2014 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;br /&gt;
=== 2014 ===&lt;br /&gt;
* [[DAS-Discussion:DAS 2024]] by Faisal Shafait&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Nibal Nayef&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:DAS_2024&amp;diff=2049</id>
		<title>DAS-Discussion:DAS 2024</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:DAS_2024&amp;diff=2049"/>
		<updated>2015-01-02T13:44:49Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Back to DAS-Discussion:Index {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |  {| |- | {{Last updated}} |}  |}  == DAS Working Subgroup Meeting: Datasets and Benchmarks == Auth...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Datasets and Benchmarks ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Faisal Shafait&lt;br /&gt;
&lt;br /&gt;
=== Progress in the last 10 years (DAS 2004) ===&lt;br /&gt;
* Open source OCR – gOCR&lt;br /&gt;
* Should we start using cameras?&lt;br /&gt;
* Handwriting recognition – only in the hand of experts&lt;br /&gt;
* Keynote talks&lt;br /&gt;
* Thomas M. Breuel The future of document imaging in the era of electronic documents [http://www.dsi.unifi.it/DAS04/Breuel-DAS04.pdf]&lt;br /&gt;
* Franco Lotti Image quality issues in digitization projects of historical documents [http://www.dsi.unifi.it/DAS04/Lotti-DAS04.pdf]&lt;br /&gt;
&lt;br /&gt;
=== Progress in the next 10 years ===&lt;br /&gt;
* Paper will be used but less and less over time&lt;br /&gt;
* Reusable ink&lt;br /&gt;
* Scanners will disappear&lt;br /&gt;
* Mobile based document recognition will mature to a state that it matches the performance of scanned documents under (almost) all capture conditions&lt;br /&gt;
* Majority of “pre-processing” problems will be solved for historical document processing&lt;br /&gt;
* Will there be a DAS 2024? Yes :)&lt;br /&gt;
* What will be the topics we will be addressing?&lt;br /&gt;
&lt;br /&gt;
=== DAS 2014 keynote talks ===&lt;br /&gt;
* An Inside Look into ABBYY OCR Technology&lt;br /&gt;
* From Academia to Industry, the knowledge transfert in Document Analysis&lt;br /&gt;
* Document Evolution drives Document Analysis&lt;br /&gt;
&lt;br /&gt;
=== The Next Big Thing? ===&lt;br /&gt;
* Understanding “Born-Digital” documents&lt;br /&gt;
* Something “Magical” for knowledge extraction from documents – similar to LSTM for handwriting recognition&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2048</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2048"/>
		<updated>2015-01-02T13:38:00Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* 2014 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;br /&gt;
=== 2014 ===&lt;br /&gt;
* [[DAS-Discussion:DAS 2024]] by Faisal Shafait&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2047</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2047"/>
		<updated>2015-01-02T13:37:22Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;br /&gt;
=== 2014 ===&lt;br /&gt;
* [[DAS-Discussion:DAS 2024]]&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2046</id>
		<title>DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2046"/>
		<updated>2015-01-02T13:33:08Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to [[DAS-Discussion:Index]]&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== DAS Working Subgroup Meeting: Datasets and Benchmarks ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Bart Lamiroy (Secretary) – Université de Lorraine&lt;br /&gt;
&lt;br /&gt;
Further Participants:&lt;br /&gt;
* Elisa Barney Smith – Boise State University&lt;br /&gt;
* Abdel Belaïd – Université de Lorraine&lt;br /&gt;
* John Fletcher – Canon&lt;br /&gt;
* Liangcai Gao – Peking University&lt;br /&gt;
* Albert Gordo – CVC Barcelona&lt;br /&gt;
* Masakazu Iwamura – Osaka Prefecture University&lt;br /&gt;
* Dan Lopresti – Lehigh University&lt;br /&gt;
* Tomohsa Matsushita – Tokyo University of Agriculture and Technology&lt;br /&gt;
* Jean-­‐Yves Ramel – Université de Tours&lt;br /&gt;
* Marc-­‐Peter Schambach – Siemens&lt;br /&gt;
* Ray Smith (Moderator) – Google Inc.&lt;br /&gt;
&lt;br /&gt;
=== Context === &lt;br /&gt;
&lt;br /&gt;
The goal of this discussion group is to address the availability, use and dissemination of benchmarks, datasets and ground truth in order to promote subjective and reproducible assessment of document analysis methods, collaboration and exchange of research results in the document analysis domain. The main idea is that “what you measure is what improves”, and that it is difficult to obtain reliable measures expressing the global progress of the state‐of‐the‐art. &lt;br /&gt;
&lt;br /&gt;
=== Topic Discussion History ===&lt;br /&gt;
As a brief reminder of the evolution of this topic as discussed during other DAS editions, we refer the interested reader to the TC‐11 website. In 2010 the main focus of discussion essentially related to making datasets other reference material available to the community and how to provide centralized access to it, how to credit and value contributors and how to maintain a level of control (data curation, availability over time, …) that would insure that the data and algorithms remain usable and useful over an as long as possible period of time. The reported discussions were essentially concerned with feasibility of these concepts, rather than impact, and focused on the TC‐11 initiative of data collection and the DAE platform (http://dae.cse.lehigh.edu). &lt;br /&gt;
&lt;br /&gt;
=== Discussion Topics ===&lt;br /&gt;
&lt;br /&gt;
During the DAS 1012 edition, the following potential discussion topics were identified after a short brainstorming session, ranked by order of (subjectively) perceived importance:&lt;br /&gt;
1. When is a problem stated? Should CFPs be more specific to what topics to address and how they should (could) be measured? How does this relate to hosting competitions? Interaction with whole or end‐to‐end evaluation systems.&lt;br /&gt;
2. What are the fundamental reasons to the perceived difficulties to sharing data sets? (public vs. copyright vs. privacy)&lt;br /&gt;
3. Would it be a good idea to more formally integrate the availability of data sets and reports of benchmarking into the acceptance criteria for publications.&lt;br /&gt;
4. Is there a risk of data sets directing research? Is this good or bad?&lt;br /&gt;
5. Open binaries/open source? &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== When is a Problem Stated?  ===&lt;br /&gt;
This question is considered by the discussion panel members as an essential preliminary step to D. Lopresti and G. Nagy's paper “When is a Problem Solved” in ICDAR 2011, and relates to the initially identified issue concerning the difficulty of measuring the overall contributions of individual research results to the improvement of the global state‐of‐the‐art. Stating a problem is related to measuring some level of achievement, and therefore directly correlated to expressing ground truth. One may conjecture that a problem is stated when there is consensus on the ground truth on the one hand and there is a data set collection of statistically proven significance. Measurement of advancement toward solving a stated problem would then consist of: &lt;br /&gt;
* track record of results over time, &lt;br /&gt;
* defined best practices by the community,&lt;br /&gt;
&lt;br /&gt;
This means that the evolution of the best practices (and the track record of the results) could give a more precise view of the improvement of the agreed upon state‐the‐art. This also means that there is a need of commenting and annotating the reference data sets by the community and also that there may be a need to evaluate individual research results within the scope of broader criteria (e.g. contribution in end‐to‐end application evaluation) The general consensus of the discussion panel is that there might be an interest in experimenting a more formal approach to managing tracks in conferences and acceptance criteria to particular events or publications, by clearly stating (at the time of the CFP) the benchmark to which contributions need to measured. This could consist of: &lt;br /&gt;
* specific problem statements, &lt;br /&gt;
* hosting competitions in direct relation with the track or conference and creating strong incentives for all submissions to compete, &lt;br /&gt;
* ensuring continuity of both data sets, ground truth, and algorithm availability year after year, &lt;br /&gt;
* requiring that reviewers have reasonable access to the data sets and have the means of checking the reported results.&lt;br /&gt;
&lt;br /&gt;
However, it is extremely important to stress that this should never be the sole criteria for acceptance and publication of papers since there is a significant risk of limiting innovating non‐mainstream approaches and the emergence of investigations into new (previously not considered, or considered uninteresting) problems. This is discussed in one of items developed below.&lt;br /&gt;
&lt;br /&gt;
=== Difficulties in Sharing Data Sets ===&lt;br /&gt;
On this issue, the discussions have rather identified a number of open issues, without necessary finding ways of solving them. The issues are: &lt;br /&gt;
* Making data sets available is not a technical issue but a cultural one, not only related to legal issues, but also to the need of acknowledgement by peers and ROI with respect to the effort/cost of creating data sets. &lt;br /&gt;
* Although data sets may be of significant interest, and not be limited by intellectual property or copyright, they may be restricted for publication because of privacy issues. In that case, anonymization processes may not necessarily be possible or appropriate, and are always very costly. Approaches of creating synthetic data sets may yield solutions in some cases. &lt;br /&gt;
* DMCA protection is probably the most convenient framework for academia to reduce the risk of distributing data sets of which the origin cannot be totally guaranteed copyright infringement free. &lt;br /&gt;
* On the other hand, companies are often reluctant to release data, either because of overly concerned legal departments and zero‐risk policies, or because of the significant competitive advantage particular data sets may yield. With respect to the issues mentioned to open and verifiable access to reported results it may be possible to conceive non--‐disclosure bound access to datasets, while still giving reasonable possibilities to verify reported results (e.g. the open access to provenance data – and not the original data – in systems like the DAE platform) &lt;br /&gt;
&lt;br /&gt;
=== Changing Acceptance Criteria ===&lt;br /&gt;
Imposing to confront results to a previously agreed upon benchmark prior to acceptance for publication may prove to be a double edged sword. The discussions have tried to identify pros and cons. &lt;br /&gt;
* As already hinted previously, this would require a shift in the way some events are publicized and organized, since the CFP would necessarily include all the required information (evaluation procedures, data sets, benchmark infrastructure, ...) &lt;br /&gt;
* Imposing stringent benchmarking criteria may not prove a good idea for smaller events, confronted to basic economics and affect the number of submissions and the acceptance rate too strongly. &lt;br /&gt;
* On the other hand, some mature topics should very strongly impose the use of standard benchmarks. &lt;br /&gt;
* This would also require a shift in the review/acceptance process: ◦ In order to preserve the possibility to publish innovative non--‐mainstream new research evaluation should integrate some level of weighting setting a cursor between “correctly benchmarked and conforming to criteria” and “out of scope with respect to criteria, but potentially groundbreaking new topic”, for instance &lt;br /&gt;
  * The possibility to have conditional acceptance and a response phase after review. &lt;br /&gt;
  * Extra load on reviewers, and requirements to be able to correctly verify claimed results. &lt;br /&gt;
* There may be an extra load of reviewers &lt;br /&gt;
* It would be interesting to get the broader community's feeling about this. &lt;br /&gt;
&lt;br /&gt;
=== Data Sets Direct Research ===&lt;br /&gt;
Before data sets become commonly accepted and agreed upon bases for benchmarking, they should undergo some community approval. This raises quite some potentially controversial issues: &lt;br /&gt;
* Data‐driven research evaluation may have a very good impact if the data is good, buy may be harmful is data is bad. &lt;br /&gt;
* Data sets progressively get out of date as knowledge evolves. (one might consider a problem underpinning a data set solved, when the set is considered obsolete) &lt;br /&gt;
* Special interest groups can try to dominate or influence decisions.&lt;br /&gt;
* Some datasets may not be considered of interest in specific cases of 3rd party supported research.&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2045</id>
		<title>DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2045"/>
		<updated>2015-01-02T13:25:01Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== DAS Working Subgroup Meeting: Datasets and Benchmarks ==&lt;br /&gt;
Authors:&lt;br /&gt;
* Bart Lamiroy (Secretary) – Université de Lorraine&lt;br /&gt;
&lt;br /&gt;
Further Participants:&lt;br /&gt;
* Elisa Barney Smith – Boise State University&lt;br /&gt;
* Abdel Belaïd – Université de Lorraine&lt;br /&gt;
* John Fletcher – Canon&lt;br /&gt;
* Liangcai Gao – Peking University&lt;br /&gt;
* Albert Gordo – CVC Barcelona&lt;br /&gt;
* Masakazu Iwamura – Osaka Prefecture University&lt;br /&gt;
* Dan Lopresti – Lehigh University&lt;br /&gt;
* Tomohsa Matsushita – Tokyo University of Agriculture and Technology&lt;br /&gt;
* Jean-­‐Yves Ramel – Université de Tours&lt;br /&gt;
* Marc-­‐Peter Schambach – Siemens&lt;br /&gt;
* Ray Smith (Moderator) – Google Inc.&lt;br /&gt;
&lt;br /&gt;
=== Context === &lt;br /&gt;
&lt;br /&gt;
The goal of this discussion group is to address the availability, use and dissemination of benchmarks, datasets and ground truth in order to promote subjective and reproducible assessment of document analysis methods, collaboration and exchange of research results in the document analysis domain. The main idea is that “what you measure is what improves”, and that it is difficult to obtain reliable measures expressing the global progress of the state‐of‐the‐art. &lt;br /&gt;
&lt;br /&gt;
=== Topic Discussion History ===&lt;br /&gt;
As a brief reminder of the evolution of this topic as discussed during other DAS editions, we refer the interested reader to the TC‐11 website. In 2010 the main focus of discussion essentially related to making datasets other reference material available to the community and how to provide centralized access to it, how to credit and value contributors and how to maintain a level of control (data curation, availability over time, …) that would insure that the data and algorithms remain usable and useful over an as long as possible period of time. The reported discussions were essentially concerned with feasibility of these concepts, rather than impact, and focused on the TC‐11 initiative of data collection and the DAE platform (http://dae.cse.lehigh.edu). &lt;br /&gt;
&lt;br /&gt;
=== Discussion Topics ===&lt;br /&gt;
&lt;br /&gt;
During the DAS 1012 edition, the following potential discussion topics were identified after a short brainstorming session, ranked by order of (subjectively) perceived importance:&lt;br /&gt;
1. When is a problem stated? Should CFPs be more specific to what topics to address and how they should (could) be measured? How does this relate to hosting competitions? Interaction with whole or end‐to‐end evaluation systems.&lt;br /&gt;
2. What are the fundamental reasons to the perceived difficulties to sharing data sets? (public vs. copyright vs. privacy)&lt;br /&gt;
3. Would it be a good idea to more formally integrate the availability of data sets and reports of benchmarking into the acceptance criteria for publications.&lt;br /&gt;
4. Is there a risk of data sets directing research? Is this good or bad?&lt;br /&gt;
5. Open binaries/open source? &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== When is a Problem Stated?  ===&lt;br /&gt;
This question is considered by the discussion panel members as an essential preliminary step to D. Lopresti and G. Nagy's paper “When is a Problem Solved” in ICDAR 2011, and relates to the initially identified issue concerning the difficulty of measuring the overall contributions of individual research results to the improvement of the global state‐of‐the‐art. Stating a problem is related to measuring some level of achievement, and therefore directly correlated to expressing ground truth. One may conjecture that a problem is stated when there is consensus on the ground truth on the one hand and there is a data set collection of statistically proven significance. Measurement of advancement toward solving a stated problem would then consist of: &lt;br /&gt;
* track record of results over time, &lt;br /&gt;
* defined best practices by the community,&lt;br /&gt;
&lt;br /&gt;
This means that the evolution of the best practices (and the track record of the results) could give a more precise view of the improvement of the agreed upon state‐the‐art. This also means that there is a need of commenting and annotating the reference data sets by the community and also that there may be a need to evaluate individual research results within the scope of broader criteria (e.g. contribution in end‐to‐end application evaluation) The general consensus of the discussion panel is that there might be an interest in experimenting a more formal approach to managing tracks in conferences and acceptance criteria to particular events or publications, by clearly stating (at the time of the CFP) the benchmark to which contributions need to measured. This could consist of: &lt;br /&gt;
* specific problem statements, &lt;br /&gt;
* hosting competitions in direct relation with the track or conference and creating strong incentives for all submissions to compete, &lt;br /&gt;
* ensuring continuity of both data sets, ground truth, and algorithm availability year after year, &lt;br /&gt;
* requiring that reviewers have reasonable access to the data sets and have the means of checking the reported results.&lt;br /&gt;
&lt;br /&gt;
However, it is extremely important to stress that this should never be the sole criteria for acceptance and publication of papers since there is a significant risk of limiting innovating non‐mainstream approaches and the emergence of investigations into new (previously not considered, or considered uninteresting) problems. This is discussed in one of items developed below.&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2044</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2044"/>
		<updated>2015-01-02T13:10:16Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* 2012 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Bart Lamiroy&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2043</id>
		<title>DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Datasets,_Benchmarks,_Competition,_and_Continuity_of_Research&amp;diff=2043"/>
		<updated>2015-01-02T13:09:06Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Participants&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Participants&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2042</id>
		<title>DAS-Discussion:Index</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=DAS-Discussion:Index&amp;diff=2042"/>
		<updated>2015-01-02T13:08:37Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* 2012 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; |&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
|}&lt;br /&gt;
== DAS Discussion Groups ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
It is a long-standing tradition to have discussion groups at the DAS workshops.&lt;br /&gt;
These are small working groups discussing about topics of special interest to attendees.&lt;br /&gt;
Till 2008, Henry Baird coordinated the discussions. In 2010, Marcus Liwicki took over the organization.&lt;br /&gt;
&lt;br /&gt;
To make the highlights of the discussions available to the community, we make the summaries of the discussions available on this website. They mainly contain lists of (a) new sound &amp;amp; reliable methods, and (b) urgent open problems.&lt;br /&gt;
&lt;br /&gt;
== Reports of previous discussions ==&lt;br /&gt;
&lt;br /&gt;
=== 2010 ===&lt;br /&gt;
* [[DAS-Discussion:Information Extraction]] by Partha Pratim Roy and Prateek Sarkar&lt;br /&gt;
* [[DAS-Discussion:Camera-Based DIA]] by Masakazu Iwamura and Elisa H. Barney Smith&lt;br /&gt;
* [[DAS-Discussion:Datasets and Benchmarks]] by Dimos Karatzas and Bart Lamiroy&lt;br /&gt;
=== 2012 ===&lt;br /&gt;
* [[DAS-Discussion:Datasets, Benchmarks, Competition, and Continuity of Research]] by Dimos Karatzas and Bart Lamiroy&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1938</id>
		<title>Datasets List</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1938"/>
		<updated>2013-08-24T21:01:45Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Software and Tools */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]]&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
See the datasets [[Datasets per Journal / Conference|sorted according to the Journal / Conference]] they first appeared in.&lt;br /&gt;
&lt;br /&gt;
= Complex Text Containers =&lt;br /&gt;
== Scene Text ==&lt;br /&gt;
* [[MSRA Text Detection 500 Database (MSRA-TD500)]]&lt;br /&gt;
* [[The Street View Text Dataset]]&lt;br /&gt;
* [[The Street View House Numbers (SVHN) Dataset]]&lt;br /&gt;
* [[NEOCR: Natural Environment OCR Dataset]]&lt;br /&gt;
* [[KAIST Scene Text Database]]&lt;br /&gt;
* [[ICDAR 2003 Robust Reading Competitions]]&lt;br /&gt;
* [[ICDAR 2005 Robust Reading Competitions]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
== Born Digital Images ==&lt;br /&gt;
* [[ICDAR 2011 Robust Reading Competition - Challenge 1: &amp;quot;Reading Text in Born-Digital Images (Web and Email)&amp;quot;]]&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Machine-printed Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Table Ground Truth for the UW3 and UNLV datasets]]&lt;br /&gt;
* [[The DocLab Dataset for Evaluating Table Interpretation Methods]]&lt;br /&gt;
* [[http://www.digitisation.eu/data/ the IMPACT data base]] The dataset contains more than half a million representative text-based images compiled by a number of major European libraries. Covering texts from as early as 1500, and containing material from newspapers, books, pamphlets and typewritten notes, the dataset is an invaluable resource for future research into imaging technology, OCR and language enrichment.&lt;br /&gt;
* [http://dataset.primaresearch.org/ PRImA Layout Analysis Dataset]&lt;br /&gt;
* [http://www.dfki.uni-kl.de/~shafait/downloads.html DFKI Dewarping Contest Dataset (CBDAR 2007)] The dataset, that was used in the CBDAR 2007 Dewarping Contest, contains 102 camera captured documents with their corresponding ASCII text ground-truth. Additionally, text-line level ground-truth was also prepared to benchmark curled text-line segmentation algorithms. Part of the dataset (76 out of 102 pages) were also scanned with a flat-bed scanner to create a ground-truth image for image based evaluation of page dewarping algorithms.&lt;br /&gt;
* [http://diuf.unifr.ch/diva/APTI/ APTI: Arabic Printed Text Image Database]&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]] This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010. The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution).&lt;br /&gt;
* [http://ciir.cs.umass.edu/downloads/ocr-evaluation/ RETAS OCR Evaluation Dataset] The RETAS dataset (used in the paper by Yalniz and Manmatha, ICDAR'11) is created to evaluate the optical character recognition (OCR) accuracy of real scanned books. The dataset contains real OCR outputs for 160 scanned books (100 English, 20 French, 20 German, 20 Spanish) downloaded from the Internet Archive website. The corresponding ground truth text for each scanned book is obtained from the Project Gutenberg database. The OCR output of each scanned book is aligned with its ground truth at the word and character level and the alignment output is provided along with estimated OCR accuracies. The dataset is provided for research purposes.&lt;br /&gt;
&lt;br /&gt;
= Graphical Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Chem-Infty Dataset: A ground-truthed dataset of Chemical Structure Images]]&lt;br /&gt;
* [[Braille Dataset - Shiraz University]]&lt;br /&gt;
* [http://www.eurecom.fr/~huet/work.html TradeMarks Image Database] - By way of Benoit Huet, 999 trademark and logo images&lt;br /&gt;
&lt;br /&gt;
= Mixed Content Documents = &lt;br /&gt;
* [http://www.umiacs.umd.edu/~zhugy/Tobacco800.html Tobacco800 Document Image Database] - composed of 1290 document images collected and scanned using a wide variety of equipment over time.&lt;br /&gt;
&lt;br /&gt;
= Handwritten Documents =&lt;br /&gt;
== On-line and Off-line ==&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2009 Signature Verification Competition (SigComp2009)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2010 Signature Verification Competition (4NSigComp2010)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2011 Signature Verification Competition (SigComp2011)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2012 Signature Verification Competition (4NSigComp2012)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.nlpr.ia.ac.cn/databases/handwriting/Home.html CASIA Online and Offline Chinese Handwriting Databases] - The Chinese handwriting datasets were produced by 1,020 writers using Anoto pen on papers, such that both online and offline data were obtained. Both the online and the offline dataset consists of three subsets for isolated characters (DB1.0–1.2, about 3.9 million samples of 7,356 classes) and three for handwritten texts (DB2.0–2.2, about 5,090 pages and 1.35 million characters). The datasets are free for academic research for handwritten document segmentation and retrieval, character and text line recognition, writer adaptation and identification.&lt;br /&gt;
&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]] This dataset contains 15 historical and old manuscript images collected from the historical records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation including bleed-through, faded ink, and blur. The dataset is the first in a series to provide document images and their ground truth as a contribution to Document image analysis and recognition (DAIR) community. It is planned to provide more data and ground-truth information in the fture.&lt;br /&gt;
&lt;br /&gt;
== On-line ==&lt;br /&gt;
* [[CROHME: Competition on Recognition of Online Handwritten Mathematical Expressions]]&lt;br /&gt;
&lt;br /&gt;
* [[Devanagari Character Dataset]]&lt;br /&gt;
&lt;br /&gt;
* [[Harbin Institute of Technology Opening Recognition Corpus for Chinese Characters (HIT-OR3C)]]&lt;br /&gt;
&lt;br /&gt;
* [[IAM Online Document Database (IAMonDo-database)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-on-line-handwriting-database IAM On-Line Handwriting Database]&lt;br /&gt;
&lt;br /&gt;
* [http://hwr.nici.kun.nl/unipen/ UNIPEN database] (Click on link 'CDROMs')&lt;br /&gt;
&lt;br /&gt;
* [http://www.tuat.ac.jp/~nakagawa/database/ Nakagawa Lab Online Handwriting Database]&lt;br /&gt;
&amp;lt;!--** Reference: [http://www.springerlink.com/content/kmhh7dg6h8cgr6a5/ Masaki Nakagawa and Kaoru Matsumoto: &amp;quot;Collection of on-line handwritten Japanese character pattern databases and their analysis,&amp;quot; International Journal on Document Analysis and Recognition, Vol. 7 No. 1, pp.69-81 (2004)].--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* [http://www.ai.rug.nl/~lambert/unipen/icdar-03-competition/ The Informal Competition of Recognizing On-line Words (ICROW)] by the Unipen Foundation&lt;br /&gt;
&lt;br /&gt;
== Off-line ==&lt;br /&gt;
* [http://www.rimes-database.fr/wiki/doku.php The Rimes Database] comprises 12,723 handwritten pages corresponding to 5605 mails of two to three pages. It was collected by asking volunteers to write a letter given one of nine predefined scenarios related to business/customer relations. The dataset has been used in numerous competitions in ICDAR and ICFHR. It is available for research purposes only, through the Web site of the authors.&lt;br /&gt;
&lt;br /&gt;
* [[IBN SINA: A database for research on processing and understanding of Arabic manuscripts images]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.cedar.buffalo.edu/Databases/CDROM1/ CEDAR Off-line Handwriting CDROM1]&lt;br /&gt;
&lt;br /&gt;
* [[CVL-Database]] - An Off-line Database for Writer Retrieval, Writer Identification and  Word Spotting&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-handwriting-database IAM Database] - A full English sentence database for off-line handwriting recognition.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/germana The GERMANA Dataset] - GERMANA is the result of digitising and annotating a 764-page Spanish manuscript entitled “Noticias y documentos relativos a Doña Germana de Foix, ́última Reina de Aragón&amp;quot;, written in 1891 by Vicent Salvador. It contains approximately 21K text lines manually marked and transcribed by palaeography experts.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/rodrigo The RODRIGO Dataset] - RODRIGO is the result of digitising and annotating a manuscript dated 1545. Digitisation was done at 300dpi in color by the Spanish Culture Ministry. The original manuscript is a 853-page bound volume, entitled &amp;quot;Historia de España del arçobispo Don Rodrigo&amp;quot;, completely written in old Castilian (Spanish) by a single author. Annotation exists for text blocks, lines and transcriptions, resulting in approximately 20K lines and 231K running words from a lexicon of 17K words.&lt;br /&gt;
&lt;br /&gt;
* [http://marg.nlm.nih.gov/ MARG- Medical Article Records Groundtruth] - A freely-available repository of document page images and their associated textual and layout data. The data has been reviewed and corrected to establish its &amp;quot;ground truth&amp;quot;. Please contact Dr. George Thoma (thoma@lhc.nlm.nih.gov) at the National Library of Medicine for more information.&lt;br /&gt;
&lt;br /&gt;
* [http://kornai.com/Hindi/ Hindi font samples] by Andras Kornai, June 5 2003&lt;br /&gt;
&lt;br /&gt;
= Software and Tools =&lt;br /&gt;
* [http://www.digitisation.eu/tools/ Tools from the IMPACT Centre of Competence] The tools offered by the Impact Centre of Competence are software components which have been developed by the different technical IMPACT partners during the IMPACT project (2008-2012). Generally, a &amp;quot;tool&amp;quot; is a piece of software which operates on image or text data modifying the data or extracting information from it. Every IMPACT tool has a specific functionality related to OCR, to the the pre- and post-processing stages. They include new approaches in areas such as image enhancement, segmentation, and document structuring, alongside existing and experimental OCR engines. &lt;br /&gt;
* [http://lampsrv02.umiacs.umd.edu/projdb/project.php?id=53 GEDI: Groundtruthing Environment for Document Images] - A generic annotation tool for scanned text documents.&lt;br /&gt;
* [http://www2.parc.com/isl/groups/pda/pixlabeler/index.html PixLabeler] - a research tool for labeling elements in a document image at a pixel level.&lt;br /&gt;
* [http://code.google.com/p/ocropus/ OCRopus(tm)] - The OCRopus(tm) open source document analysis and OCR system&lt;br /&gt;
* [http://htk.eng.cam.ac.uk/ The Hidden Markov Model Toolkit (HTK)] - a portable toolkit for building and manipulating hidden Markov models&lt;br /&gt;
* [https://github.com/meierue/RNNLIB Bidirectional Long-Short Term Memory Networks] - Implementation of Bidirectional Long-Short Term Memory Networks (BLSTM) combined with Connectionist Temporal Classification (CTC) - including examples for Arabic recognition.&lt;br /&gt;
* [http://www.speech.sri.com/projects/srilm/ SRILM - The SRI Language Modeling Toolkit] - SRILM is a toolkit for building and applying statistical language models (LMs), primarily for use in speech recognition, statistical tagging and segmentation, and machine translation.&lt;br /&gt;
* [http://torch5.sourceforge.net/ Torch 5] - a Matlab-like environment for state-of-the-art machine learning algorithms.&lt;br /&gt;
* [http://www.prtools.org/ PRTools] - a Matlab based toolbox for pattern recognition&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- [[Datasets (old page)| Old Page (contains broken links)]] --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1937</id>
		<title>Datasets List</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1937"/>
		<updated>2013-08-24T20:52:21Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Machine-printed Documents */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]]&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
See the datasets [[Datasets per Journal / Conference|sorted according to the Journal / Conference]] they first appeared in.&lt;br /&gt;
&lt;br /&gt;
= Complex Text Containers =&lt;br /&gt;
== Scene Text ==&lt;br /&gt;
* [[MSRA Text Detection 500 Database (MSRA-TD500)]]&lt;br /&gt;
* [[The Street View Text Dataset]]&lt;br /&gt;
* [[The Street View House Numbers (SVHN) Dataset]]&lt;br /&gt;
* [[NEOCR: Natural Environment OCR Dataset]]&lt;br /&gt;
* [[KAIST Scene Text Database]]&lt;br /&gt;
* [[ICDAR 2003 Robust Reading Competitions]]&lt;br /&gt;
* [[ICDAR 2005 Robust Reading Competitions]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
== Born Digital Images ==&lt;br /&gt;
* [[ICDAR 2011 Robust Reading Competition - Challenge 1: &amp;quot;Reading Text in Born-Digital Images (Web and Email)&amp;quot;]]&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Machine-printed Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Table Ground Truth for the UW3 and UNLV datasets]]&lt;br /&gt;
* [[The DocLab Dataset for Evaluating Table Interpretation Methods]]&lt;br /&gt;
* [[http://www.digitisation.eu/data/ the IMPACT data base]] The dataset contains more than half a million representative text-based images compiled by a number of major European libraries. Covering texts from as early as 1500, and containing material from newspapers, books, pamphlets and typewritten notes, the dataset is an invaluable resource for future research into imaging technology, OCR and language enrichment.&lt;br /&gt;
* [http://dataset.primaresearch.org/ PRImA Layout Analysis Dataset]&lt;br /&gt;
* [http://www.dfki.uni-kl.de/~shafait/downloads.html DFKI Dewarping Contest Dataset (CBDAR 2007)] The dataset, that was used in the CBDAR 2007 Dewarping Contest, contains 102 camera captured documents with their corresponding ASCII text ground-truth. Additionally, text-line level ground-truth was also prepared to benchmark curled text-line segmentation algorithms. Part of the dataset (76 out of 102 pages) were also scanned with a flat-bed scanner to create a ground-truth image for image based evaluation of page dewarping algorithms.&lt;br /&gt;
* [http://diuf.unifr.ch/diva/APTI/ APTI: Arabic Printed Text Image Database]&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]] This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010. The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution).&lt;br /&gt;
* [http://ciir.cs.umass.edu/downloads/ocr-evaluation/ RETAS OCR Evaluation Dataset] The RETAS dataset (used in the paper by Yalniz and Manmatha, ICDAR'11) is created to evaluate the optical character recognition (OCR) accuracy of real scanned books. The dataset contains real OCR outputs for 160 scanned books (100 English, 20 French, 20 German, 20 Spanish) downloaded from the Internet Archive website. The corresponding ground truth text for each scanned book is obtained from the Project Gutenberg database. The OCR output of each scanned book is aligned with its ground truth at the word and character level and the alignment output is provided along with estimated OCR accuracies. The dataset is provided for research purposes.&lt;br /&gt;
&lt;br /&gt;
= Graphical Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Chem-Infty Dataset: A ground-truthed dataset of Chemical Structure Images]]&lt;br /&gt;
* [[Braille Dataset - Shiraz University]]&lt;br /&gt;
* [http://www.eurecom.fr/~huet/work.html TradeMarks Image Database] - By way of Benoit Huet, 999 trademark and logo images&lt;br /&gt;
&lt;br /&gt;
= Mixed Content Documents = &lt;br /&gt;
* [http://www.umiacs.umd.edu/~zhugy/Tobacco800.html Tobacco800 Document Image Database] - composed of 1290 document images collected and scanned using a wide variety of equipment over time.&lt;br /&gt;
&lt;br /&gt;
= Handwritten Documents =&lt;br /&gt;
== On-line and Off-line ==&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2009 Signature Verification Competition (SigComp2009)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2010 Signature Verification Competition (4NSigComp2010)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2011 Signature Verification Competition (SigComp2011)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2012 Signature Verification Competition (4NSigComp2012)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.nlpr.ia.ac.cn/databases/handwriting/Home.html CASIA Online and Offline Chinese Handwriting Databases] - The Chinese handwriting datasets were produced by 1,020 writers using Anoto pen on papers, such that both online and offline data were obtained. Both the online and the offline dataset consists of three subsets for isolated characters (DB1.0–1.2, about 3.9 million samples of 7,356 classes) and three for handwritten texts (DB2.0–2.2, about 5,090 pages and 1.35 million characters). The datasets are free for academic research for handwritten document segmentation and retrieval, character and text line recognition, writer adaptation and identification.&lt;br /&gt;
&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]] This dataset contains 15 historical and old manuscript images collected from the historical records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation including bleed-through, faded ink, and blur. The dataset is the first in a series to provide document images and their ground truth as a contribution to Document image analysis and recognition (DAIR) community. It is planned to provide more data and ground-truth information in the fture.&lt;br /&gt;
&lt;br /&gt;
== On-line ==&lt;br /&gt;
* [[CROHME: Competition on Recognition of Online Handwritten Mathematical Expressions]]&lt;br /&gt;
&lt;br /&gt;
* [[Devanagari Character Dataset]]&lt;br /&gt;
&lt;br /&gt;
* [[Harbin Institute of Technology Opening Recognition Corpus for Chinese Characters (HIT-OR3C)]]&lt;br /&gt;
&lt;br /&gt;
* [[IAM Online Document Database (IAMonDo-database)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-on-line-handwriting-database IAM On-Line Handwriting Database]&lt;br /&gt;
&lt;br /&gt;
* [http://hwr.nici.kun.nl/unipen/ UNIPEN database] (Click on link 'CDROMs')&lt;br /&gt;
&lt;br /&gt;
* [http://www.tuat.ac.jp/~nakagawa/database/ Nakagawa Lab Online Handwriting Database]&lt;br /&gt;
&amp;lt;!--** Reference: [http://www.springerlink.com/content/kmhh7dg6h8cgr6a5/ Masaki Nakagawa and Kaoru Matsumoto: &amp;quot;Collection of on-line handwritten Japanese character pattern databases and their analysis,&amp;quot; International Journal on Document Analysis and Recognition, Vol. 7 No. 1, pp.69-81 (2004)].--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* [http://www.ai.rug.nl/~lambert/unipen/icdar-03-competition/ The Informal Competition of Recognizing On-line Words (ICROW)] by the Unipen Foundation&lt;br /&gt;
&lt;br /&gt;
== Off-line ==&lt;br /&gt;
* [http://www.rimes-database.fr/wiki/doku.php The Rimes Database] comprises 12,723 handwritten pages corresponding to 5605 mails of two to three pages. It was collected by asking volunteers to write a letter given one of nine predefined scenarios related to business/customer relations. The dataset has been used in numerous competitions in ICDAR and ICFHR. It is available for research purposes only, through the Web site of the authors.&lt;br /&gt;
&lt;br /&gt;
* [[IBN SINA: A database for research on processing and understanding of Arabic manuscripts images]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.cedar.buffalo.edu/Databases/CDROM1/ CEDAR Off-line Handwriting CDROM1]&lt;br /&gt;
&lt;br /&gt;
* [[CVL-Database]] - An Off-line Database for Writer Retrieval, Writer Identification and  Word Spotting&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-handwriting-database IAM Database] - A full English sentence database for off-line handwriting recognition.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/germana The GERMANA Dataset] - GERMANA is the result of digitising and annotating a 764-page Spanish manuscript entitled “Noticias y documentos relativos a Doña Germana de Foix, ́última Reina de Aragón&amp;quot;, written in 1891 by Vicent Salvador. It contains approximately 21K text lines manually marked and transcribed by palaeography experts.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/rodrigo The RODRIGO Dataset] - RODRIGO is the result of digitising and annotating a manuscript dated 1545. Digitisation was done at 300dpi in color by the Spanish Culture Ministry. The original manuscript is a 853-page bound volume, entitled &amp;quot;Historia de España del arçobispo Don Rodrigo&amp;quot;, completely written in old Castilian (Spanish) by a single author. Annotation exists for text blocks, lines and transcriptions, resulting in approximately 20K lines and 231K running words from a lexicon of 17K words.&lt;br /&gt;
&lt;br /&gt;
* [http://marg.nlm.nih.gov/ MARG- Medical Article Records Groundtruth] - A freely-available repository of document page images and their associated textual and layout data. The data has been reviewed and corrected to establish its &amp;quot;ground truth&amp;quot;. Please contact Dr. George Thoma (thoma@lhc.nlm.nih.gov) at the National Library of Medicine for more information.&lt;br /&gt;
&lt;br /&gt;
* [http://kornai.com/Hindi/ Hindi font samples] by Andras Kornai, June 5 2003&lt;br /&gt;
&lt;br /&gt;
= Software and Tools =&lt;br /&gt;
* [http://lampsrv02.umiacs.umd.edu/projdb/project.php?id=53 GEDI: Groundtruthing Environment for Document Images] - A generic annotation tool for scanned text documents.&lt;br /&gt;
* [http://www2.parc.com/isl/groups/pda/pixlabeler/index.html PixLabeler] - a research tool for labeling elements in a document image at a pixel level.&lt;br /&gt;
* [http://code.google.com/p/ocropus/ OCRopus(tm)] - The OCRopus(tm) open source document analysis and OCR system&lt;br /&gt;
* [http://htk.eng.cam.ac.uk/ The Hidden Markov Model Toolkit (HTK)] - a portable toolkit for building and manipulating hidden Markov models&lt;br /&gt;
* [https://github.com/meierue/RNNLIB Bidirectional Long-Short Term Memory Networks] - Implementation of Bidirectional Long-Short Term Memory Networks (BLSTM) combined with Connectionist Temporal Classification (CTC) - including examples for Arabic recognition.&lt;br /&gt;
* [http://www.speech.sri.com/projects/srilm/ SRILM - The SRI Language Modeling Toolkit] - SRILM is a toolkit for building and applying statistical language models (LMs), primarily for use in speech recognition, statistical tagging and segmentation, and machine translation.&lt;br /&gt;
* [http://torch5.sourceforge.net/ Torch 5] - a Matlab-like environment for state-of-the-art machine learning algorithms.&lt;br /&gt;
* [http://www.prtools.org/ PRTools] - a Matlab based toolbox for pattern recognition&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- [[Datasets (old page)| Old Page (contains broken links)]] --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Writer_Identification_and_Word_Spotting_for_the_CVL_Database&amp;diff=1936</id>
		<title>Writer Identification and Word Spotting for the CVL Database</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Writer_Identification_and_Word_Spotting_for_the_CVL_Database&amp;diff=1936"/>
		<updated>2013-08-24T12:01:26Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Description=  T…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
The CVL-database has 311 writers and was designed for writer retrieval and identification. The database consists of 7 (27 writers) respectively 5 (284 writers) different texts (101069 words at all).&lt;br /&gt;
Additionally each page is labeled and provides the coordinates of the bounding boxes of each word (punctuations are not annotated) encoded using an XML-file. Thus, the CVL database can also be used for the evaluation of word-spotting methods. In contrast to the IAM database the number of pages of each writer is distributed more equally.&lt;br /&gt;
&lt;br /&gt;
=Evaluation Protocol=&lt;br /&gt;
&lt;br /&gt;
Evaluation for Writer Identification: It is suggested to use the evaluation metrics from the Writer Identification Contest (ICDAR)&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[CVL-Database]]&lt;br /&gt;
&lt;br /&gt;
=Related Ground Truth Data=&lt;br /&gt;
* [[Bounding Boxes, IDs, and Transcription for the CVL Database]]&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
Markus Diem, Stefan Fiel, Florian Kleber and Robert Sablatnig, CVL-Database: An Off-line Database for Writer Retrieval, Writer Identification and Word Spotting, In Proc. of the 12th Int. Conference on Document Analysis and Recognition (ICDAR) 2013, forthcoming. &lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Bounding_Boxes,_IDs,_and_Transcription_for_the_CVL_Database&amp;diff=1935</id>
		<title>Bounding Boxes, IDs, and Transcription for the CVL Database</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Bounding_Boxes,_IDs,_and_Transcription_for_the_CVL_Database&amp;diff=1935"/>
		<updated>2013-08-24T11:59:43Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords=  =Des…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
&lt;br /&gt;
The GT-Viewer allows to view the ground truth. All words of the text are surrounded by a bounding &lt;br /&gt;
&lt;br /&gt;
box (punctuation is not considered), which has been automatically calculated and manually checked &lt;br /&gt;
&lt;br /&gt;
by two individuals. The information is stored in a XML file. For a description on the XML file &lt;br /&gt;
&lt;br /&gt;
please take a look at the referenced paper. A XML-Parser (C++) is available to read the GT data.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[CVL-Database]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Writer Identification and Word Spotting for the CVL Database]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
The ground truth is in the files of the database. Refer to [[CVL-Database]]&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/CVL/GtViewer.zip GT-Viewer] (7.75 MB)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=CVL-Database&amp;diff=1934</id>
		<title>CVL-Database</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=CVL-Database&amp;diff=1934"/>
		<updated>2013-08-24T11:55:37Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-07-22 |- | {{Last updated}} |}  |}  CVL-Database - A…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-07-22&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
CVL-Database - An Off-line Database for Writer Retrieval, Writer Identification and  Word &lt;br /&gt;
&lt;br /&gt;
Spotting&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Markus Diem&lt;br /&gt;
 Stefan Fiel&lt;br /&gt;
 Florian Kleber&lt;br /&gt;
 Robert Sablatnig&lt;br /&gt;
 sab@caa.tuwien.ac.at&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
CVL Database is licensed under a Creative Commons Attribution-NonCommercial 3.0 Unported License &lt;br /&gt;
&lt;br /&gt;
[http://creativecommons.org/licenses/by-nc/3.0/]&lt;br /&gt;
&lt;br /&gt;
This database may be used for non-commercial research purpose only. If you publish material based &lt;br /&gt;
&lt;br /&gt;
on this database, we request you to include a reference to the publication listed below.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Writer Identification, Word Spotting, Cursive Handwriting&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The CVL Database is a public database for writer retrieval, writer identification and word &lt;br /&gt;
&lt;br /&gt;
spotting. The database consists of 7 different handwritten texts (1 German and 6 Englisch Texts) &lt;br /&gt;
&lt;br /&gt;
and 309 different writers. For each text a rgb color image (300 dpi) comprising the handwritten &lt;br /&gt;
&lt;br /&gt;
text and the printed text sample is available as well as a cropped version (only handwritten). An &lt;br /&gt;
&lt;br /&gt;
unique id identifies the writer, whereas the Bounding Boxes for each single word are stored in an &lt;br /&gt;
&lt;br /&gt;
XML file.&lt;br /&gt;
&lt;br /&gt;
The CVL-database consists of images with cursively handwritten german and english texts which has &lt;br /&gt;
&lt;br /&gt;
been choosen from literary works. All pages have an unique writer id and the text number &lt;br /&gt;
&lt;br /&gt;
(separated by a dash) at the upper right corner, followed by the printed sample text. The text is &lt;br /&gt;
&lt;br /&gt;
placed between two horizontal separatores. Beneath the printed text individuals have been asked &lt;br /&gt;
&lt;br /&gt;
to write the text using a ruled undersheet to prevent curled text lines. The layout follows the &lt;br /&gt;
&lt;br /&gt;
style of the [http://www.iam.unibe.ch/fki/databases/iam-handwriting-database|IAM database]. &lt;br /&gt;
&lt;br /&gt;
Samples of the following texts have been used: &lt;br /&gt;
* Edwin A. Abbot - Flatland: A Romance of Many Dimension (92 words).&lt;br /&gt;
* William Shakespeare - Mac Beth (49 words).&lt;br /&gt;
* Wikipedia - Mailüfterl (73 words, under CC Attribution-ShareALike License).&lt;br /&gt;
* Charles Darwin - Origin of Species (52 words).&lt;br /&gt;
* Johann Wolfgang von Goethe - Faust. Eine Tragödie (50 words).&lt;br /&gt;
* Oscar Wilde - The Picture of Dorian Gray (66 words).&lt;br /&gt;
* Edgar Allan Poe - The Fall of the House of Usher (78 words).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Metadata and Technical Details=&lt;br /&gt;
All pages have a unique writer id and the text number (separated by a dash) at the upper right &lt;br /&gt;
&lt;br /&gt;
corner, followed by the printed sample text. The text is placed between two horizontal &lt;br /&gt;
&lt;br /&gt;
separators. The files are named according the unique writer id and the text number. In addition, &lt;br /&gt;
&lt;br /&gt;
text lines and words are extracted. Their filename convention is the same with the text line &lt;br /&gt;
&lt;br /&gt;
number and word number respectively added at the end. For word images, the GT entry is the last &lt;br /&gt;
&lt;br /&gt;
part of the filename. The Bounding Boxes for each single word are stored in an XML file according &lt;br /&gt;
&lt;br /&gt;
the unique id.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Bounding Boxes, IDs, and Transcription for the CVL Database]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Writer Identification and Word Spotting for the CVL Database]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
Markus Diem, Stefan Fiel, Florian Kleber and Robert Sablatnig, CVL-Database: An Off-line Database &lt;br /&gt;
&lt;br /&gt;
for Writer Retrieval, Writer Identification and Word Spotting, In Proc. of the 12th Int. &lt;br /&gt;
&lt;br /&gt;
Conference on Document Analysis and Recognition (ICDAR) 2013, forthcoming. &lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
Please refer to [http://caa.tuwien.ac.at/cvl/research/cvl-database/index.html &lt;br /&gt;
&lt;br /&gt;
http://caa.tuwien.ac.at/cvl/research/cvl-database/index.html] for downloading the files from the &lt;br /&gt;
&lt;br /&gt;
origninal datasets site.&lt;br /&gt;
&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/CVL/cvl-database.zip Original images] (3.92 GB)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/CVL/cvl-database-cropped.zip Cropped images] (1.12 GB)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1933</id>
		<title>Datasets List</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1933"/>
		<updated>2013-08-24T11:50:13Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Off-line */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]]&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
See the datasets [[Datasets per Journal / Conference|sorted according to the Journal / Conference]] they first appeared in.&lt;br /&gt;
&lt;br /&gt;
= Complex Text Containers =&lt;br /&gt;
== Scene Text ==&lt;br /&gt;
* [[MSRA Text Detection 500 Database (MSRA-TD500)]]&lt;br /&gt;
* [[The Street View Text Dataset]]&lt;br /&gt;
* [[The Street View House Numbers (SVHN) Dataset]]&lt;br /&gt;
* [[NEOCR: Natural Environment OCR Dataset]]&lt;br /&gt;
* [[KAIST Scene Text Database]]&lt;br /&gt;
* [[ICDAR 2003 Robust Reading Competitions]]&lt;br /&gt;
* [[ICDAR 2005 Robust Reading Competitions]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
== Born Digital Images ==&lt;br /&gt;
* [[ICDAR 2011 Robust Reading Competition - Challenge 1: &amp;quot;Reading Text in Born-Digital Images (Web and Email)&amp;quot;]]&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Machine-printed Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Table Ground Truth for the UW3 and UNLV datasets]]&lt;br /&gt;
* [[The DocLab Dataset for Evaluating Table Interpretation Methods]]&lt;br /&gt;
* [http://dataset.primaresearch.org/ PRImA Layout Analysis Dataset]&lt;br /&gt;
* [http://www.dfki.uni-kl.de/~shafait/downloads.html DFKI Dewarping Contest Dataset (CBDAR 2007)] The dataset, that was used in the CBDAR 2007 Dewarping Contest, contains 102 camera captured documents with their corresponding ASCII text ground-truth. Additionally, text-line level ground-truth was also prepared to benchmark curled text-line segmentation algorithms. Part of the dataset (76 out of 102 pages) were also scanned with a flat-bed scanner to create a ground-truth image for image based evaluation of page dewarping algorithms.&lt;br /&gt;
* [http://diuf.unifr.ch/diva/APTI/ APTI: Arabic Printed Text Image Database]&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]] This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010. The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution).&lt;br /&gt;
* [http://ciir.cs.umass.edu/downloads/ocr-evaluation/ RETAS OCR Evaluation Dataset] The RETAS dataset (used in the paper by Yalniz and Manmatha, ICDAR'11) is created to evaluate the optical character recognition (OCR) accuracy of real scanned books. The dataset contains real OCR outputs for 160 scanned books (100 English, 20 French, 20 German, 20 Spanish) downloaded from the Internet Archive website. The corresponding ground truth text for each scanned book is obtained from the Project Gutenberg database. The OCR output of each scanned book is aligned with its ground truth at the word and character level and the alignment output is provided along with estimated OCR accuracies. The dataset is provided for research purposes.&lt;br /&gt;
&lt;br /&gt;
= Graphical Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Chem-Infty Dataset: A ground-truthed dataset of Chemical Structure Images]]&lt;br /&gt;
* [[Braille Dataset - Shiraz University]]&lt;br /&gt;
* [http://www.eurecom.fr/~huet/work.html TradeMarks Image Database] - By way of Benoit Huet, 999 trademark and logo images&lt;br /&gt;
&lt;br /&gt;
= Mixed Content Documents = &lt;br /&gt;
* [http://www.umiacs.umd.edu/~zhugy/Tobacco800.html Tobacco800 Document Image Database] - composed of 1290 document images collected and scanned using a wide variety of equipment over time.&lt;br /&gt;
&lt;br /&gt;
= Handwritten Documents =&lt;br /&gt;
== On-line and Off-line ==&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2009 Signature Verification Competition (SigComp2009)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2010 Signature Verification Competition (4NSigComp2010)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2011 Signature Verification Competition (SigComp2011)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2012 Signature Verification Competition (4NSigComp2012)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.nlpr.ia.ac.cn/databases/handwriting/Home.html CASIA Online and Offline Chinese Handwriting Databases] - The Chinese handwriting datasets were produced by 1,020 writers using Anoto pen on papers, such that both online and offline data were obtained. Both the online and the offline dataset consists of three subsets for isolated characters (DB1.0–1.2, about 3.9 million samples of 7,356 classes) and three for handwritten texts (DB2.0–2.2, about 5,090 pages and 1.35 million characters). The datasets are free for academic research for handwritten document segmentation and retrieval, character and text line recognition, writer adaptation and identification.&lt;br /&gt;
&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]] This dataset contains 15 historical and old manuscript images collected from the historical records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation including bleed-through, faded ink, and blur. The dataset is the first in a series to provide document images and their ground truth as a contribution to Document image analysis and recognition (DAIR) community. It is planned to provide more data and ground-truth information in the fture.&lt;br /&gt;
&lt;br /&gt;
== On-line ==&lt;br /&gt;
* [[CROHME: Competition on Recognition of Online Handwritten Mathematical Expressions]]&lt;br /&gt;
&lt;br /&gt;
* [[Devanagari Character Dataset]]&lt;br /&gt;
&lt;br /&gt;
* [[Harbin Institute of Technology Opening Recognition Corpus for Chinese Characters (HIT-OR3C)]]&lt;br /&gt;
&lt;br /&gt;
* [[IAM Online Document Database (IAMonDo-database)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-on-line-handwriting-database IAM On-Line Handwriting Database]&lt;br /&gt;
&lt;br /&gt;
* [http://hwr.nici.kun.nl/unipen/ UNIPEN database] (Click on link 'CDROMs')&lt;br /&gt;
&lt;br /&gt;
* [http://www.tuat.ac.jp/~nakagawa/database/ Nakagawa Lab Online Handwriting Database]&lt;br /&gt;
&amp;lt;!--** Reference: [http://www.springerlink.com/content/kmhh7dg6h8cgr6a5/ Masaki Nakagawa and Kaoru Matsumoto: &amp;quot;Collection of on-line handwritten Japanese character pattern databases and their analysis,&amp;quot; International Journal on Document Analysis and Recognition, Vol. 7 No. 1, pp.69-81 (2004)].--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* [http://www.ai.rug.nl/~lambert/unipen/icdar-03-competition/ The Informal Competition of Recognizing On-line Words (ICROW)] by the Unipen Foundation&lt;br /&gt;
&lt;br /&gt;
== Off-line ==&lt;br /&gt;
* [http://www.rimes-database.fr/wiki/doku.php The Rimes Database] comprises 12,723 handwritten pages corresponding to 5605 mails of two to three pages. It was collected by asking volunteers to write a letter given one of nine predefined scenarios related to business/customer relations. The dataset has been used in numerous competitions in ICDAR and ICFHR. It is available for research purposes only, through the Web site of the authors.&lt;br /&gt;
&lt;br /&gt;
* [[IBN SINA: A database for research on processing and understanding of Arabic manuscripts images]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.cedar.buffalo.edu/Databases/CDROM1/ CEDAR Off-line Handwriting CDROM1]&lt;br /&gt;
&lt;br /&gt;
* [[CVL-Database]] - An Off-line Database for Writer Retrieval, Writer Identification and  Word Spotting&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-handwriting-database IAM Database] - A full English sentence database for off-line handwriting recognition.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/germana The GERMANA Dataset] - GERMANA is the result of digitising and annotating a 764-page Spanish manuscript entitled “Noticias y documentos relativos a Doña Germana de Foix, ́última Reina de Aragón&amp;quot;, written in 1891 by Vicent Salvador. It contains approximately 21K text lines manually marked and transcribed by palaeography experts.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/rodrigo The RODRIGO Dataset] - RODRIGO is the result of digitising and annotating a manuscript dated 1545. Digitisation was done at 300dpi in color by the Spanish Culture Ministry. The original manuscript is a 853-page bound volume, entitled &amp;quot;Historia de España del arçobispo Don Rodrigo&amp;quot;, completely written in old Castilian (Spanish) by a single author. Annotation exists for text blocks, lines and transcriptions, resulting in approximately 20K lines and 231K running words from a lexicon of 17K words.&lt;br /&gt;
&lt;br /&gt;
* [http://marg.nlm.nih.gov/ MARG- Medical Article Records Groundtruth] - A freely-available repository of document page images and their associated textual and layout data. The data has been reviewed and corrected to establish its &amp;quot;ground truth&amp;quot;. Please contact Dr. George Thoma (thoma@lhc.nlm.nih.gov) at the National Library of Medicine for more information.&lt;br /&gt;
&lt;br /&gt;
* [http://kornai.com/Hindi/ Hindi font samples] by Andras Kornai, June 5 2003&lt;br /&gt;
&lt;br /&gt;
= Software and Tools =&lt;br /&gt;
* [http://lampsrv02.umiacs.umd.edu/projdb/project.php?id=53 GEDI: Groundtruthing Environment for Document Images] - A generic annotation tool for scanned text documents.&lt;br /&gt;
* [http://www2.parc.com/isl/groups/pda/pixlabeler/index.html PixLabeler] - a research tool for labeling elements in a document image at a pixel level.&lt;br /&gt;
* [http://code.google.com/p/ocropus/ OCRopus(tm)] - The OCRopus(tm) open source document analysis and OCR system&lt;br /&gt;
* [http://htk.eng.cam.ac.uk/ The Hidden Markov Model Toolkit (HTK)] - a portable toolkit for building and manipulating hidden Markov models&lt;br /&gt;
* [https://github.com/meierue/RNNLIB Bidirectional Long-Short Term Memory Networks] - Implementation of Bidirectional Long-Short Term Memory Networks (BLSTM) combined with Connectionist Temporal Classification (CTC) - including examples for Arabic recognition.&lt;br /&gt;
* [http://www.speech.sri.com/projects/srilm/ SRILM - The SRI Language Modeling Toolkit] - SRILM is a toolkit for building and applying statistical language models (LMs), primarily for use in speech recognition, statistical tagging and segmentation, and machine translation.&lt;br /&gt;
* [http://torch5.sourceforge.net/ Torch 5] - a Matlab-like environment for state-of-the-art machine learning algorithms.&lt;br /&gt;
* [http://www.prtools.org/ PRTools] - a Matlab based toolbox for pattern recognition&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- [[Datasets (old page)| Old Page (contains broken links)]] --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1913</id>
		<title>Ground Truth for LRDE DBD binarization</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1913"/>
		<updated>2013-07-03T16:28:11Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Version 1.0 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, binarization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
125 binarized images for &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Image groundtruths have been produced using a semi-automatic process: a global thresholding followed by some manual adjustments. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_bin_gt-1.0.zip Binarization groundtruth] (0 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_OCR&amp;diff=1912</id>
		<title>Ground Truth for LRDE DBD OCR</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_OCR&amp;diff=1912"/>
		<updated>2013-07-03T16:27:49Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords= scann…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, binarization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
125 binarized images for &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Image groundtruths have been produced using a semi-automatic process: a global thresholding followed by some manual adjustments. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_bin_gt-1.0.zip Binarization groundtruth] (0 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=OCR_Evaluation_for_LRDE_DBD&amp;diff=1911</id>
		<title>OCR Evaluation for LRDE DBD</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=OCR_Evaluation_for_LRDE_DBD&amp;diff=1911"/>
		<updated>2013-07-03T16:24:54Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords= scann…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, OCR&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
OCR evaluation: Lines are extracted from the binarization outputs and OCR (Tesseract) is run in order to compare to OCR ground-truth. It is performed from binarization of “clean”, “scanned” and “original”  documents.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
Lines for OCR evaluation are also grouped by size: small, medium and large. (0 &amp;lt; small &amp;lt;= 30 &amp;lt; medium &amp;lt;= 55 &amp;lt; large &amp;lt; +inf). It shows how robust is a binarization algorithm to objects with different sizes in a single document.&lt;br /&gt;
&lt;br /&gt;
=Evaluation Protocol=&lt;br /&gt;
&lt;br /&gt;
Tools are provided to read and process all the data.&lt;br /&gt;
 &lt;br /&gt;
A setup script is provided to download and configure the benchmarking environment.&lt;br /&gt;
&lt;br /&gt;
A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
&lt;br /&gt;
C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
&lt;br /&gt;
	6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
A setup script is available to download and setup the benchmark system. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/lrde-dbd-tools-1.0.zip Tools for processing] (0.08 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1910</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1910"/>
		<updated>2013-07-03T16:22:10Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
* [[OCR Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
Please refer to [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD] for downloading the files from the origninal datasets site.&lt;br /&gt;
&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_orig-1.0.zip Original images] (213 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_clean-1.0.zip Clean Documents images] (67 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_scanned-1.0.zip Scanned Documents] (583 Mb)&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Document_Binarization_Evaluation_for_LRDE_DBD&amp;diff=1909</id>
		<title>Document Binarization Evaluation for LRDE DBD</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Document_Binarization_Evaluation_for_LRDE_DBD&amp;diff=1909"/>
		<updated>2013-07-03T16:21:45Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords= scann…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, binarization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
Pixel-wise evaluation: binarization outputs are compared to binarization ground-truth, pixel-wise. It is performed from binarization of “clean documents”.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Evaluation Protocol=&lt;br /&gt;
Binarization algorithms should handles PNG format as input and output images.&lt;br /&gt;
&lt;br /&gt;
Binarization outputs must set  background to False and objects to True.&lt;br /&gt;
&lt;br /&gt;
Tools are provided to read and process all the data.&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/lrde-dbd-tools-1.0.zip Tools for processing] (0.08 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1908</id>
		<title>Ground Truth for LRDE DBD binarization</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1908"/>
		<updated>2013-07-03T16:15:58Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, binarization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
125 binarized images for &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Image groundtruths have been produced using a semi-automatic process: a global thresholding followed by some manual adjustments. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_bin_gt-1.0.zip Binarization groundtruth] (21 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1907</id>
		<title>Ground Truth for LRDE DBD binarization</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_binarization&amp;diff=1907"/>
		<updated>2013-07-03T16:15:10Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords= scann…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, binarization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
125 binarized images for &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Image groundtruths have been produced using a semi-automatic process: a global thresholding followed by some manual adjustments. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* none&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_bin_gt-1.0.zip Binarization groundtruth] (21 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_text_line_localization&amp;diff=1906</id>
		<title>Ground Truth for LRDE DBD text line localization</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_text_line_localization&amp;diff=1906"/>
		<updated>2013-07-03T16:08:20Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, text line localization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
Text Lines Localization Information has been made available by applying text line localization algorithms. The size category of the text depends on the x-height and is considered with the following rule: 0 &amp;lt; small &amp;lt;= 30 &amp;lt; medium &amp;lt;= 55 &amp;lt; large &amp;lt; +inf&lt;br /&gt;
&lt;br /&gt;
* 123 large text lines localization (clean)&lt;br /&gt;
* 320 medium text lines localization (clean).&lt;br /&gt;
* 9551 small text lines localization (clean).&lt;br /&gt;
* 123 large text lines localization (original).&lt;br /&gt;
* 320 medium text lines localization (original).&lt;br /&gt;
* 9551 small text lines localization (original).&lt;br /&gt;
* 123 large text lines localization (scanned).&lt;br /&gt;
* 320 medium text lines localization (scanned).&lt;br /&gt;
* 9551 small text lines localization (scanned).&lt;br /&gt;
&lt;br /&gt;
The text lines dataset covers only a subset of the full-document dataset. It is generated from the binarization of the full-document images.&lt;br /&gt;
Text line localizations are stored as bounding box coordinates in text files. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* none&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_textlines-1.0.zip Text lines localization] (9.8 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1905</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1905"/>
		<updated>2013-07-03T16:04:03Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Version 1.0 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
Please refer to [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD] for downloading the files from the origninal datasets site.&lt;br /&gt;
&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_orig-1.0.zip Original images] (213 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_clean-1.0.zip Clean Documents images] (67 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_scanned-1.0.zip Scanned Documents] (583 Mb)&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1904</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1904"/>
		<updated>2013-07-03T16:03:34Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
Please refer to [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD] for downloading the files from the origninal datasets site.&lt;br /&gt;
&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_orig-1.0.zip Original images] (213 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_clean-1.0.zip Clean Documents images] (67 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/LRDE/nouvel_obs_2402_scanned-1.0.zip Scanned Documents] (583 Mb)&lt;br /&gt;
 &lt;br /&gt;
nouvel_obs_2402_clean-1.0.zip&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Binarization_of_PHIBD_2012_dataset&amp;diff=1903</id>
		<title>Binarization of PHIBD 2012 dataset</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Binarization_of_PHIBD_2012_dataset&amp;diff=1903"/>
		<updated>2013-07-03T15:50:16Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Version 1.0 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
Binarization of handwritten Document Images.&lt;br /&gt;
&lt;br /&gt;
There are actually two tasks, depending on the nature of the binarization method used.&lt;br /&gt;
&lt;br /&gt;
#	For regular binarization methods, the task is to binarize all 15 document images.&lt;br /&gt;
#	For learning-based binarization methods, the task is to use images number 1 to 5 for training, and then binarize images number 6 to 16.&lt;br /&gt;
&lt;br /&gt;
A few baseline methods have been provided: PC (phase congruency) binarization method [Ziaei2012], and SGL/BGL binarization method [Farrahi2009, Farrahi2010]. The SGL/BGL method uses a rough binarization as its initialization.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Evaluation Protocol=&lt;br /&gt;
&lt;br /&gt;
#	For regular methods, the average F-measure of the binarized images against the provided ground truth is used as the performance of the binarization method in question.&lt;br /&gt;
#	For learning-based methods, the average F-measure of the binarized images number 6 to 15 against the provided ground truth is used as the performance of the binarization method in question.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; cellpadding=&amp;quot;5&amp;quot; align=&amp;quot;center&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Method&lt;br /&gt;
!Whole set&lt;br /&gt;
!Training set&lt;br /&gt;
!Test set&lt;br /&gt;
|- &lt;br /&gt;
|Otsu (regular)&lt;br /&gt;
|82.09&lt;br /&gt;
|90.76&lt;br /&gt;
|77.75&lt;br /&gt;
|-&lt;br /&gt;
|PC (regular)&lt;br /&gt;
|90.91&lt;br /&gt;
|92.33&lt;br /&gt;
|90.20&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (upper bound; rough bin: GT)&lt;br /&gt;
|94.55&lt;br /&gt;
|95.37&lt;br /&gt;
|94.14&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (upper bound; rough bin: PC)&lt;br /&gt;
|91.79&lt;br /&gt;
|93.29&lt;br /&gt;
|91.04&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (learning) (rough bin: PC)&lt;br /&gt;
|N/A&lt;br /&gt;
|N/A&lt;br /&gt;
|89.94&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and background gray level (BGL) is provided. The executable of the method will be provided in near future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]]&lt;br /&gt;
&lt;br /&gt;
=Related Ground Truth Data=&lt;br /&gt;
* [[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian historical document dataset with introduction to PhaseGT: A ground truthing application, to be submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: 10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework for the optimization and automation of document binarization methods, Computer Vision and Image Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Training.txt  Train Set Meta Data] (0 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Test.txt   Test Set Meta Data] (0 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Otsu_PHIBD_2012.zip  Otsu_PHIBD_2012 baseline method] (0.39 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/PC_PHIBD_2012.zip PC_PHIBD_2012  baseline method] (0.33 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PHIBD_2012.zip SGLBGL_PHIBD_2012  baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PC_PHIBD_2012.zip   SGLBGL_PC_PHIBD_2012 baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PC_TrainTest_PHIBD_2012.zip   SGLBGL_PC_TrainTest_PHIBD_2012 baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_metacode.m  SGLBGL_metacode.m Sample Program] (0 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Binarization_of_PHIBD_2012_dataset&amp;diff=1902</id>
		<title>Binarization of PHIBD 2012 dataset</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Binarization_of_PHIBD_2012_dataset&amp;diff=1902"/>
		<updated>2013-07-03T15:48:52Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Description= Bi…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
Binarization of handwritten Document Images.&lt;br /&gt;
&lt;br /&gt;
There are actually two tasks, depending on the nature of the binarization method used.&lt;br /&gt;
&lt;br /&gt;
#	For regular binarization methods, the task is to binarize all 15 document images.&lt;br /&gt;
#	For learning-based binarization methods, the task is to use images number 1 to 5 for training, and then binarize images number 6 to 16.&lt;br /&gt;
&lt;br /&gt;
A few baseline methods have been provided: PC (phase congruency) binarization method [Ziaei2012], and SGL/BGL binarization method [Farrahi2009, Farrahi2010]. The SGL/BGL method uses a rough binarization as its initialization.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Evaluation Protocol=&lt;br /&gt;
&lt;br /&gt;
#	For regular methods, the average F-measure of the binarized images against the provided ground truth is used as the performance of the binarization method in question.&lt;br /&gt;
#	For learning-based methods, the average F-measure of the binarized images number 6 to 15 against the provided ground truth is used as the performance of the binarization method in question.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; cellpadding=&amp;quot;5&amp;quot; align=&amp;quot;center&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Method&lt;br /&gt;
!Whole set&lt;br /&gt;
!Training set&lt;br /&gt;
!Test set&lt;br /&gt;
|- &lt;br /&gt;
|Otsu (regular)&lt;br /&gt;
|82.09&lt;br /&gt;
|90.76&lt;br /&gt;
|77.75&lt;br /&gt;
|-&lt;br /&gt;
|PC (regular)&lt;br /&gt;
|90.91&lt;br /&gt;
|92.33&lt;br /&gt;
|90.20&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (upper bound; rough bin: GT)&lt;br /&gt;
|94.55&lt;br /&gt;
|95.37&lt;br /&gt;
|94.14&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (upper bound; rough bin: PC)&lt;br /&gt;
|91.79&lt;br /&gt;
|93.29&lt;br /&gt;
|91.04&lt;br /&gt;
|-&lt;br /&gt;
|SGL/BGL (learning) (rough bin: PC)&lt;br /&gt;
|N/A&lt;br /&gt;
|N/A&lt;br /&gt;
|89.94&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and background gray level (BGL) is provided. The executable of the method will be provided in near future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]]&lt;br /&gt;
&lt;br /&gt;
=Related Ground Truth Data=&lt;br /&gt;
* [[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian historical document dataset with introduction to PhaseGT: A ground truthing application, to be submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: 10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework for the optimization and automation of document binarization methods, Computer Vision and Image Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Training.txt  Train Set Meta Data] (0 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Test.txt   Test Set Meta Data] (0 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Otsu_PHIBD_2012.zip   baseline method] (0.39 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/PC_PHIBD_2012.zip   baseline method] (0.33 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PHIBD_2012.zip   baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PC_PHIBD_2012.zip   baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_PC_TrainTest_PHIBD_2012.zip   baseline method] (0.36 Mb)&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/SGLBGL_metacode.m   Sample Program] (0 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Binarized_images_for_PHIBD_2012_dataset&amp;diff=1901</id>
		<title>Binarized images for PHIBD 2012 dataset</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Binarized_images_for_PHIBD_2012_dataset&amp;diff=1901"/>
		<updated>2013-07-03T15:31:01Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Version 1.0 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Binarization, Text detection, Ground truth&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
For each image, a manually binarized and verified image is provided on which the text pixels are marked in black while the rest of pixels are marked in white. An automatic binarization method for ground truthing was used prior to the manual verification of the binary images in order to speed up the process. The automatic binarization method will be described in a paper (to be submitted to ICDAR’13).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/GT.zip Ground truth images] (0.28 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1900</id>
		<title>Persian Heritage Image Binarization Dataset (PHIBD 2012)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1900"/>
		<updated>2013-07-03T15:30:40Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Version 1.0 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Hossein Ziaie Nafchi, Seyed Morteza Ayatollahi, Reza Farrahi Moghaddam, and Mohamed Cheriet&lt;br /&gt;
 Synchromedia Laboratory&lt;br /&gt;
 ETS, Montreal, (Quebec) Canada&lt;br /&gt;
 H3C 1K3&lt;br /&gt;
 E-mail: mohamed.cheriet@etsmtl.ca&lt;br /&gt;
 Tel: +1(514)396-8972&lt;br /&gt;
 Fax: +1(514)396-8595&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document Image Binarization, Persian Heritage, Handwritten manuscripts&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset contains 15 historical and old manuscript images collected from the historical &lt;br /&gt;
&lt;br /&gt;
records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated &lt;br /&gt;
&lt;br /&gt;
with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation &lt;br /&gt;
&lt;br /&gt;
including bleed-through, faded ink, and blur. The dataset is the first in a series to provide &lt;br /&gt;
&lt;br /&gt;
document images and their ground truth as a contribution to Document image analysis and &lt;br /&gt;
&lt;br /&gt;
recognition (DAIR) community.&lt;br /&gt;
&lt;br /&gt;
It is planned to increase the dataset in future and to create a dataset which also covers the tasks of understanding in the near future.&lt;br /&gt;
&lt;br /&gt;
=Metadata and Technical Details=&lt;br /&gt;
As metadata, the types of degradation on each document image have been provided in two text &lt;br /&gt;
&lt;br /&gt;
files: 1) for images number 1 to 5 and 2) for images number 6 to 15. It is worth noting that &lt;br /&gt;
&lt;br /&gt;
images number 1 to 5 are considered as the training set while images number 6 to 15 are &lt;br /&gt;
&lt;br /&gt;
considered as the test set for those binarization methods that are based on a learning technique.&lt;br /&gt;
&lt;br /&gt;
Also, the estimated line height and stroke width for each image are provided in these files.&lt;br /&gt;
&lt;br /&gt;
The original document images are 4.9MB, while their ground truth images are 324KB.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and &lt;br /&gt;
&lt;br /&gt;
background gray level (BGL) is provided. The executable of the method will be provided in near &lt;br /&gt;
&lt;br /&gt;
future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-&lt;br /&gt;
&lt;br /&gt;
adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is &lt;br /&gt;
&lt;br /&gt;
the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian &lt;br /&gt;
&lt;br /&gt;
historical document dataset with introduction to PhaseGT: A ground truthing application, to be &lt;br /&gt;
&lt;br /&gt;
submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical &lt;br /&gt;
&lt;br /&gt;
Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, &lt;br /&gt;
&lt;br /&gt;
Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided &lt;br /&gt;
&lt;br /&gt;
low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: &lt;br /&gt;
&lt;br /&gt;
10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive &lt;br /&gt;
&lt;br /&gt;
binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, &lt;br /&gt;
&lt;br /&gt;
p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework &lt;br /&gt;
&lt;br /&gt;
for the optimization and automation of document binarization methods, Computer Vision and Image &lt;br /&gt;
&lt;br /&gt;
Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/Original.zip Original images] (5 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Binarized_images_for_PHIBD_2012_dataset&amp;diff=1899</id>
		<title>Binarized images for PHIBD 2012 dataset</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Binarized_images_for_PHIBD_2012_dataset&amp;diff=1899"/>
		<updated>2013-07-03T15:28:23Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Keywords= Binar…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Binarization, Text detection, Ground truth&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
For each image, a manually binarized and verified image is provided on which the text pixels are marked in black while the rest of pixels are marked in white. An automatic binarization method for ground truthing was used prior to the manual verification of the binary images in order to speed up the process. The automatic binarization method will be described in a paper (to be submitted to ICDAR’13).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
* [http://www.iapr-tc11.org/dataset/PHIBD2012/GT.zip Ground truth images] (2.8 Mb)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1894</id>
		<title>Persian Heritage Image Binarization Dataset (PHIBD 2012)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1894"/>
		<updated>2013-05-30T18:25:05Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Hossein Ziaie Nafchi, Seyed Morteza Ayatollahi, Reza Farrahi Moghaddam, and Mohamed Cheriet&lt;br /&gt;
 Synchromedia Laboratory&lt;br /&gt;
 ETS, Montreal, (Quebec) Canada&lt;br /&gt;
 H3C 1K3&lt;br /&gt;
 E-mail: mohamed.cheriet@etsmtl.ca&lt;br /&gt;
 Tel: +1(514)396-8972&lt;br /&gt;
 Fax: +1(514)396-8595&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document Image Binarization, Persian Heritage, Handwritten manuscripts&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset contains 15 historical and old manuscript images collected from the historical &lt;br /&gt;
&lt;br /&gt;
records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated &lt;br /&gt;
&lt;br /&gt;
with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation &lt;br /&gt;
&lt;br /&gt;
including bleed-through, faded ink, and blur. The dataset is the first in a series to provide &lt;br /&gt;
&lt;br /&gt;
document images and their ground truth as a contribution to Document image analysis and &lt;br /&gt;
&lt;br /&gt;
recognition (DAIR) community.&lt;br /&gt;
&lt;br /&gt;
It is planned to increase the dataset in future and to create a dataset which also covers the tasks of understanding in the near future.&lt;br /&gt;
&lt;br /&gt;
=Metadata and Technical Details=&lt;br /&gt;
As metadata, the types of degradation on each document image have been provided in two text &lt;br /&gt;
&lt;br /&gt;
files: 1) for images number 1 to 5 and 2) for images number 6 to 15. It is worth noting that &lt;br /&gt;
&lt;br /&gt;
images number 1 to 5 are considered as the training set while images number 6 to 15 are &lt;br /&gt;
&lt;br /&gt;
considered as the test set for those binarization methods that are based on a learning technique.&lt;br /&gt;
&lt;br /&gt;
Also, the estimated line height and stroke width for each image are provided in these files.&lt;br /&gt;
&lt;br /&gt;
The original document images are 4.9MB, while their ground truth images are 324KB.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and &lt;br /&gt;
&lt;br /&gt;
background gray level (BGL) is provided. The executable of the method will be provided in near &lt;br /&gt;
&lt;br /&gt;
future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-&lt;br /&gt;
&lt;br /&gt;
adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is &lt;br /&gt;
&lt;br /&gt;
the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian &lt;br /&gt;
&lt;br /&gt;
historical document dataset with introduction to PhaseGT: A ground truthing application, to be &lt;br /&gt;
&lt;br /&gt;
submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical &lt;br /&gt;
&lt;br /&gt;
Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, &lt;br /&gt;
&lt;br /&gt;
Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided &lt;br /&gt;
&lt;br /&gt;
low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: &lt;br /&gt;
&lt;br /&gt;
10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive &lt;br /&gt;
&lt;br /&gt;
binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, &lt;br /&gt;
&lt;br /&gt;
p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework &lt;br /&gt;
&lt;br /&gt;
for the optimization and automation of document binarization methods, Computer Vision and Image &lt;br /&gt;
&lt;br /&gt;
Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1893</id>
		<title>Persian Heritage Image Binarization Dataset (PHIBD 2012)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1893"/>
		<updated>2013-05-30T18:21:09Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Copyright */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Hossein Ziaie Nafchi, Seyed Morteza Ayatollahi, Reza Farrahi Moghaddam, and Mohamed Cheriet&lt;br /&gt;
 Synchromedia Laboratory&lt;br /&gt;
 ETS, Montreal, (Quebec) Canada&lt;br /&gt;
 H3C 1K3&lt;br /&gt;
 E-mail: mohamed.cheriet@etsmtl.ca&lt;br /&gt;
 Tel: +1(514)396-8972&lt;br /&gt;
 Fax: +1(514)396-8595&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document Image Binarization, Persian Heritage, Handwritten manuscripts&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset contains 15 historical and old manuscript images collected from the historical &lt;br /&gt;
&lt;br /&gt;
records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated &lt;br /&gt;
&lt;br /&gt;
with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation &lt;br /&gt;
&lt;br /&gt;
including bleed-through, faded ink, and blur. The dataset is the first in a series to provide &lt;br /&gt;
&lt;br /&gt;
document images and their ground truth as a contribution to Document image analysis and &lt;br /&gt;
&lt;br /&gt;
recognition (DAIR) community.&lt;br /&gt;
&lt;br /&gt;
It is planned to increase the dataset in future and to create a dataset which also covers the tasks of understanding in the near future.&lt;br /&gt;
&lt;br /&gt;
=Metadata and Technical Details=&lt;br /&gt;
As metadata, the types of degradation on each document image have been provided in two text &lt;br /&gt;
&lt;br /&gt;
files: 1) for images number 1 to 5 and 2) for images number 6 to 15. It is worth noting that &lt;br /&gt;
&lt;br /&gt;
images number 1 to 5 are considered as the training set while images number 6 to 15 are &lt;br /&gt;
&lt;br /&gt;
considered as the test set for those binarization methods that are based on a learning technique.&lt;br /&gt;
&lt;br /&gt;
Also, the estimated line height and stroke width for each image are provided in these files.&lt;br /&gt;
&lt;br /&gt;
The original document images are 4.9MB, while their ground truth images are 324KB.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
[[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
[[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and &lt;br /&gt;
&lt;br /&gt;
background gray level (BGL) is provided. The executable of the method will be provided in near &lt;br /&gt;
&lt;br /&gt;
future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-&lt;br /&gt;
&lt;br /&gt;
adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is &lt;br /&gt;
&lt;br /&gt;
the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian &lt;br /&gt;
&lt;br /&gt;
historical document dataset with introduction to PhaseGT: A ground truthing application, to be &lt;br /&gt;
&lt;br /&gt;
submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical &lt;br /&gt;
&lt;br /&gt;
Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, &lt;br /&gt;
&lt;br /&gt;
Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided &lt;br /&gt;
&lt;br /&gt;
low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: &lt;br /&gt;
&lt;br /&gt;
10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive &lt;br /&gt;
&lt;br /&gt;
binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, &lt;br /&gt;
&lt;br /&gt;
p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework &lt;br /&gt;
&lt;br /&gt;
for the optimization and automation of document binarization methods, Computer Vision and Image &lt;br /&gt;
&lt;br /&gt;
Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1892</id>
		<title>Persian Heritage Image Binarization Dataset (PHIBD 2012)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Persian_Heritage_Image_Binarization_Dataset_(PHIBD_2012)&amp;diff=1892"/>
		<updated>2013-05-30T18:20:56Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Created page with &amp;quot;Datasets -&amp;gt; Datasets List -&amp;gt; Current Page  {| style=&amp;quot;width: 100%&amp;quot; |- | align=&amp;quot;right&amp;quot; |   {|  |- | '''Created: '''2013-05-30 |- | {{Last updated}} |}  |}  =Contact Author=…&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Hossein Ziaie Nafchi, Seyed Morteza Ayatollahi, Reza Farrahi Moghaddam, and Mohamed Cheriet&lt;br /&gt;
 Synchromedia Laboratory&lt;br /&gt;
 ETS, Montreal, (Quebec) Canada&lt;br /&gt;
 H3C 1K3&lt;br /&gt;
 E-mail: mohamed.cheriet@etsmtl.ca&lt;br /&gt;
 Tel: +1(514)396-8972&lt;br /&gt;
 Fax: +1(514)396-8595&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document Image Binarization, Persian Heritage, Handwritten manuscripts&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset contains 15 historical and old manuscript images collected from the historical &lt;br /&gt;
&lt;br /&gt;
records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated &lt;br /&gt;
&lt;br /&gt;
with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation &lt;br /&gt;
&lt;br /&gt;
including bleed-through, faded ink, and blur. The dataset is the first in a series to provide &lt;br /&gt;
&lt;br /&gt;
document images and their ground truth as a contribution to Document image analysis and &lt;br /&gt;
&lt;br /&gt;
recognition (DAIR) community.&lt;br /&gt;
&lt;br /&gt;
It is planned to increase the dataset in future and to create a dataset which also covers the tasks of understanding in the near future.&lt;br /&gt;
&lt;br /&gt;
=Metadata and Technical Details=&lt;br /&gt;
As metadata, the types of degradation on each document image have been provided in two text &lt;br /&gt;
&lt;br /&gt;
files: 1) for images number 1 to 5 and 2) for images number 6 to 15. It is worth noting that &lt;br /&gt;
&lt;br /&gt;
images number 1 to 5 are considered as the training set while images number 6 to 15 are &lt;br /&gt;
&lt;br /&gt;
considered as the test set for those binarization methods that are based on a learning technique.&lt;br /&gt;
&lt;br /&gt;
Also, the estimated line height and stroke width for each image are provided in these files.&lt;br /&gt;
&lt;br /&gt;
The original document images are 4.9MB, while their ground truth images are 324KB.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
[[Binarized images for PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
[[Binarization of PHIBD 2012 dataset]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
A metacode of a learning-based binarization method based on stroke gray level (SGL) and &lt;br /&gt;
&lt;br /&gt;
background gray level (BGL) is provided. The executable of the method will be provided in near &lt;br /&gt;
&lt;br /&gt;
future.&lt;br /&gt;
&lt;br /&gt;
The proposed learning-based binarization method uses the SGL and the BGL to determine a locally-&lt;br /&gt;
&lt;br /&gt;
adaptive threshold value based on a parameter (alpha). The optimal selection of this parameter is &lt;br /&gt;
&lt;br /&gt;
the learning part of this method. &lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* [Ziaei2013] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam, and Mohamed Cheriet. Persian &lt;br /&gt;
&lt;br /&gt;
historical document dataset with introduction to PhaseGT: A ground truthing application, to be &lt;br /&gt;
&lt;br /&gt;
submitted to ICDAR’13.&lt;br /&gt;
* [Ziaei2012] Hossein Ziaei Nafchi, Reza Farrahi Moghaddam and Mohamed Cheriet, Historical &lt;br /&gt;
&lt;br /&gt;
Document Binarization Based on Phase Information of Images, in ACCV’12 Workshop on e-Heritage, &lt;br /&gt;
&lt;br /&gt;
Daejeon, South Korea, Nov 5-10, 2012.&lt;br /&gt;
* [Farrahi2009] Reza Farrahi Moghaddam, and Mohamed Cheriet, RSLDI: Restoration of single-sided &lt;br /&gt;
&lt;br /&gt;
low-quality document images, Pattern Recognition, Volume 42, Issue 12, p.3355–3364 (2009) DOI: &lt;br /&gt;
&lt;br /&gt;
10.1016/j.patcog.2008.10.021&lt;br /&gt;
* [Farrahi2010] Reza Farrahi Moghaddam, and Mohamed Cheriet, A multi-scale framework for adaptive &lt;br /&gt;
&lt;br /&gt;
binarization of degraded document images, Pattern Recognition, Volume 43, Issue 6, Number 6, &lt;br /&gt;
&lt;br /&gt;
p.2186–2198 (2010) DOI: 10.1016/j.patcog.2009.12.024&lt;br /&gt;
* [Cheriet2012] Mohamed Cheriet, Reza Farrahi Moghaddam, and Rachid Hedjam, A learning framework &lt;br /&gt;
&lt;br /&gt;
for the optimization and automation of document binarization methods, Computer Vision and Image &lt;br /&gt;
&lt;br /&gt;
Understanding, Volume Accepted, p.– (2012) DOI: 10.1016/j.cviu.2012.11.003&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1891</id>
		<title>Datasets List</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets_List&amp;diff=1891"/>
		<updated>2013-05-30T18:20:36Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* On-line and Off-line */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]]&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
See the datasets [[Datasets per Journal / Conference|sorted according to the Journal / Conference]] they first appeared in.&lt;br /&gt;
&lt;br /&gt;
= Complex Text Containers =&lt;br /&gt;
== Scene Text ==&lt;br /&gt;
* [[MSRA Text Detection 500 Database (MSRA-TD500)]]&lt;br /&gt;
* [[The Street View Text Dataset]]&lt;br /&gt;
* [[The Street View House Numbers (SVHN) Dataset]]&lt;br /&gt;
* [[NEOCR: Natural Environment OCR Dataset]]&lt;br /&gt;
* [[KAIST Scene Text Database]]&lt;br /&gt;
* [[ICDAR 2003 Robust Reading Competitions]]&lt;br /&gt;
* [[ICDAR 2005 Robust Reading Competitions]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--&lt;br /&gt;
== Born Digital Images ==&lt;br /&gt;
* [[ICDAR 2011 Robust Reading Competition - Challenge 1: &amp;quot;Reading Text in Born-Digital Images (Web and Email)&amp;quot;]]&lt;br /&gt;
--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Machine-printed Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Table Ground Truth for the UW3 and UNLV datasets]]&lt;br /&gt;
* [[The DocLab Dataset for Evaluating Table Interpretation Methods]]&lt;br /&gt;
* [http://dataset.primaresearch.org/ PRImA Layout Analysis Dataset]&lt;br /&gt;
* [http://www.dfki.uni-kl.de/~shafait/downloads.html DFKI Dewarping Contest Dataset (CBDAR 2007)] The dataset, that was used in the CBDAR 2007 Dewarping Contest, contains 102 camera captured documents with their corresponding ASCII text ground-truth. Additionally, text-line level ground-truth was also prepared to benchmark curled text-line segmentation algorithms. Part of the dataset (76 out of 102 pages) were also scanned with a flat-bed scanner to create a ground-truth image for image based evaluation of page dewarping algorithms.&lt;br /&gt;
* [http://diuf.unifr.ch/diva/APTI/ APTI: Arabic Printed Text Image Database]&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]] This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010. The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution).&lt;br /&gt;
* [http://ciir.cs.umass.edu/downloads/ocr-evaluation/ RETAS OCR Evaluation Dataset] The RETAS dataset (used in the paper by Yalniz and Manmatha, ICDAR'11) is created to evaluate the optical character recognition (OCR) accuracy of real scanned books. The dataset contains real OCR outputs for 160 scanned books (100 English, 20 French, 20 German, 20 Spanish) downloaded from the Internet Archive website. The corresponding ground truth text for each scanned book is obtained from the Project Gutenberg database. The OCR output of each scanned book is aligned with its ground truth at the word and character level and the alignment output is provided along with estimated OCR accuracies. The dataset is provided for research purposes.&lt;br /&gt;
&lt;br /&gt;
= Graphical Documents =&lt;br /&gt;
&lt;br /&gt;
* [[Chem-Infty Dataset: A ground-truthed dataset of Chemical Structure Images]]&lt;br /&gt;
* [[Braille Dataset - Shiraz University]]&lt;br /&gt;
* [http://www.eurecom.fr/~huet/work.html TradeMarks Image Database] - By way of Benoit Huet, 999 trademark and logo images&lt;br /&gt;
&lt;br /&gt;
= Mixed Content Documents = &lt;br /&gt;
* [http://www.umiacs.umd.edu/~zhugy/Tobacco800.html Tobacco800 Document Image Database] - composed of 1290 document images collected and scanned using a wide variety of equipment over time.&lt;br /&gt;
&lt;br /&gt;
= Handwritten Documents =&lt;br /&gt;
== On-line and Off-line ==&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2009 Signature Verification Competition (SigComp2009)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2010 Signature Verification Competition (4NSigComp2010)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICDAR 2011 Signature Verification Competition (SigComp2011)]]&lt;br /&gt;
&lt;br /&gt;
* [[ICFHR 2012 Signature Verification Competition (4NSigComp2012)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.nlpr.ia.ac.cn/databases/handwriting/Home.html CASIA Online and Offline Chinese Handwriting Databases] - The Chinese handwriting datasets were produced by 1,020 writers using Anoto pen on papers, such that both online and offline data were obtained. Both the online and the offline dataset consists of three subsets for isolated characters (DB1.0–1.2, about 3.9 million samples of 7,356 classes) and three for handwritten texts (DB2.0–2.2, about 5,090 pages and 1.35 million characters). The datasets are free for academic research for handwritten document segmentation and retrieval, character and text line recognition, writer adaptation and identification.&lt;br /&gt;
&lt;br /&gt;
* [[Persian Heritage Image Binarization Dataset (PHIBD 2012)]] This dataset contains 15 historical and old manuscript images collected from the historical records at the Documents and old manuscripts treasury of Mirza Mohammad Kazemaini (affiliated with Hazrate Emamzadeh Jafar), Yazd, Iran. The images suffer from various types of degradation including bleed-through, faded ink, and blur. The dataset is the first in a series to provide document images and their ground truth as a contribution to Document image analysis and recognition (DAIR) community. It is planned to provide more data and ground-truth information in the fture.&lt;br /&gt;
&lt;br /&gt;
== On-line ==&lt;br /&gt;
* [[CROHME: Competition on Recognition of Online Handwritten Mathematical Expressions]]&lt;br /&gt;
&lt;br /&gt;
* [[Devanagari Character Dataset]]&lt;br /&gt;
&lt;br /&gt;
* [[Harbin Institute of Technology Opening Recognition Corpus for Chinese Characters (HIT-OR3C)]]&lt;br /&gt;
&lt;br /&gt;
* [[IAM Online Document Database (IAMonDo-database)]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-on-line-handwriting-database IAM On-Line Handwriting Database]&lt;br /&gt;
&lt;br /&gt;
* [http://hwr.nici.kun.nl/unipen/ UNIPEN database] (Click on link 'CDROMs')&lt;br /&gt;
&lt;br /&gt;
* [http://www.tuat.ac.jp/~nakagawa/database/ Nakagawa Lab Online Handwriting Database]&lt;br /&gt;
&amp;lt;!--** Reference: [http://www.springerlink.com/content/kmhh7dg6h8cgr6a5/ Masaki Nakagawa and Kaoru Matsumoto: &amp;quot;Collection of on-line handwritten Japanese character pattern databases and their analysis,&amp;quot; International Journal on Document Analysis and Recognition, Vol. 7 No. 1, pp.69-81 (2004)].--&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* [http://www.ai.rug.nl/~lambert/unipen/icdar-03-competition/ The Informal Competition of Recognizing On-line Words (ICROW)] by the Unipen Foundation&lt;br /&gt;
&lt;br /&gt;
== Off-line ==&lt;br /&gt;
* [http://www.rimes-database.fr/wiki/doku.php The Rimes Database] comprises 12,723 handwritten pages corresponding to 5605 mails of two to three pages. It was collected by asking volunteers to write a letter given one of nine predefined scenarios related to business/customer relations. The dataset has been used in numerous competitions in ICDAR and ICFHR. It is available for research purposes only, through the Web site of the authors.&lt;br /&gt;
&lt;br /&gt;
* [[IBN SINA: A database for research on processing and understanding of Arabic manuscripts images]]&lt;br /&gt;
&lt;br /&gt;
* [http://www.cedar.buffalo.edu/Databases/CDROM1/ CEDAR Off-line Handwriting CDROM1]&lt;br /&gt;
&lt;br /&gt;
* [http://www.iam.unibe.ch/fki/databases/iam-handwriting-database IAM Database] - A full English sentence database for off-line handwriting recognition.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/germana The GERMANA Dataset] - GERMANA is the result of digitising and annotating a 764-page Spanish manuscript entitled “Noticias y documentos relativos a Doña Germana de Foix, ́última Reina de Aragón&amp;quot;, written in 1891 by Vicent Salvador. It contains approximately 21K text lines manually marked and transcribed by palaeography experts.&lt;br /&gt;
&lt;br /&gt;
* [http://prhlt.iti.upv.es/page/projects/multimodal/idoc/rodrigo The RODRIGO Dataset] - RODRIGO is the result of digitising and annotating a manuscript dated 1545. Digitisation was done at 300dpi in color by the Spanish Culture Ministry. The original manuscript is a 853-page bound volume, entitled &amp;quot;Historia de España del arçobispo Don Rodrigo&amp;quot;, completely written in old Castilian (Spanish) by a single author. Annotation exists for text blocks, lines and transcriptions, resulting in approximately 20K lines and 231K running words from a lexicon of 17K words.&lt;br /&gt;
&lt;br /&gt;
* [http://marg.nlm.nih.gov/ MARG- Medical Article Records Groundtruth] - A freely-available repository of document page images and their associated textual and layout data. The data has been reviewed and corrected to establish its &amp;quot;ground truth&amp;quot;. Please contact Dr. George Thoma (thoma@lhc.nlm.nih.gov) at the National Library of Medicine for more information.&lt;br /&gt;
&lt;br /&gt;
* [http://kornai.com/Hindi/ Hindi font samples] by Andras Kornai, June 5 2003&lt;br /&gt;
&lt;br /&gt;
= Software and Tools =&lt;br /&gt;
* [http://lampsrv02.umiacs.umd.edu/projdb/project.php?id=53 GEDI: Groundtruthing Environment for Document Images] - A generic annotation tool for scanned text documents.&lt;br /&gt;
* [http://www2.parc.com/isl/groups/pda/pixlabeler/index.html PixLabeler] - a research tool for labeling elements in a document image at a pixel level.&lt;br /&gt;
* [http://code.google.com/p/ocropus/ OCRopus(tm)] - The OCRopus(tm) open source document analysis and OCR system&lt;br /&gt;
* [http://htk.eng.cam.ac.uk/ The Hidden Markov Model Toolkit (HTK)] - a portable toolkit for building and manipulating hidden Markov models&lt;br /&gt;
* [https://github.com/meierue/RNNLIB Bidirectional Long-Short Term Memory Networks] - Implementation of Bidirectional Long-Short Term Memory Networks (BLSTM) combined with Connectionist Temporal Classification (CTC) - including examples for Arabic recognition.&lt;br /&gt;
* [http://www.speech.sri.com/projects/srilm/ SRILM - The SRI Language Modeling Toolkit] - SRILM is a toolkit for building and applying statistical language models (LMs), primarily for use in speech recognition, statistical tagging and segmentation, and machine translation.&lt;br /&gt;
* [http://torch5.sourceforge.net/ Torch 5] - a Matlab-like environment for state-of-the-art machine learning algorithms.&lt;br /&gt;
* [http://www.prtools.org/ PRTools] - a Matlab based toolbox for pattern recognition&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- [[Datasets (old page)| Old Page (contains broken links)]] --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1890</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1890"/>
		<updated>2013-05-30T17:37:27Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Submitted Files */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
To be available soon. Please refer to [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD] for downloading the files from the origninal datasets site.&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1889</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1889"/>
		<updated>2013-05-30T17:36:47Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1888</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1888"/>
		<updated>2013-05-30T17:36:22Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
The dataset is also available at [http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD#Setup_script_v1_0 http://www.lrde.epita.fr/cgi-bin/twiki/view/Olena/DatasetDBD#Setup_script_v1_0]&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Datasets&amp;diff=1887</id>
		<title>Datasets</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Datasets&amp;diff=1887"/>
		<updated>2013-05-30T17:31:28Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Overview – Message from TC-11 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Overview – Message from TC-11=&lt;br /&gt;
[[Image:ThisWayToTheDatasets.png|250px|right|link=Datasets List| Datasets List]]&lt;br /&gt;
&lt;br /&gt;
It is extremely important for the Document Image Analysis and Recognition community to be able to cross check and reproduce results described in published papers in the field. In order to achieve this, any datasets used as the basis for publications should be publicly available, as is the norm in many other disciplines.&lt;br /&gt;
&lt;br /&gt;
Authors are actively encouraged to submit the datasets they used to train and / or evaluate their algorithms to the TC-11 in order for them to be published on the TC-11 Web site.&lt;br /&gt;
&lt;br /&gt;
This initiative is not restricted to datasets. At TC-11 we are interested in archiving online any piece of data (ground-truth data, software, etc) which would allow to easily reproduce results, set new targets, foster healthy competition, encourage collaboration and generally advance the DIAR field as a whole.&lt;br /&gt;
&lt;br /&gt;
A wealth of datasets and corresponding ground truth data are already available through the TC-11 [[Datasets List]].&lt;br /&gt;
&lt;br /&gt;
If you wish to contribute, please read below about the procedure to submit material to the TC-11. For any comments or suggestions, please contact Marcus Liwicki, the dataset curator at liwicki (at) liwicki@dfki.uni-kl.de.&lt;br /&gt;
&lt;br /&gt;
=Submission Protocol=&lt;br /&gt;
The process of submitting a dataset / ground truth to the TC-11 is the following:&lt;br /&gt;
&lt;br /&gt;
# Please fill in the form below, and send it by email to Dimosthenis Karatzas (dimos at cvc.uab.es), the TC-11 dataset curator.&lt;br /&gt;
# The TC-11 dataset curator will review the submission request, and ensure that all information is clear and complete, and any copyright issues are properly addressed.&lt;br /&gt;
# As soon as all information is in place, the TC-11 dataset curator will ask you to sign and return by fax the final submission form.&lt;br /&gt;
# The TC-11 dataset curator will work with you to upload the dataset to the TC-11 Web site. Depending on the nature of the dataset this might be as easy as sending a CD or uploading the required files.&lt;br /&gt;
&lt;br /&gt;
The TC-11 is actively working towards a more comprehensive way of dealing with datasets and associated information.&lt;br /&gt;
&lt;br /&gt;
=Copyright Note=&lt;br /&gt;
TC-11 provides dataset hosting services as a benefit to the international research community. If it is determined that copyrighted material is improperly included in a dataset submitted to inclusion on the TC-11 website, we will immediately remove the offending material upon notification of the copyright holder.&lt;br /&gt;
&lt;br /&gt;
By submitting a dataset for inclusion to the TC-11 Web site, the author certifies that he/she has the right to publish the dataset and any associated data in the public domain and the act of doing so does not violate intellectual property rights or copyrights of some third party.&lt;br /&gt;
&lt;br /&gt;
The TC-11 will provide a service through which the submitted dataset and any associated data will be made public to the Document Analysis community worldwide. In case any legal dispute arises in the future in relation to the publishing of this dataset and associated data in the public domain, the author will hold TC-11 free from any wrongdoing and accept responsibility for the publication of these data.&lt;br /&gt;
&lt;br /&gt;
By submitting a dataset and associated data to the TC-11, you explicitly accept that any third party can independently submit additional information that relates to the original dataset (e.g. additional ground-truth data, software, etc).&lt;br /&gt;
&lt;br /&gt;
We strongly encourage the authors, where they own the copyrights of the submitted information, to consider offering it to the community under a [http://creativecommons.org/choose/| creative commons license]. See [http://wiki.creativecommons.org/Before_Licensing/| this link] for guidelines about how to choose a proper Creative Commons license.&lt;br /&gt;
&lt;br /&gt;
=Useful Definitions=&lt;br /&gt;
'''Dataset''': A collection of data along with metadata information, as required to use these data.&lt;br /&gt;
&lt;br /&gt;
'''Metadata''': Metadata is information specific to a particular dataset. Metadata are usually tightly structured within the dataset itself (e.g. information encoded within the filenames of submitted images). Metadata can only be submitted at the time of submission of the dataset.&lt;br /&gt;
&lt;br /&gt;
'''Ground Truth Specification''': The definition of the required information that accurately describes a particular aspect of the data at a high level where agreement between different human observers can be established, as well as the definition of an appropriate structure (format) for storing this information.&lt;br /&gt;
&lt;br /&gt;
'''Ground Truth Data''': A set of data conforming to a particular ground truth specification and relating to a specific dataset. Ground Truth Data can be submitted at any time, while different Ground Truth Data (corresponding to different aspects of the data) can be associated with the same dataset.&lt;br /&gt;
&lt;br /&gt;
'''Task''': A well defined process to evaluate algorithms in the context of a specific scientific problem. A task would typically provide a specific evaluation protocol, and link to specific resources as required (a dataset, and usually related ground truth data). Tasks should correspond to open challenges in the field. If you undertake any of the tasks defined and you have published results or code available, we would really like to know!&lt;br /&gt;
&lt;br /&gt;
'''Resources''': Any other type of related resources that are not specifically covered by the above definitions. Examples would include software to browse and visualise a dataset, software to create ground truth data, algorithms to do performance evaluation, codecs, reports, publications, etc.&lt;br /&gt;
&lt;br /&gt;
=Submission Form=&lt;br /&gt;
You can download the submission form from here: [[media:TC11_Dataset_Submission_Form_v09.pdf‎| PDF]] or [[Media:TC11_Dataset_Submission_Form_v09.docx| Word 2007]]&lt;br /&gt;
&lt;br /&gt;
The submission form has six sections. Please fill in the sections applicable to your situation. Not all sections need to be filled in. It would be helpful to have a look at already published datasets to get an idea of the information that is needed. Feel free to use as much space as you need, but generally a couple of paragraphs are more than enough to describe every aspect of the submitted material.&lt;br /&gt;
&lt;br /&gt;
=The TC11 Datasets=&lt;br /&gt;
The current list of datasets sorted on research topic can be found [[Datasets List|here]].&lt;br /&gt;
&lt;br /&gt;
The list of datasets sorted according to the Journal or Conference they first appeared in can be found [[Datasets per Journal / Conference|here]].&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_text_line_localization&amp;diff=1886</id>
		<title>Ground Truth for LRDE DBD text line localization</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=Ground_Truth_for_LRDE_DBD_text_line_localization&amp;diff=1886"/>
		<updated>2013-05-30T17:30:32Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Submitted Files */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
scanned, magazine, documents, text line localization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
&lt;br /&gt;
Text Lines Localization Information has been made available by applying text line localization algorithms. The size category of the text depends on the x-height and is considered with the following rule: 0 &amp;lt; small &amp;lt;= 30 &amp;lt; medium &amp;lt;= 55 &amp;lt; large &amp;lt; +inf&lt;br /&gt;
&lt;br /&gt;
* 123 large text lines localization (clean)&lt;br /&gt;
* 320 medium text lines localization (clean).&lt;br /&gt;
* 9551 small text lines localization (clean).&lt;br /&gt;
* 123 large text lines localization (original).&lt;br /&gt;
* 320 medium text lines localization (original).&lt;br /&gt;
* 9551 small text lines localization (original).&lt;br /&gt;
* 123 large text lines localization (scanned).&lt;br /&gt;
* 320 medium text lines localization (scanned).&lt;br /&gt;
* 9551 small text lines localization (scanned).&lt;br /&gt;
&lt;br /&gt;
The text lines dataset covers only a subset of the full-document dataset. It is generated from the binarization of the full-document images.&lt;br /&gt;
Text line localizations are stored as bounding box coordinates in text files. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities :&lt;br /&gt;
&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Related Dataset=&lt;br /&gt;
* [[LRDE Document Binarization Dataset (LRDE DBD)]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* none&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
* [http://www.iapr-tc11.org/mediawiki/images/LRDE_DBD_ouvel_obs_2402_textlines-1.0.zip http://www.iapr-tc11.org/mediawiki/images/LRDE_DBD_ouvel_obs_2402_textlines-1.0.zip]&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=File:LRDE_DBD_ouvel_obs_2402_textlines-1.0.zip&amp;diff=1885</id>
		<title>File:LRDE DBD ouvel obs 2402 textlines-1.0.zip</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=File:LRDE_DBD_ouvel_obs_2402_textlines-1.0.zip&amp;diff=1885"/>
		<updated>2013-05-30T17:27:39Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: Text line ground truth for http://www.iapr-tc11.org/mediawiki/index.php/Ground_Truth_for_LRDE_DBD_text_line_localization&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Text line ground truth for http://www.iapr-tc11.org/mediawiki/index.php/Ground_Truth_for_LRDE_DBD_text_line_localization&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1884</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1884"/>
		<updated>2013-05-30T17:23:42Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Contact Author */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
=Copyright=&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [http://www.nouvelobs.com/ Le Nouvel Observateur]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]&lt;br /&gt;
* [http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
	<entry>
		<id>http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1883</id>
		<title>LRDE Document Binarization Dataset (LRDE DBD)</title>
		<link rel="alternate" type="text/html" href="http://iapr-tc11.org/mediawiki/index.php?title=LRDE_Document_Binarization_Dataset_(LRDE_DBD)&amp;diff=1883"/>
		<updated>2013-05-30T17:22:58Z</updated>

		<summary type="html">&lt;p&gt;Liwicki: /* Contact Author */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Datasets]] -&amp;gt; [[Datasets List]] -&amp;gt; Current Page&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;width: 100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| align=&amp;quot;right&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
{| &lt;br /&gt;
|-&lt;br /&gt;
| '''Created: '''2013-05-30&lt;br /&gt;
|-&lt;br /&gt;
| {{Last updated}}&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Contact Author=&lt;br /&gt;
 Thierry Géraud – thierry.geraud@lrde.epita.fr&lt;br /&gt;
 EPITA Research and Development Laboratory (LRDE)&lt;br /&gt;
 14-16 rue Voltaire  F-94276 Le Kremlin-Bicetre  France&lt;br /&gt;
&lt;br /&gt;
LRDE is the copyright holder of all the images included in the dataset except for the original documents subset which are copyrighted from [[http://www.nouvelobs.com/ Le Nouvel Observateur]]. This work is based on the French magazine Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
You are allowed to reuse these documents for research purpose for evaluation and illustration. If so, please specify the following copyright: &amp;quot;Copyright (c) 2012. EPITA Research and Development Laboratory (LRDE) with permission from Le Nouvel Observateur&amp;quot;. You are not allowed to redistribute this dataset.&lt;br /&gt;
&lt;br /&gt;
If you use this dataset, please also cite the most appropriate paper from this list:&lt;br /&gt;
* [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition (IJDAR), 2013]]&lt;br /&gt;
* [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201109-ICDAR The SCRIBO Module of the Olena Platform: a Free Software Framework for Document Image Analysis. In the proceedings of the 11th International Conference on Document Analysis and Recognition (ICDAR), 2011.]]&lt;br /&gt;
&lt;br /&gt;
This data set is provided &amp;quot;as is&amp;quot; and without any express or implied warranties, including, without limitation, the implied warranties of merchantability and fitness for a particular purpose.&lt;br /&gt;
&lt;br /&gt;
=Current Version=&lt;br /&gt;
1.0&lt;br /&gt;
&lt;br /&gt;
=Keywords=&lt;br /&gt;
Document binarization, Magazine, Scanned &lt;br /&gt;
&lt;br /&gt;
=Description=&lt;br /&gt;
This dataset is composed of documents images extracted from the same French magazine: Le Nouvel Observateur, issue 2402, November 18th-24th, 2010.&lt;br /&gt;
&lt;br /&gt;
The provided dataset is composed of 375 Full-Document Images (A4 format, 300-dpi resolution)&lt;br /&gt;
&lt;br /&gt;
* 125 numerical &amp;quot;original documents&amp;quot; extracted from a PDF, with full OCR groundtruth.&lt;br /&gt;
* 125 numerical &amp;quot;clean documents&amp;quot; created from the &amp;quot;original documents&amp;quot; where images have been removed.&lt;br /&gt;
* 125 &amp;quot;scanned documents&amp;quot; based on the &amp;quot;clean documents&amp;quot;. They have been printed, scanned and registered to match the &amp;quot;clean documents&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Purpose of the three document qualities:&lt;br /&gt;
* Original : evaluate the binarization quality on perfect documents mixing text and images.&lt;br /&gt;
* Clean : evaluate the binarization quality on perfect document with text only.&lt;br /&gt;
* Scanned : evaluate the binarization quality on slightly degraded documents with text only.&lt;br /&gt;
&lt;br /&gt;
=Ground Truth Data=&lt;br /&gt;
* [[Ground Truth for LRDE DBD text line localization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD binarization]]&lt;br /&gt;
* [[Ground Truth for LRDE DBD OCR]]&lt;br /&gt;
&lt;br /&gt;
=Related Tasks=&lt;br /&gt;
* [[Document Binarization Evaluation for LRDE DBD]]&lt;br /&gt;
&lt;br /&gt;
=Software=&lt;br /&gt;
* A setup script is provided to download and configure the benchmarking environment. This is the recommanded way to run this benchmark. Note that this script also includes features to update the dataset if a new version is released.&lt;br /&gt;
* A Python script is provided to launch the benchmark and compute scores.&lt;br /&gt;
* C++ programs (and sources) are provided for performing evaluations and reading ground-truth data.&lt;br /&gt;
* 6 binarization algorithms (and their respective C++ sources) are provided and compiled to run this benchmark on their results.&lt;br /&gt;
&lt;br /&gt;
Minimum requirements: 5GB of free space, Linux (Ubuntu, Debian, …)&lt;br /&gt;
&lt;br /&gt;
Dependencies: Python 2.7, tesseract-ocr, tesseract-ocr-fra, git, libgraphicsmagick++1-dev, graphicsmagick-imagemagick-compat, graphicsmagick-libmagick-dev-compat, build-essential. libtool. automake, autoconf. g++-4.6, libqt4-dev (installed automatically with the setup script on Ubuntu and Debian).&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
* G. Lazzara, T. Géraud. Efficient Multiscale Sauvola's Binarization. In International Journal of Document Analysis and Recognition 2013 [[http://www.lrde.epita.fr/cgi-bin/twiki/view/Publications/201302-IJDAR]]&lt;br /&gt;
&lt;br /&gt;
=Submitted Files=&lt;br /&gt;
==Version 1.0==&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
This page is editable only by [[IAPR-TC11:Reading_Systems#TC11_Officers|TC11 Officers ]].&lt;/div&gt;</summary>
		<author><name>Liwicki</name></author>
		
	</entry>
</feed>