5
Ruling-Based Table Analysis for Noisy Handwritten Documents Jin Chen CSE Department 19 Memorial Drive West Bethlehem, PA 18015, USA [email protected] Daniel Lopresti CSE Department 19 Memorial Drive West Bethlehem, PA 18015, USA [email protected] ABSTRACT Table analysis can be a valuable step in document image analy- sis. In the case of noisy handwritten documents, various artifacts complicate the task of locating tables on a page and segmenting them into cells. Our ruling-based approach first detects line seg- ments to ensure high recall of table rulings, and then computes the intersections of horizontal and vertical rulings as key points. We then employ an optimization procedure to select the most probable subset of these key points which constitute the table structure. Fi- nally, we decompose a table into a 2-D arrangement of cells using the key points. Experimental evaluation involving 61 handwritten pages from 17 table classes show a table cell precision of 89% and a recall of 88%. 1. INTRODUCTION Tables are commonly expressed in various means to convey in- formation, including ASCII files, HTML webpages, PDF files, and paper [21, 6]. As one way of organizing relational data, tables have physical and logical structure [9, 20, 8]. Physical structure describes the locations of table components, e.g., headers, rows, columns, cells, rulings. Logical structure defines the way of con- necting table components to each other to form a set of relational n- tuples [21]. Note that table components may present both physical and logical structure from different perspectives. For instance, a ta- ble cell can be defined in logical structure by (Row[i], Column[j ]), and it can also be defined in the physical structure as a rectangu- lar region in the document image. In this paper, we constrain our discussions to the paper medium only. The target handwritten documents present characteristics that differ significantly from those collected in controlled environment such like research labs. To facilitate our discussion, it is useful to define substrate as a layer of all pre-printed information on a document page before people write on it. Our Arabic handwrit- ten dataset contains multiple components: machine-printed text substrate, handwriting, pre-printed rulings, signatures, logos, etc. In addition, the paper condition and scanner settings are different from those in controlled environment. For example, paper may be ripped, folded, and punched, while for scanning, we often ob- Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. MOCR ’13, August 24 2013, Washington, DC, USA Copyright 2013 ACM 978-1-4503-2114-3/13/08 ...$15.00. serve skewed pages and low image quality with plenty of “salt-and- pepper” noise, etc. Our evaluation dataset is a collection of Arabic handwritten documents containing tables and forms. Figure 1 is a sample document that illustrates several kinds of challenges present in the dataset. Hu et al. summarize the problem of table analysis as two sub- problems: detection and recognition [11]. Table detection focuses on finding table regions. Laurentini and Viada use horizontal and vertical rulings as initial evidence for tables in machine-printed documents, and then employ several tests to exclude non-tabular areas [12]. Some other work does not rely on the presence of rul- ing lines. Hu et al. introduce a table detection method that does not require ruling lines and can work on machine-printed document images and ASCII files [11]. After segmenting text into lines, they detect the inside-space between adjacent text blob and then em- ploy a dynamic programming (DP) technique to solve the problem of optimally decomposing the entire page into text lines and table rows. Shamalian et al. propose a method that uses pre-defined ta- ble model as complementary input [17]. Making use of the ink tem- plate of tables, they search for a best match of ink templates and the segmented text lines. Shafait and Smith extend table detection to multi-column documents [16]. However, existing techniques can- not be used directly on our evaluation documents because they usu- ally assume clean/simple input and/or well-segmented text lines. For complicated handwritten documents at hand, however, it is not a trivial task to make these existing techniques work on them. Table recognition usually assumes identified table regions and the goal is to find the physical structure and the logical structure of the table model [9, 20, 8]. There has been plenty of work deal- ing with machine-printed table recognition [12, 10, 2]. Gatos et al. [7] make use of the complete table rulings to recognize the ta- ble structure while Hirayama [10] relies on dynamic programming to align table columns. Richarz et al. propose a method of tabular structure recognition for their semi-supervised transcription system in handwritten historical weather reports [15]. Making use of the pre-printed table substrate, they use Hough transform to detect the horizontal and vertical rulings that constitute the tabular structure. Clawson et al. present a projection-profile based method to de- tect and extract handwritten tabular fields from historical census forms [5]. We notice that these existing techniques are evaluated on datasets where rulings are usually salient and well displaced, meaning no other lines will distract table analysis. This is, how- ever, not the case for our dataset where we need to handle severely broken lines and/or false-alarms. In this work, we address the problem of table detection and de- composition for noisy handwritten documents. These documents differ from most in the literature and they are not collected in a controlled environment but from the field. As a result, various types

Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

  • Upload
    dinhnga

  • View
    221

  • Download
    3

Embed Size (px)

Citation preview

Page 1: Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

Ruling-Based Table Analysis for NoisyHandwritten Documents

Jin ChenCSE Department

19 Memorial Drive WestBethlehem, PA 18015, [email protected]

Daniel LoprestiCSE Department

19 Memorial Drive WestBethlehem, PA 18015, [email protected]

ABSTRACTTable analysis can be a valuable step in document image analy-sis. In the case of noisy handwritten documents, various artifactscomplicate the task of locating tables on a page and segmentingthem into cells. Our ruling-based approach first detects line seg-ments to ensure high recall of table rulings, and then computes theintersections of horizontal and vertical rulings as key points. Wethen employ an optimization procedure to select the most probablesubset of these key points which constitute the table structure. Fi-nally, we decompose a table into a 2-D arrangement of cells usingthe key points. Experimental evaluation involving 61 handwrittenpages from 17 table classes show a table cell precision of 89% anda recall of 88%.

1. INTRODUCTIONTables are commonly expressed in various means to convey in-

formation, including ASCII files, HTML webpages, PDF files, andpaper [21, 6]. As one way of organizing relational data, tableshave physical and logical structure [9, 20, 8]. Physical structuredescribes the locations of table components, e.g., headers, rows,columns, cells, rulings. Logical structure defines the way of con-necting table components to each other to form a set of relational n-tuples [21]. Note that table components may present both physicaland logical structure from different perspectives. For instance, a ta-ble cell can be defined in logical structure by (Row[i], Column[j]),and it can also be defined in the physical structure as a rectangu-lar region in the document image. In this paper, we constrain ourdiscussions to the paper medium only.

The target handwritten documents present characteristics thatdiffer significantly from those collected in controlled environmentsuch like research labs. To facilitate our discussion, it is usefulto define substrate as a layer of all pre-printed information on adocument page before people write on it. Our Arabic handwrit-ten dataset contains multiple components: machine-printed textsubstrate, handwriting, pre-printed rulings, signatures, logos, etc.In addition, the paper condition and scanner settings are differentfrom those in controlled environment. For example, paper maybe ripped, folded, and punched, while for scanning, we often ob-

Permission to make digital or hard copies of all or part of this work forpersonal or classroom use is granted without fee provided that copies are notmade or distributed for profit or commercial advantage and that copies bearthis notice and the full citation on the first page. Copyrights for componentsof this work owned by others than ACM must be honored. Abstracting withcredit is permitted. To copy otherwise, or republish, to post on servers or toredistribute to lists, requires prior specific permission and/or a fee. Requestpermissions from [email protected] ’13, August 24 2013, Washington, DC, USACopyright 2013 ACM 978-1-4503-2114-3/13/08 ...$15.00.

serve skewed pages and low image quality with plenty of “salt-and-pepper” noise, etc. Our evaluation dataset is a collection of Arabichandwritten documents containing tables and forms. Figure 1 is asample document that illustrates several kinds of challenges presentin the dataset.

Hu et al. summarize the problem of table analysis as two sub-problems: detection and recognition [11]. Table detection focuseson finding table regions. Laurentini and Viada use horizontal andvertical rulings as initial evidence for tables in machine-printeddocuments, and then employ several tests to exclude non-tabularareas [12]. Some other work does not rely on the presence of rul-ing lines. Hu et al. introduce a table detection method that does notrequire ruling lines and can work on machine-printed documentimages and ASCII files [11]. After segmenting text into lines, theydetect the inside-space between adjacent text blob and then em-ploy a dynamic programming (DP) technique to solve the problemof optimally decomposing the entire page into text lines and tablerows. Shamalian et al. propose a method that uses pre-defined ta-ble model as complementary input [17]. Making use of the ink tem-plate of tables, they search for a best match of ink templates and thesegmented text lines. Shafait and Smith extend table detection tomulti-column documents [16]. However, existing techniques can-not be used directly on our evaluation documents because they usu-ally assume clean/simple input and/or well-segmented text lines.For complicated handwritten documents at hand, however, it is nota trivial task to make these existing techniques work on them.

Table recognition usually assumes identified table regions andthe goal is to find the physical structure and the logical structureof the table model [9, 20, 8]. There has been plenty of work deal-ing with machine-printed table recognition [12, 10, 2]. Gatos etal. [7] make use of the complete table rulings to recognize the ta-ble structure while Hirayama [10] relies on dynamic programmingto align table columns. Richarz et al. propose a method of tabularstructure recognition for their semi-supervised transcription systemin handwritten historical weather reports [15]. Making use of thepre-printed table substrate, they use Hough transform to detect thehorizontal and vertical rulings that constitute the tabular structure.Clawson et al. present a projection-profile based method to de-tect and extract handwritten tabular fields from historical censusforms [5]. We notice that these existing techniques are evaluatedon datasets where rulings are usually salient and well displaced,meaning no other lines will distract table analysis. This is, how-ever, not the case for our dataset where we need to handle severelybroken lines and/or false-alarms.

In this work, we address the problem of table detection and de-composition for noisy handwritten documents. These documentsdiffer from most in the literature and they are not collected in acontrolled environment but from the field. As a result, various types

Page 2: Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

Figure 1: A sample document substrate illustrates multi-ple challenges present in the evaluation dataset where text ismasked for display to protect secrete information.

of noise and artifacts are present and the low image quality makesruling lines hard to detect. After separating clutter noise from theimage, our approach is to ensure high recall of table rulings andthen compute the “key points” that intersect horizontal and verticalrulings. Then we employ an optimization procedure to select themost probable subset of rulings that constitute the table. Finally,given the selected key points, we decompose a table image into a2-D arrangement of table cell images.

2. WORKFLOW OVERVIEWThe high-level workflow is shown in Figure 2. In our implemen-

tation, an image is modeled as multi-layer structure where eachlayer consists of one document component. In addition, we ob-serve that the same document substrate is used multiple times in

ruling detection

text detection

ruling selection

table decomposition

clutter separation

Figure 2: The workflow of our table analysis technique.

the dataset. Consider this fact in practice, we can build a tablemodel that summarizes the logical and physical structure and useit to register an unknown input. Hence, table model serves as aninput to help detect and analyze tables.

2.1 Clutter DetectionClutter noise, which refers to the black margin near the image

border, is introduced by imperfect scanning. Agrawal and Doermanpresent a distance transform based approach to detect and removeclutter noise [1]. They detect it using a 2-class SVM classifier inwhich a number of connected-component based features are con-structed.

Our clutter detection, however, is based on the fact that clutternoise is usually much larger than the other components. Thus, atlower resolution, we may only be able to see the clutter noise. Todo this, we scale the image down to 1/4 and extract connected-components from the scaled image. As expected, large componentsare mostly clutter noise in the original image, so we mark theseclutter pixels so that they will not be considered in the followingprocessing operations in another layer. We show the two layers ofimage components in Figure 3(a) and Figure 3(b).

2.2 Ruling DetectionAlthough many of the ruling lines in our evaluation dataset are

pre-printed on a document substrate, they differ from the ones ex-hibit in our previous work [3] in that they may not present consis-tent spacing. In this work, we still use a probabilistic variant ofHough Transform [13] to extract salient line segments. Since manyrulings are broken, small gaps (up to 20 pixels) are allowed duringruling detection. Next, we make use of the fact that most correctrulings are parallel or orthogonal in order to exclude some of thefalse-alarms detected in the text area. Then, we employ the Adap-tive Basic Sequential Algorithmic Scheme (Adaptive BSAS [19])to group clustered line segments based on their ρ values, and com-pute their parameters (slope, intercept, etc.) using standard linefitting. Due to the low image quality, we adjust parameters in theHough Transform to obtain high recall of line segments, which,of course, may result in false-alarms in packed text regions. Fig-ure 3(c) shows the results of clustered horizontal and vertical rul-ings, marked by red and green, respectively.

2.3 Text DetectionSpatial displacement of text can be valuable information to ex-

ploit for tabular structure detection and recognition [11, 3]. In ourcurrent work, text blob are detected by separating the layers of rul-ings and clutter noise and then detecting connected-components.Next, we transform the text blob based on the skew angle andthen exclude those having unexpected aspect ratios (α < 0.1 orα > 10.0). This effectively excludes most of the line segments leftin the current image layer, as shown in Figure 3(d).

3. RULING SELECTIONOut ruling selection algorithm is designed to work on a set of

rulings that are either parallel or orthogonal, which should be sat-isfied by the filtering in Section 2.2. We now adapt our previousidea of key points in this work [4]. Key points are defined to be awhite space region within a local 2× 2 array of text blob in whichany horizontal or vertical cuts will not affect table cells. For ta-bles that are ruled or unruled, key points usually refer to the whitespace between text blob such as words. For ruled tables that haverow/column spans, however, we need to adapt key points to be theintersections of table rulings. Therefore, the problem of ruling se-lection is converted into key point selection. In addition, this con-

Page 3: Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

(a) An image layer containingclutter noise.

(b) An image layer containingthe remaining components.

(c) Detected ruling lines. (d) An image layer containingtext blob (masked for display).

Figure 3: Detection and separation of various document components. Note that secrete text is masked for display but the algorithmworks on original images without redaction.

version also reduces the search space because not every ruling in-tersects another.

As a result, several rulings in the text area isolated from the tableheader are excluded at this stage, as shown in Figure 4(a). Note thatwe have allowed relatively large gaps (up to 100 pixels) betweenrulings due to degraded images in order to obtain high recall of keypoints in the table structure.

Then we employ the Adaptive BSAS clustering again to groupkey points into horizontal and vertical clusters. Each key point is in-dexed by the horizontal and vertical clusters, e.g., (H1,...p, V1,...q),where p, q are the numbers of horizontal and vertical key point clus-ters, respectively.

Next, we formulate key points selection as an optimization prob-lem:

argminW∈Ωf(W,M) (1)

where W is a configuration of selected key points for scoring, Ω isthe set of all configurations, and M is the table model containingthe number of rows r and columns c, and cell dimension infor-mation. We decide the key points for table structure by selectinghorizontal and vertical key points separately. Taking the horizontalcase as an example, the formulation is specified as follows:

argminWh∈Ωhf(Wh,Mh) (2)

with the constraints that ‖Wh‖ = r + 1 and ‖Ωh‖ =(

pr+1

). The

vertical key point selection is similar. The cost function f(·) com-putes a real value indicating how close the current key point clusterconfiguration matches with the table model. In our implementation,we compute is as follows:

f(·) = ‖Wh −Mh‖+ C(Wh)

=

p∑i=1

‖Whi −Mhi‖+ C(Wh) (3)

where Whi and Mhi are heights for table row i, respectively. Inwords, the first term in this equation computes the accumulateddifferences of row heights between selected and the table model.The second term C(Wh) is the cost for text displacement againstthe horizontal rulings. Currently, we only consider the text dis-placement in the table header, i.e., the accumulated cost of bound-ing boxes of text blob adjacent to the second horizontal ruling. Ofcourse, more sophisticated cost functions are possible for futureimprovement.

Since p, q are small numbers, it is feasible to enumerate the con-

(a) Detected key points. (b) Key points for the table.

Figure 4: Intermediate results from ruling selection.

figuration space to find the global optimum that constitutes the tablestructure. Figure 4(b) shows the result of selecting horizontal keypoint clusters, marked by circles. Then we run the selection algo-rithm again using the corresponding vertical key point clusters, andobtain the correct result. Note that for tables with opened sides (leftand right in Figure 4(a)) we need to add imaginary key points ac-cordingly to comply with the constraints in Equation 2. Finally, wescan through the obtained key point grid to decompose the tabularstructure into a 2-D arrangement of table cells, using one 2× 2 tu-ple at a time. To handle row or column spanning, we simply skipcorresponding adjacent key points vertically/horizontally.

4. EXPERIMENTAL SETUP

4.1 Data PreparationOur evaluation involves a noisy Arabic handwritten document

dataset that contains field data [14]. Since this is not prepared incontrolled environment, it presents various kinds of noise, artifacts,and complicated document layouts. A sample document substratepresenting some of the challenges is shown in Figure 1. So far, wehave annotated and evaluated 61 Arabic documents from 17 docu-ment substrates with the corresponding table models (158 imagesin total). Table 1 shows some statistics in this 158-image datasetbased on manual investigation.

The table model specifies the number of rows and columns, rowor column spanning, and approximate cell dimensions. Figure 5shows an example of such a table model file. Note that since the

Page 4: Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

Table 1: Statistics of human perception on the Arabic handwritten table documents.Page-wise Document Characteristics Max Median Min

# of image/model 29 6 1# of vertical rulings 27 5 3

# of open-ended solid horizontal rulings 27 4 0# of open-ended dashed horizontal rulings 24 0 0

# of close-ended horizontal rulings 12 0 0# of headers containing printed Arabic phrases 27 10 4

# of other printed Arabic phrases (e.g., from title) 89 14 0

same table substrate appears many times in the dataset, the celldimensions (heights, widths) are only the average values.

4.2 EvaluationWe evaluate the system by computing precision and recall on

table cell images:

precision =number of correctly detected table cells

total number of detected table cells

recall =number of correctly detected table cells

total number of ground-truthed table cells(4)

Currently, we consider a correct table cell detection as long as thetext information defined by the 2×2 tuple of key points correspondsto the ground-truth.

4.3 ComparisonIn this work, we compare our algorithm with a cross matrix

based method by Shi et al. [18]. The idea is to first compute a2-D matrix of rulings which have intersections with their orthogo-nal rulings, for both the table model and input image (Model[·][·]and Scene[·][·]). Next, considering relatively small number of ver-tical rulings, they enumerate all possible combinations of verticalrulings. In each enumeration, they design a dynamic programming(DP) algorithm to select the optimal horizontal rulings. In this way,they transform the problem of selecting optimal subsets of rulingsto optimally aligning two sets of rulings, which is similar to the editdistance computation between two strings. Finally, the horizontalrulings are computed by back tracking the score matrix in the DPframework. In essence, both methods use similar logic to selecthorizontal/vertical rulings as the tabular structure, but our methodmakes use of cell information in the table model, which is expectedto be more reliable than the cross matrix based approach.

We restate the DP framework here for reference. Denote hor-izontal model rulings as Mi, i ∈ [1, n], where n is the numberof horizontal rulings in the table model. Likewise Sj , j ∈ [1, N ]where N is the number detected in the image. Now the matchingscore of a model ruling and a scene ruling is defined as the Ham-ming distance between two rows in the cross matrices.

C(Mi, Sj) =

K∑k=1

‖Model[i][k]− Scene[j][k]‖ (5)

where k ∈ [1, . . . ,K] denote the index to the current vertical rulingconfiguration.

The definition for the score matrixH resembles the computationof edit distance, as follows:

H[i][j] = min

H[i− 1][j − 1] + Csub(Mi, Sj)H[i− 1][j] + Cdel(Mi)H[i][j − 1] + Cins(Sj)

(6)

where i ∈ [1, n], j ∈ [1, N ], and Cdel(·), Cins(·) are the costfor deletion and insertion in computing the edit distance betweentwo strings. In the task of ruling selection, we need to forbid dele-tions by assigning it a high cost, otherwise deletion may reducethe number of rulings selected after the alignment. Thus, we setCdel(·) = 100 and Cins(j) = C(0, Sj). After computing the scorematrix, we back track from H[n][N ] to obtain the alignment ofmodel and scene rulings, in which the positions of substitution arerecorded as the indices of selected rulings.

5. EXPERIMENTAL RESULTSWe evaluated our algorithm using 61 document from 17 table

classes involving 3,627 table cells and found out that the precisionis 88.60% and the recall is 87.90%. On the other hand, the crossmatrix based method obtained precision and recall of 85.90% and84.20%, respectively. We should clarify that both methods did notundergo sophisticated parameter tuning. Note that the detected keypoints do not have to form a complete 2-D grid since we can com-pute the rest based on the table model.

Our method did outperform the cross matrix based method, how-ever, there are several observations that we would like to point outwhich may be useful for future improvements. First, spacing be-tween adjacent rulings is characteristic information to exploit forselecting the correct rulings. Second, the cross matrix method as-sumes less input information but relies on high recall of line seg-ments, which is challenging in such a degraded dataset, i.e., if theexpected intersection in one ruling is missing, there will be a costin the cross matrix. Lastly, the spatial displacement of text blob canbe useful, especially the text in table headers, because they are usu-ally well spaced by their adjacent rulings, while those in other partof the table or page may lie on rulings and/or in different directions.

An example of failure cases is shown in Figure 6. First, thereare a number of false-alarms detected due to the fact that the textarea is packed, hence false-alarm rulings are detected. On the otherhand, this table model is not precise enough for this specific inputdocument, so the optimization procedure gives us an incorrect so-lution. However, this example does provide us with some hints onfuture improvements. For example, it is reasonable to incorporatenumber of key points in selecting the subsets of key points.

6. CONCLUSIONTables analysis can be a valuable step in document image analy-

sis and over the years, there has been much research on table detec-tion and recognition on carefully prepared datasets. In this paper,we have shown the complexity of processing handwritten docu-ments that are from real life. Our table analysis algorithm is basedon detecting and analyzing salient ruling lines in the documents,which may contain false-alarms including paper folds, pre-printedrulings for handwriting, etc. Our approach is to aim for high-recallof the line segments and then employ Adaptive BSAS clustering

Page 5: Ruling-Based Table Analysis for Noisy Handwritten Documentsjic207/pub/mocr2013.pdf · Ruling-Based Table Analysis for Noisy Handwritten Documents ... 2.1 Clutter Detection Clutter

Figure 5: An example of the table model specification.

(a) (b)

Figure 6: An example of failure cases.

and orthogonal analysis to obtain horizontal and vertical rulings.We then adapt the idea of “key points” and employed an optimiza-tion procedure to select the most probably subset that constitutesthe table structure. Finally, we use the key points and the tablemodel to decompose the table image into a 2-D arrangement of ta-ble cell images. Experimental results involving 61 pages from 17table classes show a table cell precision of 88.60% and a recall of87.90%. As for the future work, we would like to enhance the ro-bustness of our algorithm and design more powerful cost functionsfor the optimization procedure.

AcknowledgementThe authors acknowledge stimulating discussions with George Nagy.This work is supported by a DARPA IPTO grant administered byRaytheon BBN Technologies.

7. REFERENCES[1] M. Agrawal and D. Doermann. Clutter noise removal in

binary document images. In International Conference onDocument Analysis and Recognition, pages 556–560, 2009.

[2] F. Cesarini, S. Marinari, L. Sarti, and G. Soda. Trainabletable location in document images. In InternationalConference on Pattern Recognition, pages 236–240, 2002.

[3] J. Chen and D. Lopresti. A model-based ruling line detectionalgorithm for noisy handwritten documents. In Proceedingsof the 11th International Conference on Document Analysisand Recognition, pages 404–408, September 2011.

[4] J. Chen and D. Lopresti. Model-based tabular structuredetection and recognition in noisy handwritten documents. In

Proceedings of the 13th International Conference onFrontiers in Handwriting Recognition, pages 75–80, 2012.

[5] R. Clawson, K. Bauer, G. Chidester, M. Tyler-Pohontsch,D. Kennard, J. Ryu, and W. Barrett. Automated recognitionand exrtaction of tabular fields for the indexing of censusrecords. In Proc. Document Recognition and Retrieval XVIII(IS&T/SPIE International Symposium on ElectronicImaging), pages 8658–17, 2013.

[6] D. Embley, M. Hurst, D. Lopresti, and G. Nagy.Table-processing paradigms: a research survey. InternationalJournal on Document Analysis and Recognition, 8(2):66–86,2006.

[7] B. Gatos, D. Danatsas, I. Pratikakis, and S. Perantonis.Automatic table detection in document images. InProceedings of the Third International Conference onAdvances in Pattern Recognition, pages 609–618, 2005.

[8] J. Handley. Electronic imaging technology, Chapter 8(Document Recognition). IS&T/SPIE Optical EngineeringPress, 1999.

[9] R. Haralick. Document image understanding: geometric andlogical layout. In Proceedings of the 1994 Computer Visionand Pattern Recognition, pages 385–390, 1994.

[10] Y. Hirayama. A method for table structure analysis using DPmatching. In International Conference on DocumentAnalysis and Recognition, pages 583–586, 1995.

[11] J. Hu, R. Kashi, D. Lopresti, and G. Wilfong.Medium-independent table detection. In DocumentRecognition and Retrieval VIII (IS&T/SPIE ElectronicImaging, pages 44–55, 2001.

[12] A. Laurentini and P. Viada. Identifying and understandingtabular material in compound documents. In InternationalConference on Pattern Recognition, pages 405–409, 1992.

[13] J. Matas, C. Galambos, and J. Kittler. Robust detection oflines using the progressive probabilistic Hough transform.Computer Vision and Image Understanding, 78:119–137,2000.

[14] B. Montgomery. The Iraqi secret police files: A documentaryrecord of the Anfal genocide. Archivaria, pages 81–82, 2001.

[15] J. Richarz, S. Vajda, and G. Fink. Towards semi-supervisedtranscription of handwritten historical weather reports. InProceedings of the 10th International Workshop onDocument Analysis Systems. To appear, 2012.

[16] F. Shafait and R. Smith. Table detection in heterogeneousdocuments. In Proceedings of the 9th IAPR InternationalWorkshop on Document Analysis Systems, pages 65–72,2010.

[17] H. Shamalian, H. Baird, and T. Wood. A retargetable tablereader. In International Conference on Document Analysisand Recognition, pages 158–163, 1997.

[18] Z. Shi, S. Setlur, and V. Govindaraju. Table cell detectionand content extraction from degraded document images. InProc. 12th International Conference on Document Analysisand Recognition. To Appear, 2013.

[19] S. Theodoridis and K. Koutroumbas. Pattern Recognition.Academic Press, 2009.

[20] X. Wang. Tabular abstraction, editing, and formatting. PhDthesis, University of Waterloo, 1996.

[21] R. Zanibbi, D. Blostein, and J. Cordy. A survey of tablerecognition: models, observations, transformations, andinferences. International Journal on Document Analysis andRecognition, 7(1):1–16, 2003.