Nothing Special   »   [go: up one dir, main page]

AU2008260018C1 - Refining text extraction in colour compound documents - Google Patents

Refining text extraction in colour compound documents Download PDF

Info

Publication number
AU2008260018C1
AU2008260018C1 AU2008260018A AU2008260018A AU2008260018C1 AU 2008260018 C1 AU2008260018 C1 AU 2008260018C1 AU 2008260018 A AU2008260018 A AU 2008260018A AU 2008260018 A AU2008260018 A AU 2008260018A AU 2008260018 C1 AU2008260018 C1 AU 2008260018C1
Authority
AU
Australia
Prior art keywords
text
connected component
ccs
graphics
photo
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
AU2008260018A
Other versions
AU2008260018A1 (en
AU2008260018B2 (en
Inventor
Yu-Ling Chen
Ping Liu
Trevor Lee Mcdonell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to AU2008260018A priority Critical patent/AU2008260018C1/en
Priority to US12/637,446 priority patent/US8351691B2/en
Publication of AU2008260018A1 publication Critical patent/AU2008260018A1/en
Application granted granted Critical
Publication of AU2008260018B2 publication Critical patent/AU2008260018B2/en
Publication of AU2008260018C1 publication Critical patent/AU2008260018C1/en
Ceased legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/413Classification of content, e.g. text, photographs or tables

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Description

S&F Ref: 888293 AUSTRALIA PATENTS ACT 1990 COMPLETE SPECIFICATION FOR A STANDARD PATENT Name and Address Canon Kabushiki Kaisha, of 30-2, Shimomaruko 3 of Applicant : chome, Ohta-ku, Tokyo, 146, Japan Actual Inventor(s): Yu-Ling Chen Ping Liu Trevor Lee McDonell Address for Service: Spruson & Ferguson St Martins Tower Level 35 31 Market Street Sydney NSW 2000 (CCN 3710000177) Invention Title: Refining text extraction in colour compound documents The following statement is a full description of this invention, including the best method of performing it known to me/us: 5845c(1898902_1) REFINING TEXT EXTRACTION IN COLOUR COMPOUND DOCUMENTS TECHNICAL FIELD The present invention relates to document layout analysis of complex colour documents for storage and other processing and, in particular, to improving the accuracy of extracting text components from an image of the document. 5 BACKGROUND The proliferation of imaging technology, combined with ever increasing computational processing power, has led to many advances in the area of document analysis systems. These systems may be used to extract semantic information from a scanned document, and the technology is being used in a growing number of 10 applications. One such application of this technology is to the problem of the accurate identification of text and non-text regions of images of a scanned document, particularly those including complex features such as photographs and line graphics. Such content analysis is an important precursor to many document management, synthesis, and display 15 processing systems. Popular applications of content analysis in a document management system include the generation of editable text, via optical character recognition (OCR), and the extraction of photographs and line graphics for later reuse. The advent of powerful document editing tools and the proliferation of inexpensive colour printing systems have enabled the creation of complex colour 20 documents, whose layouts are no longer restricted to the traditional rectangular layout style. Due to unpredictable colour combinations and potentially unrestrained intermingling of text and non-text regions present in such documents, a problem facing 1898437_1 888293_speci lodge -2 document analysis systems is the increase of image reproduction artefacts such as colour bleeding and noise. Such colour pollution may have been caused by printer anti-aliasing, scanner aliasing, halftone estimation, chromatic aberrations, and other blurring, haloing or fringing effects. The increase in layout complexity and colour reproduction errors 5 consequently makes accurate extraction of content from colour documents significantly more challenging than for traditional black and white documents. A basic approach to determining page content operates on a binary (black and white) version of the input scan image of the document, in order to avoid colour complexity. The conventional methods for text extraction in black and white documents 10 are considered mature, and may still produce acceptable results for documents with limited colour combinations and rectangular layout styles. However, those methods do not address the problems associated with image reproduction errors or complex colour combinations. This binarisation approach fundamentally relies on applying a threshold to the input image, and may fail to correctly distinguish between foreground regions such 15 as text, and the surrounding background pixels. An alternative method is to process the colour document directly, often utilising a quantised colour version of the input to reduce processing complexity and remove several reproduction artefacts, such as halftone colours. Additionally, a method utilising colour quantisation can often correctly distinguish between colours that binarisation fails 20 to separate. However, colour quantisation in and of itself cannot solve colour misrepresentation errors present in the source document such as colour bleeding, which may be retained by the quantisation process. This could result in information bearing objects that have colour fringing effects, or that have been quantised into multiple colours instead of remaining as flat-filled objects. Text decoration effects such as 1898437_1 888293_speci_lodge -3 outline and shadow will also be retained by the colour segmentation process, which pose challenges for character recognition methods if both the text and decoration component are to be extracted. There is a need to accurately identify and categorise the information bearing 5 objects contained within mixed content document images possessing complex colour layouts. SUMMARY Disclosed is a connected component (CC) based text extraction method that includes refining text CC classification using global colour context statistics, grouping 10 text CCs into text blocks, recovering misclassified non-text CCs into a nearby text block and removing erroneous CCs from each text block using local colour context statistics. In accordance with one aspect of the present disclosure, there is a computer implemented method of text extraction in colour compound documents, said method comprising the steps of: 15 (a) connecting similarly coloured pixels of an image of a colour compound document into connected components (CCs); (b) classifying each CC as either text or non-text; (c) refining text CC classification for each text CC using global colour context statistics; 20 (d) grouping text CCs into text blocks; (e) recovering misclassified non-text CCs into a nearby text block; and (f) removing extraneous CCs from each text block using local colour context statistics to thereby provide the extracted text in the text blocks. Other aspects are also disclosed. 1898437_1 888293_speci-lodge -3A An example of such another aspect is a computer implemented method of identifying at least one of a photo region and a graphics region in a colour bitmap image of a compound document, said method comprising: 5 (a) connecting similarly coloured pixels of said image into connected components and placing the connected components in an enclosure tree, the enclosure tree forming a hierarchy of ancestor and descendent connected components where an ancestor connected component encloses a descendent connected component wherein each node of the enclosure tree is a connected component; 10 (b) classifying each connected component into one of a plurality of classes wherein at least one class represents non-text connected components; (c) selecting a connected component from one of the non-text connected components of the enclosure tree; (d) calculating context statistics for the selected connected component based 15 on descendent and touching sibling connected components of the selected connected component; (e) identifying whether the selected connected component contains at least one of a photo region and a graphics region based on the context statistics; and (f) storing the selected connected component identified as either graphics or 20 photo object to memory.
-4 BRIEF DESCRIPTION OF THE DRAWINGS At least one embodiment of the present invention will now be described with reference to the following drawings, in which: Fig. 1 is an application environment of a document analysis module in which the 5 arrangements to be described can be used; Figs. 2A and 2B form a schematic block diagram of a general purpose computer system upon which the arrangements described can be practiced; Fig. 2C is a block diagram of the constituent software architecture of a document 10 analysis apparatus; Fig. 3 is a processing flow of the document analysis module; Fig. 4 is an example input page with photo, graphics and text groups; Fig. 5 is an example input page with a number of objects occluding the background and other objects; 15 Fig. 6 is a representation of an enclosure tree which is generated by applying a CC generation process to Fig. 5; Fig. 7 is a processing flow diagram showing the traversal of the CC enclosure tree for CC classification; Fig. 8 is a detailed flowchart of the CC classification process in Fig. 7; 20 Fig. 9A is an example of a bad text container; Fig. 9B is an example of a good text container; Fig. 10 is a processing flow of the graphics extraction process in Fig. 2; Fig. 11 is a processing flow for each CC processed by the graphics extraction process; 25 Fig. 12 is an example of a graphics object with one child CC; 1898437_1 888293_specilodge -5 Fig. 13 is an example of a graphics object with multiple child CCs; Fig. 14 is an example of a graphics object with no child CCs and one touching CC; Fig. 15 is a processing flow to find the graphics container CC for graphics 5 processing; Fig. 16 is a processing flow to find the statistics of children of the graphics container CC; Fig. 17 is a processing flow of the graphics grouping process; Fig. 18 is an example of the extracted graphics region from the snowman 10 graphics in Fig. 4; Fig. 19 is the processing flow of the photo object extraction process in Fig. 3; Fig. 20 is a flowchart of the photo seeding process in Fig. 19; Fig. 21 is a flowchart of the photo region growing process in Fig. 19; Fig. 22 is a flowchart of the photo region consolidation process in Fig. 19; 15 Fig. 23 is a flowchart of the text extraction process; Fig. 24 is a flowchart of refining TEXT CC classification in Fig. 23. Fig. 25 is a flowchart of calculating text colour likelihood for each colour bucket in Fig. 24; Fig. 26 is a flowchart of grouping TEXT CCs into logical text blocks in Fig. 23; 20 Fig. 27 is a flowchart of recovering misclassified CCs into existing text blocks in Fig. 23; Fig. 28 is a flowchart of removing misclassified TEXT CCs from text blocks in Fig. 23; Fig. 29 is a flowchart of removing misclassified TEXT CCs from a text block; 1898437_1 888293_specilodge -6 Fig. 30A-30C are examples demonstrating the TEXT CC classification refinement process; Fig. 31A-31C are examples demonstrating how misclassified non-text CCs are recovered into a text block; 5 Fig. 32A-32B are examples showing fragmented TEXT CCs before and after merging; and Fig. 33A-33B are examples demonstrating how misclassified TEXT CCs are removed from a text block. DETAILED DESCRIPTION INCLUDING BEST MODE 10 Fig. I shows an application environment 100 of a document analysis system. A hard copy document 102 is scanned by a scanner 104 to form an input scanned document image 106, which is often simply called the "input document". The scanned input document 106 is then subject to document analysis in a module 110 which extracts from the image portions or objects that may be independently further processed. The 15 extracted objects may be output from the module 110 as a logical tree 115. Areas of a scanned document image can be compressed more efficiently using an algorithm suitable for the image content in each of those areas. The extracted objects from a document can be compressed separately each using an algorithm suitable for the extract object type. The extracted graphics objects can be vectorised (130) for re-use or compressed (120). 20 Extracted photo objects can be cropped (140) for image retrieval and extracted text can be processed by OCR (150) to produce a searchable or an editable document. Figs. 2A and 2B collectively form a schematic block diagram of a general purpose computer system 200, upon which the various arrangements described can be practiced. 1898437_1 888293_speci_lodge -7 As seen in Fig. 2A, the computer system 200 is formed by a computer module 201, input devices such as a keyboard 202, a mouse pointer device 203, the scanner 104, a camera 227, and a microphone 280, and output devices including a printer 215, a display device 214 and loudspeakers 217. An external Modulator 5 Demodulator (Modem) transceiver device 216 may be used by the computer module 201 for communicating to and from a communications network 220 via a connection 221. The network 220 may be a wide-area network (WAN), such as the Internet or a private WAN. Where the connection 221 is a telephone line, the modem 216 may be a traditional "dial-up" modem. Alternatively, where the connection 221 is a high capacity 10 (eg: cable) connection, the modem 216 may be a broadband modem. A wireless modem may also be used for wireless connection to the network 220. The computer module 201 typically includes at least one processor unit 205, and a memory unit 206 for example formed from semiconductor random access memory (RAM) and semiconductor read only memory (ROM). The module 201 also includes an 15 number of input/output (I/0) interfaces including an audio-video interface 207 that couples to the video display 214, loudspeakers 217 and microphone 280, an 1/0 interface 213 for the keyboard 202, mouse 203, scanner 104, camera 227 and optionally a joystick (not illustrated), and an interface 208 for the external modem 216 and printer 215. In some implementations, the modem 216 may be incorporated within the 20 computer module 201, for example within the interface 208. The computer module 201 also has a local network interface 211 which, via a connection 223, permits coupling of the computer system 200 to a local computer network 222, known as a Local Area Network (LAN). As also illustrated, the local network 222 may also couple to the wide network 220 via a connection 224, which would typically include a so-called "firewall" 1898437_1 888293_specilodge -8 device or device of similar functionality. The interface 211 may be formed by an EthernetTM circuit card, a Bluetooth T m wireless arrangement or an IEEE 802.11 wireless arrangement. The interfaces 208 and 213 may afford either or both of serial and parallel 5 connectivity, the former typically being implemented according to the Universal Serial Bus (USB) standards and having corresponding USB connectors (not illustrated). Storage devices 209 are provided and typically include a hard disk drive (HDD) 210. Other storage devices such as a floppy disk drive and a magnetic tape drive (not illustrated) may also be used. An optical disk drive 212 is typically provided to act as a 10 non-volatile source of data. Portable memory devices, such optical disks (eg: CD-ROM, DVD), USB-RAM, and floppy disks for example may then be used as appropriate sources of data to the system 200. The components 205 to 213 of the computer module 201 typically communicate via an interconnected bus 204 and in a manner which results in a conventional mode of 15 operation of the computer system 200 known to those in the relevant art. Examples of computers on which the described arrangements can be practised include IBM-PC's and compatibles, Sun Sparcstations, Apple MacTM or alike computer systems evolved therefrom. The methods of document layout analysis described herein may be implemented 20 using the computer system 200 wherein the processes of Figs. 2C to 33B to be described, may be implemented as one or more software application programs 233 executable within the computer system 200. In particular, the steps of the methods of document analysis are effected by instructions 231 in the software 233 that are carried out within the computer system 200. The software instructions 231 may be formed as 1898437_1 888293_speci_lodge -9 one or more code modules, each for performing one or more particular tasks. The software may also be divided into two separate parts, in which a first part and the corresponding code modules performs the document analysis methods and a second part and the corresponding code modules manage a user interface between the first part and 5 the user. The software 233 is generally loaded into the computer system 200 from a computer readable medium, and is then typically stored in the HDD 210, as illustrated in Fig. 2A, or the memory 206, after which the software 233 can be executed by the computer system 200. In some instances, the application programs 233 may be supplied 10 to the user encoded on one or more CD-ROM 225 and read via the corresponding drive 212 prior to storage in the memory 210 or 206. Alternatively the software 233 may be read by the computer system 200 from the networks 220 or 222 or loaded into the computer system 200 from other computer readable media. Computer readable storage media refers to any storage medium that participates in providing instructions and/or data 15 to the computer system 200 for execution and/or processing. Examples of such storage media include floppy disks, magnetic tape, CD-ROM, a hard disk drive, a ROM or integrated circuit, USB memory, a magneto-optical disk, or a computer readable card such as a PCMCIA card and the like, whether or not such devices are internal or external of the computer module 201. Examples of computer readable transmission media that 20 may also participate in the provision of software, application programs, instructions and/or data to the computer module 201 include radio or infra-red transmission channels as well as a network connection to another computer or networked device, and the Internet or Intranets including e-mail transmissions and information recorded on Websites and the like. 1898437_1 888293_speci_lodge - 10 The second part of the application programs 233 and the corresponding code modules mentioned above may be executed to implement one or more graphical user interfaces (GUIs) to be rendered or otherwise represented upon the display 214. Through manipulation of typically the keyboard 202 and the mouse 203, a user of the computer 5 system 200 and the application may manipulate the interface in a functionally adaptable manner to provide controlling commands and/or input to the applications associated with the GUI(s). Other forms of functionally adaptable user interfaces may also be implemented, such as an audio interface utilizing speech prompts output via the loudspeakers 217 and user voice commands input via the microphone 280. 10 Fig. 2B is a detailed schematic block diagram of the processor 205 and a "memory" 234. The memory 234 represents a logical aggregation of all the memory devices (including the HDD 210 and semiconductor memory 206) that can be accessed by the computer module 201 in Fig. 2A. When the computer module 201 is initially powered up, a power-on self-test 15 (POST) program 250 executes. The POST program 250 is typically stored in a ROM 249 of the semiconductor memory 206. A program permanently stored in a hardware device such as the ROM 249 is sometimes referred to as firmware. The POST program 250 examines hardware within the computer module 201 to ensure proper functioning, and typically checks the processor 205, the memory (209, 206), and a basic 20 input-output systems software (BIOS) module 251, also typically stored in the ROM 249, for correct operation. Once the POST program 250 has run successfully, the BIOS 251 activates the hard disk drive 210. Activation of the hard disk drive 210 causes a bootstrap loader program 252 that is resident on the hard disk drive 210 to execute via the processor 205. This loads an operating system 253 into the RAM memory 206 upon 1898437_1 888293_speci_lodge - 11 which the operating system 253 commences operation. The operating system 253 is a system level application, executable by the processor 205, to fulfil various high level functions, including processor management, memory management, device management, storage management, software application interface, and generic user interface. 5 The operating system 253 manages the memory (209, 206) in order to ensure that each process or application running on the computer module 201 has sufficient memory in which to execute without colliding with memory allocated to another process. Furthermore, the different types of memory available in the system 200 must be used properly so that each process can run effectively. Accordingly, the aggregated 10 memory 234 is not intended to illustrate how particular segments of memory are allocated (unless otherwise stated), but rather to provide a general view of the memory accessible by the computer system 200 and how such is used. The processor 205 includes a number of functional modules including a control unit 239, an arithmetic logic unit (ALU) 240, and a local or internal memory 248, 15 sometimes called a cache memory. The cache memory 248 typically includes a number of storage registers 244 - 246 in a register section. One or more internal buses 241 functionally interconnect these functional modules. The processor 205 typically also has one or more interfaces 242 for communicating with external devices via the system bus 204, using a connection 218. 20 The application program 233 includes a sequence of instructions 231 that may include conditional branch and loop instructions. The program 233 may also include data 232 which is used in execution of the program 233. The instructions 231 and the data 232 are stored in memory locations 228-230 and 235-237 respectively. Depending upon the relative size of the instructions 231 and the memory locations 228-230, a 1898437_1 888293_specilodge - 12 particular instruction may be stored in a single memory location as depicted by the instruction shown in the memory location 230. Alternately, an instruction may be segmented into a number of parts each of which is stored in a separate memory location, as depicted by the instruction segments shown in the memory locations 228-229. 5 In general, the processor 205 is given a set of instructions which are executed therein. The processor 205 then waits for a subsequent input, to which it reacts to by executing another set of instructions. Each input may be provided from one or more of a number of sources, including data generated by one or more of the input devices 202, 203, data received from an external source across one of the 10 networks 220, 222, data retrieved from one of the storage devices 206, 209 or data retrieved from a storage medium 225 inserted into the corresponding reader 212. The execution of a set of the instructions may in some cases result in output of data. Execution may also involve storing data or variables to the memory 234. The disclosed document analysis arrangements use input variables 254, that are 15 stored in the memory 234 in corresponding memory locations 255-258. The document analysis arrangements produce output variables 261, that are stored in the memory 234 in corresponding memory locations 262-265. Intermediate variables may be stored in memory locations 259, 260, 266 and 267. The register section 244-246, the arithmetic logic unit (ALU) 240, and the 20 control unit 239 of the processor 205 work together to perform sequences of micro operations needed to perform "fetch, decode, and execute" cycles for every instruction in the instruction set making up the program 233. Each fetch, decode, and execute cycle comprises: 1898437_1 888293_specilodge - 13 (a) a fetch operation, which fetches or reads an instruction 231 from a memory location 228; (b) a decode operation in which the control unit 239 determines which instruction has been fetched; and 5 (c) an execute operation in which the control unit 239 and/or the ALU 240 execute the instruction. Thereafter, a further fetch, decode, and execute cycle for the next instruction may be executed. Similarly, a store cycle may be performed by which the control unit 239 stores or writes a value to a memory location 232. 10 Each step or sub-process in the processes of Figs. 2C to 33B is associated with one or more segments of the program 233, and is performed by the register section 244 247, the ALU 240, and the control unit 239 in the processor 205 working together to perform the fetch, decode, and execute cycles for every instruction in the instruction set for the noted segments of the program 233. 15 Fig. 2C shows schematically the exemplary software architecture of the document analysis module 110 which is seen to include the processor, or "computational processing unit (CPU) 205, the memory module 234, and a number of software code modules, particularly a CC generation module 274, a CC classification module 270, a graphics extraction module 280, a photo extraction module 276 and a text extraction 20 module 290. The modules seen in Fig. 2C are operationally, if not physically, coupled together via the common bus 204. The document analysis module 110 is illustrated in Fig. 2C using a bus architecture diagram. Other types of configuration may be used to implement one or more of the components. In the present arrangement, a computer program is stored in the memory 234 and is executable by in the CPU 205 to implement 1898437_1 888293_specilodge -14 the functionality of each of the CC generation module 274, the CC classification module 270, the graphics extraction module 280, the photo extraction module 276 and the text extraction module 290. A further module 295, also implementable as part of the computer program, is operable to combine the extracted objects to output the logical tree 5 of extracted objects 115. An image of the document 102 as scanned by the scanner 104 is stored in the memory 234 whereupon the image is acted upon by the modules which in turn use the memory 234 for intermediate storage of image components and text extracted or otherwise processed from the scanned image. Fig. 3 is a processing flow of a document analysis method 300 that is performed 10 by the aforementioned computer program of the document analysis module 110. The processes 320, 330, 340, 350, and 360 of the method 300 of Fig. 3 correspond to the modules 274, 270, 280, 276, and 290 in Fig. 2C respectively. Process 320 generates CCs from a scanned input image 106 and places the generated CCs in an enclosure tree. Process 330 classifies the generated CCs into different types according to the 15 characteristics of each CC. Process 340 finds or identifies those CCs with salient graphics features and then performs semantic analysis on them to extract graphics objects. Process 350 finds photo seeds from CCs with high photo probability and then checks the semantic relationship in the enclosure tree from the neighbourhood of the photo seeds to extract photo objects. Process 360 groups CCs that remain as text after 20 graphics and photo extraction and refines those grouped CCs to form text lines or paragraphs. Each of these processes will be explained in more detail below. Each of the processes 320-360 operates to group CCs with similar properties to form semantic objects. There are three kinds of groups mentioned in this description text groups, photo groups and graphics groups. A text group represents a text line or 1898437_1 888293_specilodge - 15 paragraph on a page. A graphics group represents a graphics object on a page. A photo group represents an isolated photographic region enclosed by the main background. A group is a logical entity that is represented in the memory 234 as a stored list of all CCs that make up the group. 5 Fig. 4 shows a sample page 400 which may exemplify an input scanned image 106. The sample page 400 has a main background 410 and groups 420, 430, 440, 450, 460, and 470. Each of the groups 420, 430, 440, 450, 460, 470 is made up of multiple CCs and corresponds to a semantic object on the input page 400. The CCs belonging to the same group are grouped together and are maintained as a list in the memory 234. 10 Each of the groups 420 and 430 corresponds to a photo group. The group 440 (a snowman) consists of an outline part 441 and three interior parts 442, 443 and 444 which are grouped together to represent a graphics object. The group 460 is a text group consisting of a line of text 461, and the group 470 is a text group consisting of lines of text 471 and 472. The group 450 is a text group which overlaps with the interior part 15 444 and thus the group 440. The detail of each process 320-360 will be explained further in the following sections. CC Generation Process 320 builds an enclosure tree of CCs. CCs represent a group of connected 20 pixels with a perceptually uniform colour. There are many different ways of generating CCs from an image. CC generation methods are well known in the art. One method is to divide the colour space into a small number of compartments and place all the colours on the document image 106 into one of these compartments by determining the colour similarity. This forms a colour quantised document as a digital image with limited 1898437_1 888293_specilodge -16 colours. A connected component is formed by contiguous pixels of the same colour on the colour quantised document image. Fig. 5 shows a sample page 500 and the constructed enclosure tree for this page is shown in Fig. 6. In Fig. 5, the sample input page 500 includes a main page background 510 with 5 other objects of interest. Areas 520 and 521 are background regions having different colours from each other, as depicted by the different shading thereof. Each of the letters "D" 550, "E" 551 and "F" 552 are placed on top of these solid background areas 520 and 521 and have different colours from the areas 520 and 521. Letter "E" 551 occludes parts of both background areas 520 and 521, while letters "D" 550 and "F" 552 only 10 occlude parts of the areas 520 and 521 respectively. Additionally, the letter "D" 550 also results in a background segment 530, which is cut or otherwise isolated from a remainder of the background 520. Fig. 6 shows an example of an enclosure tree 600 produced by the CC analysis that is applied to the page 500. A root 605 of the enclosure tree, Root CC, is a virtual 15 node that contains the main background 510 of the page 610. Nodes of the tree 600 are linked using two kinds of link: an enclosing relationship link, such as the link 680 shown by a solid arrow, and a touching relationship link, such as the link 690 shown by a dotted line. An object on the page is enclosed by a second object when there is a closed path of pixels from the second object that surrounds the first object. In the example page 500 in 20 Fig. 5, the letter "D" 550 is enclosed by background region 520, but letter "E" 551 is not enclosed by either background region 520 or 521, as there is no surrounding closed pixel path from either of these background regions. The letter "E" is enclosed by the page background object 510. As such, the enclosure information produces a tree as described further below. Background regions 520 and 521 produce nodes 620 and 621 respectively 1898437_1 888293_specilodge - 17 in the tree 600, which are each children of the page background node 610. Letters "D" 550, "E" 551, and "F" 552 produce nodes 650, 651, and 652 respectively. The region 530 inside of letter "D" 550 produces the node 630 which is a child of the node 650, containing the letter "D". Along with the enclosing information, some nodes include 5 touching information. Touching link 690 between the pair of nodes 620 and 621 corresponds to the touching of the pair objects 520 and 521. Other touching links in the graph 600 correspond to the touching relationships between the objects on the page. CCs with the same parent node are siblings. A CC may touch its siblings, children, or parent. Most CCs touch their parent because there is a closed path of pixels 10 that enclose the child. This is seen in Fig. 6 where the node 651 touches two siblings 620 and 621, but does not touch its parent node 610. In order to correctly form an enclosure tree, care has to be taken with how pixels are connected to form CCs. If pixels are connected 8-ways, it is possible to construct two CCs which enclose each other, destroying the tree property of the enclosure tree. Connecting pixels 4-ways or 6-ways 15 does not have this problem and allows a proper enclosure tree to be formed. When a CC is generated and placed in the enclosure tree, the following properties are also derived: CC Properties Definitions Bounding Box (BB) The minimum axis-aligned rectangle that completely circumscribes the CC Bounding Box Area The area of the Bounding Box rectangle Aspect Ratio (AR) The ratio of the Bounding Box width to Bounding Box height Self Pixel Count The number of pixels that make up the CC Total Pixel Count The sum of Self Pixel Count of the CC and all its enclosed 1898437_1 888293_specilodge - 18 descendants Coverage The ratio of Self Pixel Count to Bounding Box Area of the CC Pixel Run Count The number of Pixel Runs that make up the CC wherein a Pixel Run is a set of consecutive pixels in the same row Edge Ratio The ratio of the number of perimeter pixels of a CC to Self Pixel Count Number of Touching The number of touching siblings Number of Enclosed The number of enclosed CCs Touching List A list of touching siblings Enclosing list A list of enclosed child CCs Descendent Count The number of CCs included in the sub-tree Noise Pixel Count The number of salt and pepper noise pixels in the CC Enclosed Pixel Count Total Pixel Count - Self Pixel Count CC Classification The process 330 classifies the CCs in the enclosure tree into different types based on each CC's properties. After each CC is assigned an initial classification, more 5 properties can be deduced from the context created by the semantic relationship of the enclosure tree. These additional properties may then be used to optimise the CC classification, and some CCs may have their classifications changed as a result. Using the properties of each CC, a CC is classified into one of the following types: 10 NOISE Small element without meaning TEXT Part of a text character FRAME Thin and may be part of a table frame 1898437_1 888293_spedlodge -19 TB_CELL Enclosed by a FRAME CC and may be a table cell G_FILL Part of the inside solid region of a graphics object G_OUTLINE Part of the outline of a graphics object PHOTO Part of a photo region BG Part of a solid or flat background region LINE Part of a line object All thresholds and checks mentioned below are optimised for documents scanned at a resolution of 300 dots per inch (DPI) and may be scaled as necessary for documents scanned at a different resolution. In the following sections, non-text (NT) refers to 5 classification types - TB_CELL, PHOTO, GFILL, GOUTLINE, FRAME, LINE and BG. Colour similarity referred to in this description is the visual similarity of two colours. The description hereafter assumes two colours are similar if the Euclidean distance between them is less than a predetermined threshold (eg. 40 when the colour space is YCbCr, with each colour channel normalised to the range of 0 to 255). 10 The CC enclosure tree is traversed in a depth first order and parent CCs are classified before their children. Fig. 7 is a flowchart of the process 330 showing the order in which CCs in the enclosure tree are classified. Traversal of the tree 600 starts from the root CC 705 of the page. Step 710 checks if the CC has any children. If the CC has children then the first child is retrieved at step 720 and is then sent to the classifier at 15 step 730. If the CC does not have children as determined at step 710, then a new CC for classification is found at step 750 by checking for unprocessed siblings. If no unprocessed siblings are found in step750, then a new CC for classification is found by traversing to the parent of the CC at step 770. Where the parent is the root CC, as determined at step 780, the process ends at step 790. If not the root CC, the parent node 1898437_1 888293_specilodge - 20 is then checked again at step 750 to determine if the parent has any unprocessed siblings. This process is repeated until an unprocessed sibling is found. Where unprocessed siblings are found at step 750, step 760 follows and the next unprocessed sibling is retrieved for classification in step 730. The flow proceeds in depth first order, the 5 children of a CC are classified before its siblings, except when the classifier of step 730 classifies the CC as NOISE or TEXT, which is tested in step 740. In the case when the classifier classifies the CC as TEXT or NOISE, then the next CC to be classified is found by looking for an unprocessed sibling at step 750. The classification process of step 730 does not classify children of TEXT or 10 NOISE CCs, whereas children of non-text CCs are classified, as is evident from step 740. This is based on the observation that text is unlikely to be a container for other page elements. Referring back to Fig. 6, showing the CC enclosure tree for the page of Fig. 5, the order in which the nodes are processed is 610, 620, 650, 651, 621, and 652. The classification process 730 classifies the CC based on the coverage, aspect 15 ratio, height and width, and colour of the touching siblings and parent. The classification process labels all the CCs on the CC enclosure tree and this forms a basis for further analysis in the later stages. The output of the classification process 730 is cascaded through the document analysis pipeline to refine the accuracy of classification. Fig. 8 is a flowchart for the CC classification process 730 which operates on a 20 current CC 805. In step 810, a CC is labelled as NOISE 812 if the SelfPixel Count property of the CC, as discussed above, is less than a threshold (eg. 4), or if the CC contains too many noise pixels (eg. Noise Pixel Count greater than 25% of Self Pixel Count). 1898437_1 888293_speci_lodge -21 In step 820, a CC is labelled as LINE 822 if all of the following conditions are satisfied: (i) Aspect Ratio is greater than a threshold (eg. 10), indicating a horizontal line, or is less than a threshold (eg. 0.1), indicating a vertical line; 5 (ii) the CC does not have any children (because a line does not contain/enclose other regions); (iii) Width or height of the Bounding Box of the CC is greater than a threshold (eg. 100, any smaller than this might not be a line but a character like '-'or 'j'); In step 830, a CC is labelled as TBCELL 832 if all of the following conditions 10 are satisfied: (i) The CC is enclosed by a CC classified as FRAME; (ii) Does not have too many siblings (eg. less than a predetermined number, such as 20); (iii) Edge Ratio is lower than a threshold (eg. 0.7); 15 (iv) The CC does not have any touching siblings, or the CC is bigger than any of the CC's touching siblings which overlaps the CC's Bounding Box by more than 80%. In step 840, a CC is labelled as TEXT 842 if all of the following conditions are satisfied: (i) Both the dimensions of the Bounding Box of the CC are less than a 20 threshold (eg. 320 pixels); (ii) The depth of the sub-tree of the CC is less than a threshold (eg. 4); (iii) Descendent Count is less than a threshold (eg. 15); (iv) Coverage is greater than a threshold (eg. 14%); 1898437_1 888293_specilodge - 22 (v) Pixel Run Count is within a predefined range derived from BB height, AR, and the larger of BB dimension. In step 850, a CC is labelled as a FRAME 852 if all of the following conditions are satisfied: 5 (i) The Self Pixel Count of the CC is greater than a threshold (eg. 256); (ii) The ratio of the Self Pixel Count of the CC to the Bounding Box Area of the CC is less than a threshold (eg. 0.25). (iii) Edge Ratio is greater than a threshold (eg. 0.3) In step 860, a CC is labelled G_FILL 862 if the CC does not have any 10 descendants. In step 870, a CC not classified in steps 810-860 is labelled as BG. After all the CCs have been classified, the classification process revisits the CCs labelled as FRAME and examines whether the CC encloses a table structure. A FRAME CC is determined to hold a table structure, if any of the two conditions below is met: 15 (i) The majority (eg. more than 90% of Number of Enclosed of the FRAME CC) of the children of the FRAME CC are rectangular (eg. the ratio of Total Pixel Count of the child CC to the Bounding Box Area of the child CC is greater than a threshold (eg. 0.9); (ii) The ratio of the sum of Total Pixel Count of all the child CCs to Enclosed 20 Pixel Count of the FRAME CC is greater than a threshold (eg. 0.9). If none of the above conditions is satisfied then the classification of the CC is changed from FRAME to GOUTLINE. The CC classifier 730 additionally creates a text profile histogram for each of the non-text CCs in the enclosure tree. The text profile and text histogram is calculated for 1898437_1 888293_speci_lodge - 23 each of the non-text CCs with TEXT CC children. The histogram is generated by first finding the mean pixel count of the enclosed TEXT classified CCs, which is the sum of the Self Pixel Count of all the TEXT children divided by the total number of TEXT children. i.e.: 5 Mean, m = (Self Pixel Count of TEXT CC) / n, n where n is the total number of TEXT CCs. The TEXT children CCs of a non-text CC are categorised into three bins or groups, small, average and large, based on their Self Pixel Count. The size classifications for the text size histogram bins are (a) less than half of the sum of Self 10 Pixel Count of the smallest TEXT CC and the average pixel count for the small bin; (b) greater than half the sum of Self Pixel Count of the largest TEXT CC and the average pixel count for the large bin; and (c) all TEXT CCs which are not placed in either the small bin or large bin are placed into the average bin. Along with the count of TEXT CCs in the histogram bins, other measures of statistical interest and importance include; 15 (a) the mean of Self Pixel Count of TEXT children; and (b) the variance of Self Pixel Count of the TEXT children of a given container. The Coverage of the TEXT CCs of a non-text CC are also categorised into three bins - low, medium and high. The pixel coverage classifications for the three coverage histogram bins are (a) less than 40% for the low bin; (b) greater 80% for the high bin; 20 and (c) all TEXT CCs which are not placed in either the low bin or high bin are placed into the medium bin. Text characters that constitute a text line or paragraph normally have similar characteristics. These characters are typically enclosed by a common background region. 1898437_1 888293_specilodge -24 As a result, analysing the profiled statistics of TEXT classified children can help to determine whether the non-text CC is a good text container. Figs. 9A and 9B show examples of how the profiled statistics of TEXT classified children is used to determine a good text container. In Fig. 9A a CC 900 is a non-text 5 CC that encloses other TEXT children CCs - 901, 902, 903, 904, 905, 906, and 907. Similarly the non-text CC 912 in Fig. 9B encloses some TEXT children CCs. A good text container (i.e. containing actual text) contains TEXT classified children with the following characteristics: - Number of TEXT classified children is greater than a threshold. (eg. 15; When 10 there is not enough samples, the result is not reliable.) - The mean of Self Pixel Count of TEXT children is greater than a threshold. (eg. 40) - The variance of Self Pixel Count of the TEXT children is less than a threshold (eg. mean * mean * 3). 15 Fig. 9B is therefore an example of a good text container and Fig. 9A is an example of a bad text container. Only a non-text CC can contain TEXT CCs. Children of a TEXT CC are considered to be the cut-off parts of background. The profiled statistics of TEXT classified children for each non-text CC is further used in the photo extraction process and will be explained further in the photo extraction 20 section. Graphics Extraction Fig. 10 shows the flowchart of the graphics extraction process 340 from a CC enclosure tree 600. 1898437_1 888293_specilodge - 25 The CC tree 600 is traversed in bottom up, children first order and each CC is processed for graphics extraction. The process 340 commences at step 1010 at the root CC of the page image. If the CC has children, as determined in step 1020, then the first child CC is considered in step 1030. These two steps then continue until a CC without 5 any children is found. This CC is then processed in step 1040 for graphics extraction. A new CC is then found by checking in step 1050if the current CC has an unprocessed sibling. If such a sibling exists, then that sibling is found or retrieved in step 1070 by traversing the tree 600 in memory 234. The process 340 then returns to step 1020 where that CC is again checked for children, and any children found are processed first. If the 10 CC has no unprocessed siblings (step1050), then the parent CC is retrieved in step 1060. If the retrieved parent CC is the root CC as tested in step 1080, then processing stops at step 1090. Otherwise the parent CC is processed at step 1040, as all of its children have already been processed. Graphics objects are normally made of a collection of different shapes and lines. 15 The most salient component in a graphics object is the outline of each shape because of its unique thin and enclosing features. These are generally the outline which is used to define the graphics object and which is traditionally thin so that the outline doesn't impede upon any fill (which may be a bitmap). As such the outline has low pixel density. Returning to Fig. 4, an example of a graphics object is shown which consists of 20 a collection of CCs. The outline part 441 of the graphics object 440 has the most salient graphics feature. The CC with the most salient graphics feature is therefore a good starting point for locating a graphics object. Fig. 11 shows the flowchart of the per CC graphics extraction in process 1040. The process 1040 aims to find CCs with salient 1898437_1 888293_specilodge - 26 graphics feature and to then examine the semantic context in the neighbourhood in order to locate graphics objects. The process 1040 operates on the CC currently being processed, named currentcc 1110. Initially step 1120 tests the currentcc 1110 and if such is not 5 classified as GOUTLINE, then currentcc is ignored and no graphics extraction processing is done on this CC, and the process 1040 terminates at step 1130. If currentcc is identified at step 1120 as classified as GOUTLINE, then the currentcc 1110 is likely to be a CC with a salient graphics feature. Further checking is performed on currentcc to check whether currentcc is worth the processing time for semantic 10 feature extraction. A CC which is worth semantic feature extraction is called a graphics container. Process 1140 performs this checking to find a graphics container from currentcc. The graphics container may be current_cc itself, or a child CC, or a touching sibling CC. If a graphics container is not found, as tested in step 1150, then currentcc is reclassified as either NOISE or BG in step 1160, at which point the 15 process 1040 terminates for the currentcc 1110. Step 1160 reclassifies current_cc 1110 as BG if currentcc is a good text container. Otherwise currentcc is reclassified in step 1160 as NOISE. If a graphics container CC is found in step 150, then the children of the graphics container CC are profiled to check for graphics characteristics by a process 1170. Data 20 gathered from this profiling step 1170 is stored in the memory 234 and used in a following process 1180 to determine how to group the CCs in the neighbourhood of the graphics container CC into a graphics group, namely, the children of the graphics container CC and the touching sibling CCs of currentcc. If a group is created, the group is stored in the memory 234. 1898437_1 888293_specilodge - 27 Fig. 12 shows a thin frame CC 1210 which is classified as GOUTLINE by the CC classification process 330. The frame CC 1210 includes a CC 1220 which is the interior of the frame, and is a child of the frame CC 1210. Note that the interior CC 1220 is only slightly smaller than the frame CC 1210. The interior CC 1220 has one 5 child, CC 1230, having the shape of a star. Fig. 13 shows a thin frame CC 1310 which is classified as GOUTLINE by the CC classification process 330. The frame CC 1310 contains multiple child CCs 1320, 1330, 1340 and 1350. Note that none of these CCs are of similar size to the frame CC 1310. 10 Fig. 14 shows a thin frame CC 1410 which is classified as GOUTLINE by the CC classification process 330. The frame CC 1410 contains no children due to a broken right side 1440 of the frame. However the frame CC 1410 touches a CC 1420 which has a child CC 1430. Fig. 15 shows the flowchart of the process 1140 to find a graphics container CC. 15 An initial step 1520 tests if the current CC currentcc 1510 has a child which has Total Pixel Count over a threshold (eg. 80% of Enclosed Pixel Count of currentcc). If so, then this child CC is assigned as the graphics container CC in step 1530, and the process 1140 terminates or otherwise returns. For example, this is the case if the CC being processed is a thin frame, making the graphics container CC the interior of the frame. 20 An example of this is shown in Fig. 12. If currentcc 1510 has a number of children as determined in step 1540, but none with Total Pixel Count over a threshold (eg. 80% of Enclosed Pixel Count of current cc), then the currentcc 1510 itself is identified as the graphics container in step 1550. An example of this is shown in Fig. 13. Otherwise touching sibling CCs are checked at step 1570 to find a suitable graphics container CC at 1898437_1 888293_speci_lodge -28 step 1580. A suitable graphics container is a CC which is classified as BG or G_FILL, has Total Pixel Count over a threshold (eg. 90% of the Bounding Box Area of current cc) and has a Bounding Box which is within current_cc's Bounding Box. For example this is the case when there is a thin frame which is broken. The broken frame 5 does not contain the frame interior, but touches it. An example of this is shown in Fig. 14. If no good touching sibling CC can be found to be a graphics container, then no graphics container CC has been found, as seen at step 1590. Fig. 16 shows the flowchart for the process 1170, mentioned above, for profiling the children of the graphics container CC. The profiling is to extract features from the 10 children of the graphics container CC in order to obtain semantic context. The features can be categorised into the following: Features Statistics Definitions Graphics area graphicsarea The number of pixels belong to a graphics object Text text-ent The number of TEXT child CCs text_gccnt The number of TEXT grandchild CCs Noise noise cnt The number of NOISE CCs noisepix The sum of Total Pixel Count of NOISE CCs Solidity solidity_cnt The number of solid like CCs soliditypix The sum of Total Pixel Count of solid like CCs Descendant descendantent The sum of Descendent Count of all child CCs descendantpix The sum of Enclosed Pixel Count of all child CCs Graphics interior cnt The number of graphics interior CCs intenor 1898437_1 888293_speci_lodge - 29 The above statistics are initialised to zero before the profiling. The process 1170 operates in a loop 1610 where each child CC of the graphics container CC is checked for certain properties and statistics which are updated and stored in the memory 234. If the child CC belongs to a graphics group, as tested in step 1615, then the graphics area 5 statistics are updated in step 1620 and the process 1170 continues to next child CC, obtained in step 1685. If the child CC is classified as TB_CELL, as tested at step 1625, then interiorent and text_gecnt are updated in step 1630. TBCELL classified CCs are likely to be CCs directly enclosed by an outline of a graphics object and may contain text. After step 1630, the TBCELL classified child CC is checked in step 1635 to 10 determine whether it is a good text container. If the test in step 1635 is true, the solidity statistics are updated in step 1670. If the child CC is classified as BG or G_FILL as determined in step 1640, then the solidity statistics are also updated in step 1670. If the child CC is classified as TEXT, as determined in step 1645, then the graphics container CC is checked in step 1650 as to whether it is a good text container. If the test in step 15 1650 is true, text Cnt is updated in step 1660. Otherwise the TEXT classified CC is checked whether it is large (eg. Self Pixel Count > 500) in decision step 1655. If it is large, the solidity statistics are updated in step 1670. Otherwise the process flow continues to step 1680 to update the descendant statistics. If the child CC is classified as NOISE, as determined at step 1665, the noise statistics are updated in step 1675 and the 20 flow continues to the next child CC at step 1685. After all the child CCs are profiled and their statistics tabulated in the memory 234, objcnt is calculated by subtracting noisecnt from Number of Enclosed of the graphics container CC. objcnt indicates the number of child CCs that are useful for semantic context analysis. These values may then also be stored in the memory 234. 1898437_1 888293_specilodge -30 The features obtained from process 1170 are analysed to decide whether the graphics container CC indeed belongs to a graphics object. There are three outcomes: (i) The graphics container CC contains graphics elements only and therefore all the descendent CCs are to be grouped into a graphics group. 5 (ii) The graphics container CC contains graphics elements and potential text as well. The descendent CCs are selectively grouped into a graphics group to exclude the potential text. (iii) The semantic features show that the graphics container CC does not belong to a graphics object and it is ignored. 10 Fig. 17 shows the flowchart for the process 1180, discussed above, of deciding whether the graphics container CC belongs to a graphics object and how the neighbourhood CCs are grouped together. Step 1710 initially checks whether the graphics container CC only contains graphics elements. The check returns true if any of the following three conditions (each having multiple sub-conditions) are met: 15 A. (i) currentcc is not a good text container; and (ii) graphics_area is greater than 50% of Bounding Box Area of current-cc; and (iii) text-gccnt equals zero. OR: 20 B. (i) The graphics container CC is classified as GOUTLINE; and (ii) objcnt equals solid cnt plus textcnt; and (iii) descendant_cnt equals zero. OR: C. (i) The graphics container CC is classified as GOUTLINE; and 1898437_1 888293_specilodge -31 (ii) interior_cnt is greater than 50% of objent; and (iii) solidpix is greater than 90% Enclosed Pixel Count of the graphics container CC; and (iv) descendantpix is less than noisepix. 5 If any of the above conditions are met then the graphics container and all descendants of the graphics container CC are grouped into a graphics group in step 1720. In addition, if the graphics container is not current-cc, then currentcc is also added to the graphics group. If the test of step 1710 returns false, then a test is performed in step 1730 to 10 determine whether the graphics container CC contains graphics elements and potential text as well, based on the following conditions: (i) The Bounding Box Area of currentcc is less than a threshold compared to the page size (for example 12.5%); and (ii) solidpix is more than 25% of Total Pixel Count for the graphics 15 container CC; and (iii) Number of Enclosed of the graphics container CC is less than a threshold (eg. 100). If the above conditions are true, the children are selectively grouped into a graphics group together with the graphics container CC and currentcc in step 1750. 20 Otherwise current-cc is ignored at step 1740 and no graphics grouping is performed, at which point the process 1180 returns. Step 1750 selectively groups each child CC of the graphics container CC into a graphics group if any of these conditions are true: (i) The child CC is classified as G_FILL or GOUTLINE; or 1898437_1 888293_specilodge -32 (ii) The child CC is classified as TBCELL or BG and its Number of Enclosed is less than a threshold (eg. 50); or (iii) The child CC is classified as TEXT and the graphics container CC is not a good text container and: 5 (iii-i) the child CC is solid (eg. Coverage > 90%); or (iii-ii) the child CC has low text confidence (eg. using machine learning for TEXT classification). To consolidate the graphics region, after step 1720 and step 1750, the child CCs enclosed by currentcc and the touching sibling CCs of currentcc are further 10 examined in step 1760 to check whether they should be grouped into the graphics group to complete a graphics region. Step 1760 traverses the subtree of each child CC of currentcc. The descendant CC is grouped into the graphics group if either of the following two conditions are true: (i) The descendant CC is classified as G_FILL or TB_CELL; or 15 (ii) The descendant CC has already been grouped into a different graphics group. Step 1760 groups in the touching sibling CC if either of the following conditions are true: (i) The touching sibling CC is classified as G_FILL or GOUTLINE; 20 or (ii) The touching sibling CC is classified as BG or TBCELL and its Bounding Box is fully contained within the graphics group's Bounding Box. Fig. 18 shows an example of the extracted graphics region from the snowman graphics in Fig. 4. The black part is the extracted graphics region. The graphics 1898437_1 888293_spedlodge - 33 container CC 441 is determined to contain graphics elements and potential text by step 1730 and TEXT CCs are therefore left out by step 1750. The net effect of step 1180 is to decide whether the graphics container contains a whole or part of a graphics object based on the semantic context statistics. 5 Photo Extraction Fig. 19 shows the flowchart of the photo extraction process 350 which is broken into four linearly configured stages - photo seeding 1910, photo region growing 1920, photo region consolidation 1930, and erroneous photo seed removal 1940. The photo seeding stage 1910 finds the CCs with rich photo features from the CC enclosure tree 10 and designates those CCs as photo seeds based on contextual information from the neighbourhood of surrounding CCs. The photo region growing stage 1920 selectively groups CCs in the neighbourhood of each of the photo seeds into a photo group in order to grow photo regions from the photo seeds. A photo group is a photo region isolated from the background and representing a logical photo object in the original document. A 15 photo group is formed using a list of PHOTO CCs. Following the photo region growing process 1920 is the photo region consolidation process 1930 which consolidates nearby photo groups and non-text CCs. All ungrouped photo seeds after the photo region consolidation process are removed from photo extraction by the process 1940. Each of these processing stages will be explained in further detail with reference to Fig. 20 to 20 Fig. 22. Fig. 20 shows the flowchart of the photo seeding process 1910. For photo seeding, the CC tree 600 is traversed in bottom up, children first order and each CC is processed for photo seeding. The tree traversing order is the same as used for the graphics extraction process 340 as shown in Fig. 10, where the process 1040 is replaced 1898437_1 888293_spedlodge -34 by the photo seeding process 1910. The process 1910 examines each CC (current-cc) for photo rich features. As seen in Fig. 20, initially the currentcc is checked for suitability in step 2010. A suitable CC for photo feature analysis is a CC that satisfies the following two 5 conditions: (i) The CC is a non-text; and (ii) The sub-tree depth is less than 5. (CCs with sub-tree depth higher than 4 have compound features.) If currentcc is not suitable for photo feature analysis, currentcc is ignored at 10 step 2030, at which point the next CC in the traversal can be considered. Otherwise the flow proceeds to step 2020 to analyse TEXT classified children of the currentcc. The fine details of a photo often have text characteristics at CC level because they are normally segmented to text-sized CCs. These text-sized CCs are often classified as TEXT. However when these TEXT classified CCs are considered together, they have 15 different characteristics from a group of real text CCs as normally found in a text paragraph. Step 2050 then determines whether the currentcc has photo features when any of the following conditions A-E is met: A. (i) The number of all the TEXT classified children is greater than a threshold (eg. 10); and 20 (ii) The mean of Self Pixel Count of all the TEXT children is smaller than a threshold (eg. 15). OR: B. (i) The number of TEXT classified children is greater than a threshold (eg. 10); and 1898437_1 888293_speci.Jodge -35 (ii) The mean of Self Pixel Count of all the TEXT children is smaller than a threshold (eg. 40); and (iii) The average bin contains higher than 50% of all the TEXT classified children; and 5 (iv) The medium bin contains higher than 70% of all the TEXT classified children. OR: C. (i) The number of TEXT classified children is greater than a threshold (eg. 10); and 10 (ii) The mean of Self Pixel Count of all the TEXT classified children is smaller than a threshold (eg. 40); and (iii) The average bin contains higher than 20% of all the TEXT classified children; and (iv) The small bin contains more than the average bin. 15 OR: D. (i) The number of TEXT classified children is greater than a threshold (eg. 10); and (ii) The high bin contains higher than 40% of all the TEXT classified children; and 20 (iii) The small bin contains higher than 40% of all the TEXT classified children. OR: E. (i) The total Self Pixel Count of all the TEXT children is greater than 1/16 of Total Pixel Count of current_cc; and 1898437_1 888293_speci_lodge -36 (ii) The Total Pixel Count of current-cc is less than 2/3 of the Bounding Box Area of current cc; and (iii) The variance of Self Pixel Count of all the TEXT children is greater than a threshold (eg. mean * mean * 10). 5 If there are photo features found, the currentcc and the sub-tree of currentcc are grouped into a photo group. All CCs in the photo group are reassigned with the PHOTO classification. When there are no photo features found in current_cc, the flow proceeds to step 2060 which checks whether currentcc contains any PHOTO child or is fragmented relative to the Bounding Box Area of current-cc (eg. Pixel Run Count is 10 greater than 3% of Bounding Box Area of current-cc). If the test in step 2060 is false, the flow proceeds to step 2070 which checks whether currentcc has busy surroundings. Currentcc has busy surroundings if the following conditions are met: (i) Number of Touching is greater than 100; and (ii) Number of Touching is greater than 1% of Selfpixel Count. 15 If the test in step 2070 is false, the flow proceeds to step 2080 which checks whether currentcc is in a colour difference range to any of its touching siblings. Photo regions often have shades of the same hue. The colour difference range is set to find colours that are visually different, but not too different. A colour distance function with a predetermined threshold is used to discriminate visually different colours. Given two 20 input colours, the colour distance function returns a number describing their similarity. In a preferred implementation, this is calculated as the Manhattan distance between the colours in YCbCr colours space, scaled by the difference in hue and saturation. To avoid converting large background area to a PHOTO CC, the touching siblings are limited to CCs with Bounding Box Area less than 1/16 of the page size and CCs that are not good 1898437_1 888293_specilodge - 37 text containers. If the test in step 2080 is false, the currentcc is ignored in step 2030. If currentcc is determined to be photo-like in steps 2060, 2070 or 2080, then currentcc is converted to PHOTO in step 2090. The photo seeding process in Fig. 20 is performed on each CC in the enclosure tree 600 in the order demonstrated in Fig. 10 5 until all the CCs are processed. After the photo seeding process 1910, all CCs with high probability as part of a photo object are converted to PHOTO. The CCs in the neighbourhood of each of the photo seeds will be examined for photo region growing from the photo seed. The neighbourhood of a PHOTO CC includes the touching siblings and the parent CC. Fig. 10 21 shows a flowchart of the photo region growing process 1920 for each photo seed on the touching sibling CCs and the parent CC. In a preferred implementation, the CC tree 600 is again traversed in bottom up, children first order and each CC is processed for photo region growing. ,Where appropriate or desirable, other approaches to accessing the CCs may be used in alternate implementations. Referring to the bottom up tree 15 traversing flow in Fig. 10, the process 1040 is replaced by process 1920 which processes current-cc. In an initial step 2110 of the process 1920, if currentcc is not classified as PHOTO, the CC is ignored at step 2130 and processing proceeds to the next CC. If currentcc is classified as PHOTO at step 2110, its parent CC is checked in step 2120 20 for photo region growing. The photo region is expanded into the parent if all of the following conditions are met: (i) The parent CC is classified as PHOTO or GOUTLINE; and (ii) The Total Pixel Count of the parent CC is less than 4 of the page size; and 1898437_1 888293_speci lodge -38 (iii) The parent CC does not touch any FRAME or GOUTLINE classified CCs; and (iv) The parent CC is not a good text container or the Self Pixel Count of the parent CC is less than the Total Pixel Count of currentcc. 5 After the parent CC is examined for photo region growing, the processing proceeds to step 2140 wherein each of the touching siblings of currentcc is examined for photo region growing. If the touching sibling CC is a good text container, as determined at step 2150, the touching sibling CC is excluded from photo region growing and the processing moves on to step 2195 to acquire the next touching sibling CC, which 10 is then processed at step 2140. Otherwise, step 2160 then checks whether the otherwise bad text container is classified as PHOTO. If not, the touching sibling CC is checked whether it is classified as TEXT at step 2165. The TEXT classified touching sibling CC is then checked at step 2170 to determine whether it is fully contained within the bounding box area of the expanded photo region of currentcc. A TEXT classified CC, 15 touching a PHOTO CC but not enclosed by the PHOTO CC, has high probability of being a falsely classified text character and therefore should be added to the touching photo group. The non TEXT classified touching sibling CC is checked at step 2180 to assess whether it increases the coverage of the expanded photo region of currentcc. Coverage is defined as the total number of pixels contributing to the region divided by 20 the bounding box of the region. If the coverage of the combined region is greater than either the Coverage of the touching sibling CC or the coverage of the expanded photo region of current-cc, the touching sibling CC is determined to be part of the logical photo object. When the touching sibling CC, either TEXT or non-text, is determined to 1898437_1 888293_speci lodge -39 be part of the touching photo object, it is added by way of step 2190 to the photo group to which current_cc belongs. After the photo region growing process 1920, each isolated photo object on the page should have been formed into a photo group. Because the photo region growing 5 process 1920 is based on touching siblings and enclosing parents, a logical photo object in the original document may have been split into more than one photo group due to imperfect colour quantisation and CC segmentation. The photo region consolidation process 1930 as shown in Fig. 22 further examines the already formed photo groups and non-text CCs in order to consolidate 10 nearby photo regions into logical photo objects. The photo region consolidation process 1930 loops through each photo group at step 2210 and considers all good candidates. Good candidates include all the other photo groups that have a bounding box that overlaps with the current photo group, as assessed at step 2220, and all non-text CCs that have bounding box overlaps with the current photo group, as assessed at step 2250. For 15 an overlapped photo group, identified at step 2220, the merging condition assessed at step 2230 is whether the coverage of the two merged photo groups is higher than the coverage of either of the two photo groups. If the test in step 2230 is true, the overlapped photo group is merged to the current photo group at step 2240 and all the statistics associated are updated accordingly. The process repeats via step 2245, which 20 identifies the next bounding box overlapped photo group, until all overlapped photo groups are considered. Next the overlapped non-text CCs are considered in a loop commencing at step 2250. For a non-text CC to be determined to be part of the nearby photo group, the following conditions have to be met: (i) The non-text CC is not a good text container; and 1898437_1 888293_specilodge - 40 (ii) The merging can produce coverage higher than either the Coverage of the non-text CC or the coverage of the photo group. If the above conditions are met, as assessed in step 2260, the non-text CC is added into the current photo group in step 2270. The process repeats by accessing the 5 next bounding box of the overlapped non-text CC at step 2275, by returning to step 2250 until all overlapped non-text CCs are considered. When the loops at 2220 and 2250 have completed, the next photo group is accessed at step 2280 and the process 1930 returns to step 2210. Following the photo region consolidation process 1930, there may be still some 10 photo seeds that are not grouped into any of the photo groups. It is unlikely that a logical photo object only consists of a CC. To avoid false photo extraction, all ungrouped photo seeds are changed to BG. As a result, the photo region growing and photo region consolidation processes 1920 and 1930 also serve as a confirmation step for the photo seeding process. 15 Text Extraction Fig. 23 shows the flowchart of the text extraction process 360, which groups CCs of text characteristics (TEXT CCs) to form logical text blocks such as text lines and paragraphs. The text blocks can then be extracted for applications such as OCR. The TEXT CC grouping process 360 normally employs a nearest-neighbour 20 search technique to find suitable candidates in the proximity for each TEXT CC. A grouping decision for each TEXT CC is then made based on its distance to all candidates. Typically there are a lot more TEXT CCs in a colour compound document image with mixed contents than in a black and white document image. The increase in the number of TEXT CCs is caused by the increase of image reproduction artefacts 1898437_1 888293_speci_lodge -41 aforementioned in the background art. These artefacts are often of text size and as a result they are classified as TEXT. Performing text grouping directly with a significant number of misclassified TEXT CCs is not only time consuming but also affects the accuracy of the logical text 5 blocks. In a preferred implementation as seen in Fig. 23, connected components 2305 are input and the TEXT CC classification is initially refined in step 2310 to reduce misclassification based on the statistics derived from the global colour context information. The refined TEXT CCs are then grouped into logical text blocks by a process 2320. Documents with mixed-contents may have text regions that overlap with 10 non-text regions. The actual text CCs in the overlapped portion may have been misclassified as non-text CCs by the CC classification, graphics or photo extraction processes. The text blocks formed by the text grouping process can provide additional context to correct the misclassification of actual text CCs based on the text characteristics derived from a text block. The actual text CCs misclassified as non-text 15 CCs are recovered into a nearby existing text block by a process 2330. In the preferred implementation, the TEXT CC classification refinement process takes a conservative approach to reject misclassified TEXT CCs because the global colour context information may not be able to provide enough information to make a confident decision in all situations. In such occasions the misclassified TEXT CCs will be grouped into a 20 text block together with the actual text. A further process 2340 then extracts local colour context information from a text block so that a confident decision can be made to reject misclassified TEXT CCs from a text block, to then output 2345 correct text blocks, which may be stored in the memory 234. 1898437_1 888293_specilodge - 42 Refine CC classifications There are known methods for text CC classification in the art. These methods normally use pixel properties extracted from CCs and classify them into TEXT or non text. Colour information is generally not used because one cannot assume which colour 5 is used to represent text characters. Though text characters are normally represented with darker colours compared to the surrounding background, assuming such a colour combination will inevitably miss out lighter text colours on dark background. However it is reasonable to make the assumption that text characters are of uniform colours in the majority of colour documents. Such assumption is true except for word arts which may 10 use non uniform colours as fills such as gradation or patterns. Typically there are only a limited number of colours used to represent text in a document and text characters of the same colour are normally used to represent a word, a sentence or a paragraph. There is generally some space between text characters. Based on these facts, one can assume that a TEXT CC should not touch another TEXT CC of a different colour. When touching 15 TEXT CCs exist, such contradicts with the facts mentioned above and causes a conflict. This conflict shows that some of these touching TEXT CCs must have been introduced as a result of imperfect segmentation caused by the aforementioned artefacts, or form part of complex background. Such misclassified TEXT CCs are preferably corrected before text grouping. 20 Fig. 24 is a flowchart explaining the TEXT CC classification refinement process 2310, which attempts to solve the conflict between touching TEXT CCs. The conflict resolution is determined based on which of the touching TEXT CCs is less likely to be the real text. The likelihood of being a real text character is decided by the global CC classification histogram of each colour present on the page. Each colour present on the 1898437_1 888293_specilodge -43 page has a likelihood of being used for text representation. There are two stages to the process 2310 - text colour likelihood calculation 2410 and resolving conflicts 2420. Fig. 25 explains the detail of process 2410 which profiles the CC classification results for each colour and then calculates the likelihood of being a text colour based on 5 the statistics in each colour bucket. Due to colour fluctuations caused by printing, scanning and colour quantisation, similar colours present on the page are profiled into the same colour bucket. A loop is established at step 2500 such that the process 2410 selectively places CCs on the page into a colour bucket while accumulating CC classification statistics. Since the CC classification profiling for a specific colour is to 10 determine its likelihood of being a text colour, it is important to exclude some CCs from the profiling so that a more accurate likelihood estimation for a colour can be obtained. Step 2510, being a first step in the loop, makes a decision on which CCs should be ignored. This includes CCs without a valid classification, such as those in the sub-tree of a TEXT classified CC which are not classified during the CC classification process 15 330. These unclassified CCs are of no importance to the text likelihood calculation and are ignored. Noise-sized TEXT CCs are also ignored so that the statistics are not polluted. TEXT CCs smaller than a certain size (eg. Self Pixel Count < 15) may actually be noise and are excluded from profiling. Such noise-sized TEXT CCs are placed into a noise colour bucket without updating the bucket statistics. The bucket choice for each 20 CC is based on colour similarity as determined in step 2520. If there is no existing colour bucket with a similar colour found, a new colour bucket is created for the CC. Each colour bucket contains the following information: Statistics Definitions TEXT CC Count The number of all the TEXT classified CCs being placed in 1898437_1 888293_speci_lodge -44 the bucket Non-text CC Count The number of all the non-text classified CCs being placed in the bucket NOISE CC Count The number of all the NOISE classified CCs being placed in the bucket TEXT Pixel Count The sum of Self Pixel Count of all the TEXT classified CCs being placed in the bucket Non-text Pixel Count The sum of Self Pixel Count of all the non-text classified CCs being placed in the bucket NOISE Pixel Count The sum of SelfPixel Count of all the NOISE classified CCs being placed in the bucket TEXT CC Size Mean The mean of Self Pixel Count of all the TEXT classified CCs being placed in the bucket TEXT CC Size The variance of Self Pixel Count of all the TEXT classified Variance CCs being placed in the bucket Total CC Count The number of all the CCs being placed in the bucket Total Pixel Count The sum of Self Pixel Count of all the CCs being placed in the bucket Step 2530 then updates the above statistics accordingly when a new CC is placed into the bucket. Once all the useful CCs on the page are placed into a suitable colour bucket, the likelihood of each colour bucket being a text colour is calculated in step 2540 5 based on the accumulated statistics. The likelihood estimation is desirably expressed as a percentage, and is based on the following pseudo code: CC eighted Likelihood = TEXT CC Count / Total CC Count * 100 Pixel Weighted Likelihood = TEXT Pixel Count / Total Pixel Count * 100 10 1898437_1 888293_spedlodge -45 IF NOISE CC Count > Total CC Count * 0.95 OR TEXT CC Count = 0 THEN likelihood = 0 5 ELSE IF TEXT CC Count >= 3 AND TEXT CC Size Variance < TEXT CC Size Mean * 2 AND TEXT CC Count > (Total CC Count - NOISE CC Count) * 0.7 AND TEXT Pixel Count > Total Pixel Count * 0.7 AND 10 THEN likelihood = 80 ELSE IF TEXT CC Count >= 20 AND CC Weighted Likelihood > 80 AND Pixel Weighted Likelihood < 10 15 THEN likelihood = 20 ELSE IF TEXT CC Size Variance < TEXT CC Size Mean AND TEXT Pixel Count > Total Pixel Count * 0.5 THEN likelihood = CC Weighted Likelihood * 1.2 20 ELSE IF TEXT CC Size Variance < TEXT CC Size Mean * 5 THEN likelihood = CC Weighted Likelihood * 0.8 ELSE likelihood = CC Weighted Likelihood * 0.25 25 The resolving conflicts stage 2420 in Fig. 24 traverses the enclosure tree 600 and examines each TEXT classified CC TXT in a loop established at step 2425. Step 2430 rejects a TEXT classified CC as bad text if any of the following conditions is true: - TXT is in a colour bucket with zero text colour likelihood; - TXT has a PHOTO classified parent CC and TXT is in a colour bucket 30 with low text colour likelihood (eg. < 20); and 1898437_1 888293_specilodge - 46 - the colour difference between TXT and its parent is not significant (eg. < 50). The classification of a rejected TEXT CC is changed to BG by step 2440 and the process 2420 continues to the next TEXT CC at step 2494. Otherwise the process 2420 5 continues to step 2450 to find whether TXT has any conflicts with any of its touching sibling TEXT CCs (TCH). A conflict sometimes is caused by imperfect segmentation due to colour fluctuation. Such a conflict should be resolved by merging the two CCs in order to fix text character fragmentation. Fig. 32A shows two examples of text character fragmentation, where CC 3201 and CC 3203 originally belong to the same part of a text 10 character. Imperfect segmentation causes them to become two CCs. CC 3202 and CC 3204 are another such example. Step 2460 checks whether TCH and TXT are both part of a fragmented text character using the following two conditions: A. (i) The colour distance between TCH and TXT is not very high (eg. < 80); and 15 (ii) TCH and TXT overlap; and (iii) Both TCH and TXT have high Edge Ratio (eg. > 0.7) and both are in a colour bucket with high text colour likelihood (eg. >= 80). OR: B. (i) TCH and TXT are of similar colour; and 20 (ii) TCH and TXT do not overlap with each other. If either of the above two conditions is met, TCH and TXT are merged together into a single TEXT CC at step 2470. Fig. 32B shows two examples of a fragmented text character being merged into a complete text character, complementing the examples of Fig. 32A. If TCH and TXT are not both part of a fragmented text character, one of them 1898437_1 888293_speci_lodge - 47 might be a misclassified TEXT CC resulted from the aforementioned artefacts. Further analysis is performed at step 2480 which checks whether TCH is more likely a misclassified TEXT CC compared to TXT. TCH in the pair is determined to be a misclassified TEXT CC if any of the following conditions is true: 5 (i) TCH is in a colour bucket with zero text colour likelihood; (ii) TXT is in a colour bucket with high text colour likelihood (eg. > 50), and TCH's Self Pixel Count is very small (eg. < 40), and TXT's Self Pixel Count is much greater than TCH's Self Pixel Count (eg. at least four times greater); (iii) TXT is in a colour bucket with higher TEXT CC Count and higher 10 TEXT CC Size Mean than TCH's. If TCH is determined to be a misclassified TEXT CC, TCH's classification is changed to NOISE at step 2490. Fig. 30A is an example of a text character 'K' on complex background. After CC classification, the letter 'K' (CC 3010) and two cut outs of the blob-shaped background 3005, CC 3020 and CC 3030, are all classified as TEXT, 15 as seen in Fig. 30B. The misclassified TEXT CCs, 3020 and 3030, are corrected by comparing the text colour likelihood between the colour of the letter 'K' and the colour of the blob-shaped background 3005. Normally the text colour likelihood of a real text is much higher than the non-text colours, which correspondingly have a significantly lower text colour likelihood because text of a specific colour hardly exists by one TEXT 20 CC. When step 2480 cannot make a decision, the potentially misclassified TEXT CCs remain the TEXT classification. The potential misclassifications are left to be analysed by the process 2340, discussed above, where a better local context statistics can be obtained from a text block. TEXT misclassifications resulted from the aforementioned 1898437_1 888293_speci_lodge -48 artefacts have the same characteristics to the example in Fig. 30B except the misclassified TEXT CCs are relatively smaller than CC 3020 and CC 3030. After step 2490, the next TCH is retrieved at step 2492 and the process 2420 returns to step 2450 as discussed above. 5 When all of the touching sibling TEXT CCs of TXT are analysed, process for TXT is completed. Once all TEXT CCs in the enclosure tree are analysed, the resolving conflicts stage 2420 finishes. Group text CCs into logical block structure After the CC classification refinement process 2310, the refined TEXT CCs are 10 ready to be grouped into logical text blocks at step 2320 mentioned above. Fig. 26 explains the process 2320 in detail, which groups TEXT CCs into text lines, which are then grouped into text paragraphs. A text line has a reading order direction in either the horizontal or vertical direction. Typically a horizontal text line consists of TEXT CCs with closer neighbouring TEXT CCs in horizontal direction than the neighboring 15 TEXT CCs in vertical direction. As seen in Fig. 26, initially step 2610 calculates the adjacency for each TEXT CC to other nearby TEXT CCs. In the present preferred implementation a nearest neighbour graph for each TEXT CC is calculated. Each TEXT CC has four nearest neighbours, one in each of the top, left, right, and bottom directions. Based on the neighbour information, each TEXT CC is assigned a text direction 20 depending on the distance to its neighbours. A TEXT CC has a horizontal direction if the neighbour distance between the TEXT CC to its left or right neighbour is smaller than the neighbour distance between the TEXT CC to its top or bottom neighbour. The neighbour distance between two TEXT CCs is decided by the geometric distance in text 1898437_1 888293_specilodge -49 direction, the alignment difference in the direction orthogonal to text direction and the text height difference. Step 2620 forms text lines based on the assigned text direction for each TEXT CC. The text direction determines the direction of grouping. CCs with horizontal 5 direction are grouped with its horizontal neighbours. CCs with vertical direction are grouped with its vertical neighbours. The dimension of the CCs at the direction orthogonal to the text direction is considered as text line height. CCs with same direction are grouped together if they are close to each other and align well in the text directions. 10 After text lines are formed, the text line adjacency analysis is performed at step 2630. Adjacent text lines of the same text direction within a predefined line distance are then grouped together to form text paragraphs in step 2640. Recover misclassified non-text CCs into existing text blocks Once all TEXT CCS on the page are grouped into text blocks by process 2320, 15 another classification refinement is performed by process 2330 which corrects text that have been misclassified as non-text. Fig. 27 explains the process 2330 of recovering misclassified non-text CCs into existing text blocks in detail. The process 2330 examines each text block at least once in at loop founded on step 2700 to recover the nearby misclassified non-text CCs that should have been part of the text block. 20 Step 2710 then operating of a text block T, collects a list of nearby non-text CCs for text block as the candidates for text recovery. Since the aim is to extend T, the bounding box of T is expanded slightly by a predefined distance D in the left, top, right and bottom directions. D is calculated based on T's average text line height. Preferably, 1898437_1 888293_spedlodge -50 D is two times of the average text line height of text block T. The non-text CCs that overlap with the expanded bounding box are collected. Each non-text CC NT from the list collected in step 2710 is examined in step 2720 to check whether adding NT would improve T's logical structure quality. T's 5 logical structure quality is improved, and NT is added into T at step 2730, if all of the following conditions are true: (i) the difference between T's text line height and NT's height is small (eg. < 12); and (ii) NT's Coverage is high enough (eg. > 0.12); and 10 (iii) the distance of NT to the closest CC in T is small in text direction (eg. < average line height / 2) and their colour difference is within a threshold (eg. < 50); and (iv) the union bounding box of NT and T does not overlap with other text blocks on the page; and 15 (v) NT and T align very well in text direction (eg. character alignment difference < 10). After all collected non-text CCs have been examined, step 2740 checks whether it is necessary to reconsider the non-text CCs that previously fail the test in step 2720. T's Bounding Box Area may be expanded by the addition of non-text CCs in step 2730. 20 Non-text CCs that previously are not selected as candidates may become qualified after the bounding box expansion of T by step 2730. Step 2740 makes the decision on whether or not to repeat the text recovery process for T based on how much the bounding box of T has been expanded. If any of the left, right, top or bottom side is expanded 1898437_1 888293_speci_lodge -51 more than 80% of D, text recovery for T is repeated. Otherwise the process moves on to the next text block at step 2750. Fig. 31 A - Fig. 31 C are the sample diagrams demonstrating how text recovery process works. Fig. 31A is the original image which contains a text paragraph on a 5 complex background. Fig. 31B is the output of process 2310 - process 2320. The black parts in this figure are all perfect text and have been grouped together into a single text block, whereas the grey tints in this figure (i.e. those without the strikethrough line), including 3101 and 3102, are the CCs that have been mis-classified as non-text. It will be apparent to a human observer that these grey tints are all perfect text as well, and 10 should be grouped into the text block together with the other text characters on the page. Fig. 31C shows the output of process 2330, from which all the grey tints have been reclaimed back as TEXT and grouped into the nearby text block. Remove misclassified text CCs from text blocks The final stage 2340 of the text extraction process of Fig. 23 is the removal of 15 incorrectly classified TEXT components from the text blocks, which is expanded as the flow diagram of Figure 28. The process 2340 operates in a loop at step 2805 for each of the text blocks 2800 on the page, where the process 2340 attempts to identify and remove noise-like components of the text block, which may be a result of misrepresentation errors present in the source image, such as colour bleeding, or the 20 result of incorrect text classifications. Additionally, it may be desirable to remove text decorations, such as outlines and drop shadows, which pose challenges to character recognition methods if both the text and decoration components are extracted. The process 2340 examines each of the text blocks individually, and begins by first determining at step 2810 if the text block consists of only a small number of CCs, 1898437_1 888293_speci_lodge - 52 for example, fewer than three. If the text block does not contain many connected components, it is then checked at step 2820 to assess if the text block is closely spatially located to a large number of NOISE CCs. If this is the case, then it is most likely that the CCs of the text block have been falsely classified as TEXT, and so the entire contents of 5 the text block are reclassified as NOISE at step 2830. In one example, the method 2340 determines if the text block is in a high noise region at step 2820 by examining any NOISE-classified CCs that touch the CCs of the text block. For each of those touched NOISE CCs, the bounding box area is accumulated, and if the sum total area is greater than twice the bounding box area of the text block, the text CCs are considered NOISE 10 and removed at step 2830. In this fashion the removal of the extraneous CCs from a text block removes a text block if the contained CCs touch a significant noise area wherein the significance is measured as the ratio of accumulated area of the touched noise CCs to the area of the text block itself. Otherwise, the text block is considered to at least partially consist of real text, 15 and a more detailed analysis is conducted at step 2840 to determine if individual noise like components, such as text decorations, can be removed. This process is expanded as the block diagram shown in Fig. 29, which is described in further detail below. Following this process, the next text block of the page is retrieved at step 2850 from the memory 234 and is examined to remove other small isolated text blocks and 20 text decoration components. The process 2840 of removing misclassified or extraneous components from a text block is shown in Fig. 29, to which is input a single text block 2900. The process module 2840 distinguishes undesirable TEXT components based on information about the colours present in the text block, which is generated by the first processing stage 1898437_1 888293_specilodge -53 2910. For each of the colour groups present in the text block, the process records information including: (i) Total CCs: The total number of CCs of this colour in the text block (ii) Pixel counts: The total SelfPixel Count for all CCs of this colour, and the 5 average Self Pixel Count (iii) Inter-colour touching counts: The number of CCs of each other colour in the text block that CCs of this colour touch, as well as the total number of differently coloured CCs that CCs of this colour touch. The processing of step 2840 as seen in Fig. 29 cycles through each of the colour 10 groups present in the text block via a loop 2920, and evaluates each selected (reference) colour group against all other (neighbour) colour groups via a loop 2925. The process compares at step 2930 the reference and neighbour colour groups to determine if the neighbour colour group constitutes undesired noise such as text decorations. If so, the process 2840 removes all CCs of the neighbour colour from the text block at step 2950. 15 In a preferred implementation the neighbour colour group is determined to be noise-like in step 2940 if both of the following conditions are fulfilled: (i) the average Self Pixel Count of the neighbour colour CCs is much smaller than that of the reference colour CCs (for example, less than 60%); and (ii) a significant portion of the neighbour coloured CCs touch CCs of 20 any different colour, or the reference colour specifically (for example, greater than 40% and 60% respectively). Step 2955 follows step 2950 and a "no" decision of step 2940 and obtains the next neighbour colour, returning the process to step 2925. Once all the neighbour colours have been processed, step 2960 accesses the next reference colour to return to 1898437_1 888293_speci lodge - 54 the loop at step 2920. In this manner, all of the colours constituting a text block are examined, and small noise artefacts as well as larger text decoration components are excluded from the extracted output. An example of the text removal process is shown in Fig. 33A, where it is desired 5 to remove the shadowed text decoration from the extracted output. The input consists of a sample page 3300 containing a single text block 3310 in two colours having a main text component 3320, in black, which is desired to be retained, and the shadow component 3330, in grey, which is desired to be removed. It can be seen from Fig. 33A that the strokes of the shadow text characters are often "broken", where they have been 10 overlapped by the main character strokes. This results in the decoration component consisting of a larger number of connected components than the main text, with a correspondingly smaller average pixel size. Accordingly, the process can determine at step 2940 that the shadow component 3330 should be removed, leaving only the main text component 3320 in the extracted output, as shown in Fig. 33B. 15 INDUSTRIAL APPLICABILITY The arrangements described are applicable to the computer and data processing industries and particularly for the extraction of text and other data types from documents by processes for document analysis. Whilst the arrangements are described as being operable within a general purpose computer system (Figs. 2A and 2B), the automation 20 and absence of user control makes the arrangements well suited to incorporation into copier or photocopier apparatus where document pages are routinely scanned. The foregoing describes only some embodiments of the present invention, and modifications and/or changes can be made thereto without departing from the scope and spirit of the invention, the embodiments being illustrative and not restrictive. 1898437_1 888293_specilodge -55 (Australia Only) In the context of this specification, the word "comprising" means "including principally but not necessarily solely" or "having" or "including", and not "consisting only of'. Variations of the word "comprising", such as "comprise" and "comprises" have correspondingly varied meanings. 5 1898437_1 888293_specilodge

Claims (16)

1. A computer implemented method of identifying at least one of a photo region and a graphics region in a colour bitmap image of a compound document, said method 5 comprising: (a) connecting similarly coloured pixels of said image into connected components and placing the connected components in an enclosure tree, the enclosure tree forming a hierarchy of ancestor and descendent connected components where an ancestor connected component encloses a descendent connected component wherein 10 each node of the enclosure tree is a connected component; (b) classifying each connected component into one of a plurality of classes wherein at least one class represents non-text connected components; (c) selecting a connected component from one of the non-text connected components of the enclosure tree; 15 (d) calculating context statistics for the selected connected component based on descendent and touching sibling connected components of the selected connected component; (e) identifying whether the selected connected component contains at least one of a photo region and a graphics region based on the context statistics; and 20 (f) storing the selected connected component identified as either graphics or photo object to memory.
2. A method according to claim I wherein said selecting step comprises: 2441468_1 888293_speci amend_01 - 57 (ca) selecting a non-text connected component with either a thin characteristic or low pixel density compared to a bounding box area of the non-text connected component; (cb) selecting a child connected component of said non-text connected 5 component as the selected connected component if said non-text connected component encloses the child connected component with a total pixel count greater than a first threshold based on an enclosed pixel count of said non-text connected component; (cc) selecting the non-text connected component as the selected connected component if said non-text connected component encloses more than one child 10 connected component and none of the child connected components have a total pixel count greater than the first threshold; (cd) selecting a touching connected component of said non-text connected component as the selected connected component if the touching connected component has children and a total pixel count greater than a second threshold based on a pixel 15 count of the bounding box of said non-text connected component.
3. A method according to claim I wherein the calculating context statistics step calculates, from descendent CCs, at least one statistic selected from the group of statics consisting of: 20 (i) number of pixels belonging to a graphics object; (ii) number of TEXT child CCs; (iii) number of TEXT grandchild CCs; (iv) number of NOISE CCs; (v) sum of Total Pixel Count of NOISE CCs; 2441468_1 888293_speci amend_01 - 58 (vi) number of solid like CCs; (vii) sum of Total Pixel Count of solid like CCs; (viii) sum of Descendent Count of all child CCs; (ix) sum of Enclosed Pixel Count of all child CCs; and 5 (x) number of graphics interior CCs.
4. A method according to claim 1 wherein the calculating context statistics calculates, from touching CCs, at least one (Fig. 16, 1170) statistic of either a busy state of surrounding connected components or a colour difference range to touching siblings 10 of the selected connected component.
5. A method according to claim 1 wherein the calculating context statistics step further comprises: calculating for said selected connected component at least one (Fig. 16, 1170) 15 statistic selected from the group consisting of: (i) a size and density histogram of TEXT classified children; and (ii) a fragmentation state.
6. A method according to claim 1 wherein the identifying step further comprises: 20 classifying the selected connected component as a good text container if a size and density histogram of TEXT classified children has each of the following characteristics: (i) a number of TEXT classified children is greater than a third predetermined threshold; 2441468_1 888293_speci_amend_01 - 59 (ii) a mean of Self Pixel Count of TEXT classified children is greater than a forth predetermined threshold; (iii) a variance of Self Pixel Count of TEXT classified children is less than a fifth predetermined threshold. 5
7. A method according to claim 1 wherein a decision on whether the selected connected component contains a photo region depends on said size and density histogram of TEXT classified children having no characteristics of a text paragraph. 10
8. A method according to claim I wherein said selected connected component identified as a graphics region is optimised by the steps of: [1760] (ga) consolidating grouped graphics descendants; (gb) consolidating touching childless CCs; and (gc) consolidating flat CCs that are fully contained by a grouped graphics 15 group.
9. A method according to claim I wherein selected connected component identified as a photo region is optimised by the steps of: (ha) consolidating a parent CC if the parent CC is not a good text container; 20 (hb) consolidating any touching sibling non-text CCs if consolidating any of the touching sibling non-text CCs will increase a pixel density of a bounding box area of said photo region; and (hc) consolidating any touching sibling text CCs if any of the touching sibling text CCs are fully enclosed by the bounding box area of said photo region. 2441468_1 888293_speciamend_01 -60
10. A computer readable storage medium having a computer program recorded thereon, the program being executable by computer apparatus to identify at least one of a photo region and a graphics region in a colour bitmap image of a compound document, 5 said program comprising: code for connecting similarly coloured pixels of said image into connected components and placing the connected components in an enclosure tree, the enclosure tree forming a hierarchy of ancestor and descendent connected components where an ancestor connected component encloses a descendent connected component wherein 10 each node of the enclosure tree is a connected component; code for classifying each connected component into one of a plurality of classes wherein at least one class represents non-text connected components; code for selecting a connected component from one of the non-text connected components of the enclosure tree; 15 code for calculating context statistics for the selected connected component based on descendent and touching sibling connected components of the selected connected component; code for identifying whether the selected connected component contains at least one of a photo region and a graphics region based on the context statistics; and 20 code for storing the selected connected component identified as either graphics or photo object to memory.
11. Computer apparatus, said apparatus comprising a processor and a memory coupled to the processor, the memory having recorded thereon a program for identifying 2441468_1 888293_speci_amend_01 -61 at least one of a photo region and a graphics region in a colour bitmap image of a compound document, the image being stored in the memory, the program comprising: code means executable by the processor for connecting similarly coloured pixels of said image into connected components and placing the connected components in an 5 enclosure tree, the enclosure tree forming a hierarchy of ancestor and descendent connected components where an ancestor connected component encloses a descendent connected component wherein each node of the enclosure tree is a connected component; code means executable by the processor for classifying each connected component into one of a plurality of classes wherein at least one class represents non 10 text connected components; code means executable by the processor for selecting a connected component from one of the non-text connected components of the enclosure tree; code means executable by the processor for calculating context statistics for the selected connected component based on descendent and touching sibling connected 15 components of the selected connected component; code means executable by the processor for identifying whether the selected connected component contains at least one of a photo region and a graphics region based on the context statistics; and code means executable by the processor for storing in the memory the selected 20 connected component identified as either graphics or photo object.
12. A copier comprising: a scanner for scanning a compound document to form a colour bitmap image; a processor coupled to the scanner; 2441468_1 888293_speci amend_01 -62 a memory coupled to the processor for storing the colour bitmap image, said memory having recorded thereon a program executable by the processor for identifying at least one of a photo region and a graphics region in the colour bitmap image of the compound document, the program comprising: 5 code means for connecting similarly coloured pixels of said image into connected components and placing the connected components in an enclosure tree, the enclosure tree forming a hierarchy of ancestor and descendent connected components where an ancestor connected component encloses a descendent connected component wherein each node of the enclosure tree is a connected component; 10 code means for classifying each connected component into one of a plurality of classes wherein at least one class represents non-text connected components; code means for selecting a connected component from one of the non-text connected components of the enclosure tree; code means for calculating context statistics for the selected connected 15 component based on descendent and touching sibling connected components of the selected connected component; code means for identifying whether the selected connected component contains at least one of a photo region and a graphics region based on the context statistics; and 20 code means for storing in the memory the selected connected component identified as either graphics or photo object..
13. A computer implemented method of identifying at least one of a photo region and a graphics region in a colour bitmap image of a compound document, said method being 2441468_1 888293_speci amend_01 - 63 substantially as described herein with reference to any one of the embodiments as that embodiment is illustrated in the drawings.
14. A computer readable storage medium having a computer program recorded 5 thereon, the program being substantially as described herein with reference to any one of the embodiments as that embodiment is illustrated in the drawings.
15. Computer apparatus adapted to perform the method of any one of claims I to 9 or 13. 10
16. A copier comprising apparatus according to claim 15. DATED this seventh Day of September, 2010 15 CANON KABUSHIKI KAISHA Patent Attorneys for the Applicant Spruson&Ferguson 2441468_1 888293_speciamend_01
AU2008260018A 2008-12-18 2008-12-18 Refining text extraction in colour compound documents Ceased AU2008260018C1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
AU2008260018A AU2008260018C1 (en) 2008-12-18 2008-12-18 Refining text extraction in colour compound documents
US12/637,446 US8351691B2 (en) 2008-12-18 2009-12-14 Object extraction in colour compound documents

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
AU2008260018A AU2008260018C1 (en) 2008-12-18 2008-12-18 Refining text extraction in colour compound documents

Publications (3)

Publication Number Publication Date
AU2008260018A1 AU2008260018A1 (en) 2010-07-08
AU2008260018B2 AU2008260018B2 (en) 2010-09-23
AU2008260018C1 true AU2008260018C1 (en) 2011-01-20

Family

ID=42313393

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2008260018A Ceased AU2008260018C1 (en) 2008-12-18 2008-12-18 Refining text extraction in colour compound documents

Country Status (1)

Country Link
AU (1) AU2008260018C1 (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0404236A1 (en) * 1989-06-21 1990-12-27 Océ-Nederland B.V. Image segmentation method and device
US5202933A (en) * 1989-12-08 1993-04-13 Xerox Corporation Segmentation of text and graphics
JPH10283475A (en) * 1997-04-04 1998-10-23 Minolta Co Ltd Image processing method and device
US5956468A (en) * 1996-07-12 1999-09-21 Seiko Epson Corporation Document segmentation system
WO2000036546A1 (en) * 1998-12-11 2000-06-22 Koninklijke Philips Electronics N.V. Method for detecting a face in a digital image
EP0685959B1 (en) * 1994-05-31 2000-07-26 NEC Corporation Image processing apparatus for identifying character, photo and dot images in the image area
EP1061731A2 (en) * 1999-06-14 2000-12-20 Riso Kagaku Corporation Method of and system for distinguishing regions in black and white image information
EP1315126A2 (en) * 2001-11-26 2003-05-28 Ricoh Company, Ltd. Image processing method, image processing apparatus, image output apparatus and storage medium
EP0828375B1 (en) * 1996-09-09 2004-04-21 Sharp Kabushiki Kaisha Image-region discriminating method and image-processing apparatus
US20060088214A1 (en) * 2004-10-22 2006-04-27 Xerox Corporation System and method for identifying and labeling fields of text associated with scanned business documents
US20070172137A1 (en) * 2006-01-23 2007-07-26 Konica Minolta Business Technologies, Inc. Image processing apparatus capable of accurately and quickly determining character part included in image

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0404236A1 (en) * 1989-06-21 1990-12-27 Océ-Nederland B.V. Image segmentation method and device
US5202933A (en) * 1989-12-08 1993-04-13 Xerox Corporation Segmentation of text and graphics
EP0685959B1 (en) * 1994-05-31 2000-07-26 NEC Corporation Image processing apparatus for identifying character, photo and dot images in the image area
US5956468A (en) * 1996-07-12 1999-09-21 Seiko Epson Corporation Document segmentation system
EP0828375B1 (en) * 1996-09-09 2004-04-21 Sharp Kabushiki Kaisha Image-region discriminating method and image-processing apparatus
JPH10283475A (en) * 1997-04-04 1998-10-23 Minolta Co Ltd Image processing method and device
WO2000036546A1 (en) * 1998-12-11 2000-06-22 Koninklijke Philips Electronics N.V. Method for detecting a face in a digital image
EP1061731A2 (en) * 1999-06-14 2000-12-20 Riso Kagaku Corporation Method of and system for distinguishing regions in black and white image information
EP1315126A2 (en) * 2001-11-26 2003-05-28 Ricoh Company, Ltd. Image processing method, image processing apparatus, image output apparatus and storage medium
US20060088214A1 (en) * 2004-10-22 2006-04-27 Xerox Corporation System and method for identifying and labeling fields of text associated with scanned business documents
US20070172137A1 (en) * 2006-01-23 2007-07-26 Konica Minolta Business Technologies, Inc. Image processing apparatus capable of accurately and quickly determining character part included in image

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
ETEMAD, K. et al., 'Multiscale Document Page Segmentation Using Soft Decision Integration', IEEE Transactions on Pattern Analysis and Machine Intelligence, January 1997, Volume 19 Issue 1, pages 92-96, June 2000 *
Patent Abstracts of Japan & JP 10283475 A (MINOLTA CO LTD) 23 October 1998 *
ZHANG, W. et al., 'Using Artificial Neural Networks to Identify Headings in Newspaper Documents', Proceedings of the International Joint Conference on Neural Networks 2003, Volume 3, 20-24 July 2003, pages 2283-2287 *

Also Published As

Publication number Publication date
AU2008260018A1 (en) 2010-07-08
AU2008260018B2 (en) 2010-09-23

Similar Documents

Publication Publication Date Title
US8351691B2 (en) Object extraction in colour compound documents
US10191889B2 (en) Systems, apparatuses and methods for generating a user interface by performing computer vision and optical character recognition on a graphical representation
US8532374B2 (en) Colour document layout analysis with multi-level decomposition
US8045800B2 (en) Active segmentation for groups of images
JP6151763B2 (en) Word segmentation for document images using recursive segmentation
Lazzara et al. Efficient multiscale Sauvola’s binarization
US8503781B2 (en) Finding text regions from coloured image independent of colours
US7627176B2 (en) Apparatus, method, and computer program for analyzing document layout
US20110007366A1 (en) System and method for classifying connected groups of foreground pixels in scanned document images according to the type of marking
JP6188976B2 (en) Method, apparatus and computer-readable recording medium for detecting text contained in an image
US6711292B2 (en) Block selection of table features
JPH05217019A (en) Business form identifying system and image processing system
CN110598581B (en) Optical music score recognition method based on convolutional neural network
JP5539488B2 (en) Judgment of transparent fill based on reference background color
AU2008260018C1 (en) Refining text extraction in colour compound documents
Bhardwaj et al. Spectral–spatial active learning with attribute profile for hyperspectral image classification
CN112861843A (en) Method and device for analyzing selection frame based on feature image recognition
AU2008260130A1 (en) Graphics object extraction in compound colour documents
Dey et al. A comparative study of margin noise removal algorithms on marnr: A margin noise dataset of document images
Alhéritière et al. Straight line reconstruction for fully materialized table extraction in degraded document images
US20240153294A1 (en) Automatic template recommendation
Kieri Context dependent thresholding and filter selection for optical character recognition
Vu Building extraction from high-resolution satellite image for tsunami early damage estimation
AU2012268796A1 (en) Directional stroke width variation feature for script recognition
Akcay et al. Automated detection of objects using multiple hierarchical segmentations

Legal Events

Date Code Title Description
DA2 Applications for amendment section 104

Free format text: THE NATURE OF THE AMENDMENT IS AS SHOWN IN THE STATEMENT(S) FILED 07 SEP 2010.

DA3 Amendments made section 104

Free format text: THE NATURE OF THE AMENDMENT IS AS SHOWN IN THE STATEMENT(S) FILED 07 SEP 2010

FGA Letters patent sealed or granted (standard patent)
MK14 Patent ceased section 143(a) (annual fees not paid) or expired