Character-Level Interaction in Multimodal Computer-Assisted Transcription of Text Images (Pattern Recognition and Image Analysis)

Abstract

To date, automatic handwriting text recognition systems are far from being perfect and heavy human intervention is often required to check and correct the results of such systems. As an alternative, an interactive framework that integrates the human knowledge into the transcription process has been presented in previous works. In this work, multimodal interaction at character-level is studied. Until now, multimodal interaction had been studied only at whole-word level. However, character-level pen-stroke interactions may lead to more ergonomic and friendly interfaces. Empirical tests show that this approach can save significant amounts of user effort with respect to both fully manual transcription and non-interactive post-editing correction.

Introduction

At present time, the use of automatic handwritten text recognition systems (HTR) for the transcription of manuscript document images is far from being useful, mainly because of the unrestricted vocabulary and/or handwriting styles involved in such documents. Typically, the automatic transcriptions obtained by these HTR systems need a heavy human post-editing process in order to obtain transcriptions of standard quality. In practice, such a post-editing solution becomes rather inefficient, expensive and hardly acceptable by professional transcribers.

In previous works [7,5], a more effective, interactive on-line approach was presented. This approach, called "Computer Assisted Transcription of Handwritten Text Images" (CATTI), combines the accuracy ensured by the human transcriber with the efficiency of the HTR systems to obtain final perfect transcriptions. Empirical results show that the use of CATTI systems can save a substantial quantity of human effort with respect to both pure manual transcriptions and post-editing.


So far, human corrective feedback for CATTI has been studied at two different levels: a) whole-word interactions (both typed and handwritten using an e-pen interface [7]) and b) (typed) character-level corrections [5]. According to the results of these works, keystroke corrections can save a significant quantity of human effort with respect to whole-word corrections, while multimodal, e-pen interaction seems more ergonomic for human transcribers, which is a key point in the design of friendly and usable user interfaces.

In this work, we focus on character level interaction using the more ergonomic e-pen handwriting modality, which is perhaps the most natural way to provide the required feedback in CATTI systems. It is important to note, however, that the use of this kind of non-deterministic feedback typically increases the overall interaction cost in order to correct the possible feedback decoding errors. Nevertheless, by using informations derived from the interaction process, we will show how the decoding accuracy can be significantly improved over using a plain e-pen handwriting recognizer which can not take advantage of the interaction context.

CATTI Overview

In the original CATTI framework, the human transcriber (named user from now on) is directly involved in the transcription process since he is responsible of validating and/or correcting the HTR outputs. The process starts when the HTR system proposes a full transcription of a feature vector sequence x, extracted from a handwritten text line image. The user validates an initial part of this transcription, p’, which is error-free and introduces a correct word, v, thereby producing correct transcription prefix, p — p’v. Then, the HTR system takes into account the available information to suggest a new suitable continuation suffix, s. This process is repeated until a full correct transcription of x is accepted by the user [7].

At each step of this process, both the image representation, x, and a correct transcription prefix p are available and the HTR system should try to complete this prefix by searching for the most likely suffix s as:

tmp9F5-115_thumb

Since the concatenation ofp and s constitutes a full transcription hypothesis, P(x | p,s) can be approximated by concatenated character Hidden Markov Models (HMMs) [2,4] as in conventional HTR. On the other hand, P(s | p) is usually approximated by dynamically modifying a n-gram in order to cope with the increasingly consolidated prefixes [7]. Let p — pk be a consolidated prefix and s — si a possible suffix:

tmp9F5-116_thumb

In order to make the system more ergonomic and friendly to the user, interaction based on characters (rather than full words) has been studied in [5] with encouraging results. Now, as soon as the user types a new keystroke (character), the system proposes a suitable continuation following the same process described above. As the user operates now at the character level, the last word of the prefix may be incomplete. In order to autocomplete this last word, it is assumed that the prefix p is divided into two parts: the fragment of the prefix formed by complete words (p”) and the last incomplete word of the prefix (vp). In this case the HTR decoder has to take into account x, p” and vp, in order to search for a transcription suffix s, whose first part is the continuation of vp:

tmp9F5-117_thumb

Again,the concatenation oftmp9F5-118_thumb_ and s constitutes a full transcription hypothesis and tmp9F5-119_thumb, can be modelled with HMMs. On the other hand, to modeltmp9F5-120_thumb we assume that the suffix s is divided into two fragments: vs and s". vs is the first part of the suffix that corresponds with the final part of the incomplete word of the prefix, i.e, vpvs = v where v is an existing word in the task dictionarytmp9F5-121_thumband s" is the rest of the suffix. So, the search must be performed over all possible suffixes s of p, and the language model probabilitytmp9F5-122_thumbmust ensure that the concatenation of the last part of the prefix vp, and the first part of the suffix, vs, form an existing word (v) in the task dictionary. This probability can be decomposed into two terms: tmp9F5-128_thumb

The first term accounts for the probability of all the whole-words in the suffix, and can be modelled directly by (2). The second term should ensure that the first part of the suffix (usually a word-ending-part) vs, will be a possible suffix of the incomplete word vp, and can be stated as:

tmp9F5-129_thumb

Multimodal CATTI (MM-CATTI) at the Character Level

One way to increase the ergonomy and the usability in CATTI is to allow the user to provide his or her validating and/or corrective feedback by means of more comfortable peripheral devices, such as e-pen or touchscreen.

Clearly, decoding this kind of non-deterministic feedback consists in on-line HTR. As previously mentioned, the information available in the interaction process, allows us to boost the accuracy of this on-line HTR subsystem with respect to a conventional on-line HTR decoder (which do not make use of the interaction-derived information).

Let x be the representation of the input image and pp a user-validated prefix of the transcription. Let t be the on-line touchscreen pen strokes provided by the user. These data are related to the suffix suggested by the system in the previous interaction step, s’, and are typically aimed at accepting or correcting parts of this suffix. Using this information, the system has to find a new suffix, s, as a continuation of the previous prefix p’, considering all possible decodings, d, of the on-line data t and some information from the previous suffix s’. That is:

tmp9F5-130_thumb

An approximate two-step solution to this difficult optimization problem is followed (see Figure 1). In the first step, an "optimal" decoding, d, of the on-line pen-strokes t is computed using only the first two terms of equation (6). After observing this decoding, d, the user may type additional keystrokes, k, to correct possible errors in d. In the second step, the first two terms of (6) are ignored and d is replaced with ds in the last two terms. This way, a new consolidated prefix p = p’d is obtained, which leads to a formulation identical to (1). These two steps are repeated until p is accepted by the user as a full correct transcription of x.

Example of multimodal CATTI at character level interaction. The process starts when the HTR system proposes a full transcription of the handwritten text image x. Then, each interaction consists in two steps. In the first step the user handwrites some touchscreen to amend the suffix proposed by the system in the previous step. This defines a correct prefix p', which can be used by the on-line HTR subsystem to obtain a decoding of t. After observing this decoding, d, the user may type additional keystrokes, k, to correct possible errors in d. On the second step, a new prefix is built from the previous correct prefix p', the decoded on-line handwritten text, d, and the typed text k. Using this information, the system proposes a new potential suffix. The process ends when the user enters the special character "#". System suggestions are printed in boldface and typed text in typewriter font. In the final transcription, T, underlined italic characters are those which were typed by the user.

Fig. 1. Example of multimodal CATTI at character level interaction. The process starts when the HTR system proposes a full transcription of the handwritten text image x. Then, each interaction consists in two steps. In the first step the user handwrites some touchscreen to amend the suffix proposed by the system in the previous step. This defines a correct prefix p’, which can be used by the on-line HTR subsystem to obtain a decoding of t. After observing this decoding, d, the user may type additional keystrokes, k, to correct possible errors in d. On the second step, a new prefix is built from the previous correct prefix p’, the decoded on-line handwritten text, d, and the typed text k. Using this information, the system proposes a new potential suffix. The process ends when the user enters the special character "#". System suggestions are printed in boldface and typed text in typewriter font. In the final transcription, T, underlined italic characters are those which were typed by the user.

Assuming whole-word e-pen feedback, this approach was studied and tested in [7], with good results. Here we consider single-character e-pen strokes, which we think may lead to improved productive and usability. Therefore, we henceforth assume that d consists of a single character. As in section 2, the prefix p’ is divided into two parts: p” (fragment of p’ formed by complete words) and vp’ (the last incomplete word of p’). Therefore the first step of the optimization (6) can be written as:

tmp9F5-132_thumb

where, P (t | d) is provided by a morphological (HMM) model of the character d and tmp9F5-133_thumbcan be approached by a language model dynamically constrained by information derived from the interaction process. Equation (7) may lead to several scenarios depending on the assumptions and constraints adopted fortmp9F5-134_thumbWe examine some of them bellow.

The first and simplest scenario corresponds to a naive approach where any kind of interaction-derived information is considered; that is,tmp9F5-135_thumb

In a slighty more restricted scenario, we take into account just the information from the previous off-line HTR prediction s’. The user interacts providing t in order to correct the wrong character of s’, e, that follows the validated prefix p’. Clearly, the erroneous character e should be prevented to be a decoding on-line HTR result. This error-conditioned model can be written astmp9F5-139_thumb

Another, more restrictive scenario, using the information derived from the validated prefix p’, arises when we regard the portion of word already validatedtmp9F5-140_thumb

In this case the decoding should be easier as we know beforehand what should be a suitable continuation of the part of word accepted so far.

Finally, the most restrictive scenario corresponding to the additional consideration of the information provided by p”, is left for future studies.

Dynamic Language Modelling for Character-Level MM-CATTI

Language model restrictions are implemented on the base of n-grams, depending on each multimodal scenario considered. As mentioned above, the simplest scenario is that which does not take into account any information derived from the interaction. In this case, P(d) can be modelled directly using uni-grams. This is the baseline case.

The second case, P(d | e), only considers the first wrong character. The language model probability is given by

tmp9F5-145_thumb

The next scenario, given bytmp9F5-146_thumb, the on-line HTR subsystem counts not only on the first wrong character but also on the last incomplete word of the validated prefix vp’ . This scenario can be approached in two different ways: using a character language model or a word language model. In the first one, the on-line HTR subsystem uses a modified character n-gram model:

tmp9F5-148_thumb

In the second approach (10), we use a word language model to generate a more refined character language model. This can be written as:

tmp9F5-149_thumb

where:

tmp9F5-150_thumb

beingtmp9F5-151_thumban existing word of E.

Off- and On-Line HTR System Overview

Both the off-line and on-line HTR systems employ a similar conceptual architecture composed of three modules: preprocessing, feature extraction and recognition. The first two entail different well-known standard techniques depending on the data type, but the last one is identical for both systems. The Off-line HTR preprocessing involves skew and slant corrections and size normalization operation [8]. On the other hand, on-line handwriting preprocessing encompasses repeated points elimination and noise reduction. Regarding feature extraction, the off-line case converts the preprocessed text into a sequence of 60-dimensional feature vectors, whereas the on-line preprocessed coordinates are transformed into a sequence of 7-dimensional feature vectors [6].

As explained above, the recognition process is similar in both cases. Characters are modelled by continuous density left-to-right HMMs with a Gaussian mixture per state. Each lexical word is modelled by a Stochastic Finite-State automaton, and text sentences are modelled using bi-grams with Kneser-Ney back-off smoothing. All these finite-state models can be easily integrated into a single global model in which decoding process is efficiently performed by the Viterbi algorithm.

Experimental Framework

For test the effectiveness of MM-CATTI at character level different experiments were carried out. The corpora and the performance measures used are explained below.

Assessment Measures

Some types of measures have been adopted to assess the performance of character-level transcription. On the one hand, to make the post-editing process more accurately comparable to CATTI at character level, we introduce a post-editing autocompleting approach. Here, when the user enters a character to correct some incorrect word, the system automatically completes the word with the most probable word on the task vocabulary. Hence we define the Post-editing Key Stroke Ratio (PKSR), as the number of keystrokes that the user must enter to achieve the reference transcription, divided by the total number of reference characters. On the other hand, the effort needed by a human transcriber to produce correct transcriptions using CATTI at character level is estimated by the Key Stroke Ratio (KSR), which can be defined as the number of (character level) user interactions that are necessary to achieve the reference transcription of the text image considered, divided by the total number of reference characters. These definitions make PKSR and KSR comparable and the relative difference between them gives us a good estimate of the reduction in human effort that can be achieved by using CATTI at character level with respect to using a conventional HTR system followed by human autocompleting postediting. This estimated effort reduction will be denoted as "EFR".

Finally, since only single-character corrections are considered, the conventional classification error rate (ER) will be used to assess the accuracy of the on-line HTR feedback subsystem under the different constraints entailed by the MM-CATTI at character level interaction process.

Corpus

The character level CATTI was evaluated on the IAMDB corpus. For the MM-CATTI, the on-line UNIPEN corpus was employed to simulate the user touchscreen interactions.

The database is provided at different segmentation levels: characters, words, lines.Here we use sentence-segmented images. Each sentence is accompanied by its ground through transcription as the corresponding sequence of words. To better focus on the essential issues of the considered problems, no punctuation marks, diacritics, or different word capitalizations are included in the transcriptions. From 2,324 sentences that forms the corpus, 200 were used as test, leaving the rest as training partition.

The UNIPEN corpus [1] comes organized in several categories: lower and upper-case letters, digits, symbols, isolated words and full sentences. For our experiment, three UNIPEN categories were used: 1a (digits), 1c (lowercase letters) and 1d (symbols). Three arbritary writers were chosen as test partition and 17 as training data [7].

Results

Different experiments have been carried out to asses the feasibility and potential of CATTI at character level. Two types of results are reported for CATTI at character level: the PKSR (first column of table 2) and the KSR (second column of table 2). The 12.5% of KSR corresponds to a total of 1,627 characters that the user has to correct. In the MM-CATTI at character level these characters would have to be handwritten by the user on the touchscreen. It is simulated here using character samples belonging to a same writer from the UNIPEN corpus.

As we mentioned earlier, the introduction of multimodal interactivity leads, on the one hand, to an ergonomic and easier way of working, but on the other hand, to a situation where the system has to deal with non-deterministic feedback signals. Therefore, two of the most important concerns here is the accuracy of the on-line HTR subsystem and the determination of how much this accuracy can be boosted by taking into account informations derived from the interaction process. Table 1 reports the writer average feedback on-line recognition error rate of characters considering the different scenarios studied in section 3. As observed, feedback decoding accuracy increases significantly as more interaction derived constraints are taken into account. In addition, Table 1 also shows the relative accuracy improvements obtained respect to the baseline case.

Table 1. On-line HTR subsystem error rates for the four language models: plain character unigram (CU, baseline), error conditioned character uni-gram (CUe), prefix-and-error conditioned character bi-gram (CBe) and prefix-and-error conditioned word uni-gram (WUe). The relative accuracy improvements for CUe, CBe and WUe are shown in the last three columns. The same GSF value (15) is used for all the cases. All values are in percentages.

Erro

r Rate

Rel<

tive Imp

rov.

CU

CUe

CBe

WUe

CUe

CBe

WUe

7.0

6.9

6.7

5.0

1.4

4.3

28.6

Table 2. From left to right: PKSR obtained with the post-editing autocompleting approach, KSR achieved with CATTI at character level and KSR obtained with the baseline and best scenarios for MM-CATTI approach. EFR for KSR of CATTI with respect to PKSR and for KSR for the two scenarios of MM-CATTI with respect to PKSR. All results are in percentages.

CATTI

MM-I

2ATT1

EFR

PKSR

KSR

CU-KSR

WUe-KSR

CATTI

MM-CATTI (CU)

MM-CATTI (WUe)

15.8

12.5

13.4

13.1

20.9

15.2

17.1

As a final overview, Table 2 summarizes all the CATTI and MM-CATTI results obtained in this work. The third and fourth columns show the MM-CATTI KSR for the baseline as well as the best scenarios. This values are calculated under the simplifying assumption that the cost of keyboard-correcting a feedback on-line decoding error is similar to that of another on-line touchscreen interaction step. That is, each correction is counted twice: one for the failed touchscreen attempt and another for the keyboard correction itself. According to these results, the expected user effort for the best MM-CATTI approach is only barely higher than that of CATTI.

Conclusions

In this paper, we have studied the character level interaction in the CATTI system presented in previous works using pen strokes handwritten on a touchscreen as a complementary means to introduce the required CATTI correction feedback. From the results, we observe that the use of this more ergonomic feedback modality comes at the cost of a reasonably small number of additional interaction steps needed to correct the few feedback decoding errors. The number of these extra steps is kept very small thanks to the ability to use interaction-derived constraints to considerably improve the on-line HTR feedback decoding accuracy.

Next post:

Previous post: