TY - GEN
T1 - Partially observable markov decision process (POMDP) technologies for sign language based human-computer interaction
AU - Ong, Sylvie C.W.
AU - Hsu, David
AU - Lee, Wee Sun
AU - Kurniawati, Hanna
PY - 2009
Y1 - 2009
N2 - Sign language (SL) recognition modules in human-computer interaction systems need to be both fast and reliable. In cases where multiple sets of features are extracted from the SL data, the recognition system can speed up processing by taking only a subset of extracted features as its input. However, this should not be realised at the expense of a drop in recognition accuracy. By training different recognizers for different subsets of features, we can formulate the problem as the task of planning the sequence of recognizer actions to apply to SL data, while accounting for the trade-off between recognition speed and accuracy. Partially observable Markov decision processes (POMDPs) provide a principled mathematical framework for such planning problems. A POMDP explicitly models the probabilities of observing various outputs from the individual recognizers and thus maintains a probability distribution (or belief) over the set of possible SL input sentences. It then computes a policy that maps every belief to an action. This allows the system to select actions in real-time during online policy execution, adapting its behaviour according to the observations encountered. We illustrate the POMDP approach with a simple sentence recognition problem and show in experiments the advantages of this approach over "fixed action" systems that do not adapt their behaviour in real-time.
AB - Sign language (SL) recognition modules in human-computer interaction systems need to be both fast and reliable. In cases where multiple sets of features are extracted from the SL data, the recognition system can speed up processing by taking only a subset of extracted features as its input. However, this should not be realised at the expense of a drop in recognition accuracy. By training different recognizers for different subsets of features, we can formulate the problem as the task of planning the sequence of recognizer actions to apply to SL data, while accounting for the trade-off between recognition speed and accuracy. Partially observable Markov decision processes (POMDPs) provide a principled mathematical framework for such planning problems. A POMDP explicitly models the probabilities of observing various outputs from the individual recognizers and thus maintains a probability distribution (or belief) over the set of possible SL input sentences. It then computes a policy that maps every belief to an action. This allows the system to select actions in real-time during online policy execution, adapting its behaviour according to the observations encountered. We illustrate the POMDP approach with a simple sentence recognition problem and show in experiments the advantages of this approach over "fixed action" systems that do not adapt their behaviour in real-time.
KW - Human-computer interaction
KW - Planning under uncertainty
KW - Sign language recognition
UR - http://www.scopus.com/inward/record.url?scp=70350322962&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-02713-0_61
DO - 10.1007/978-3-642-02713-0_61
M3 - Conference contribution
AN - SCOPUS:70350322962
SN - 3642027121
SN - 9783642027123
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 577
EP - 586
BT - Universal Access in Human-Computer Interaction
T2 - 5th International Conference on Universal Access in Human-Computer Interaction, UAHCI 2009. Held as Part of HCI International 2009
Y2 - 19 July 2009 through 24 July 2009
ER -