TY - GEN
T1 - A probabilistic language model for hand drawings
AU - Akce, Abdullah
AU - Bretl, Timothy
PY - 2010
Y1 - 2010
N2 - Probabilistic language models are critical to applications in natural language processing that include speech recognition, optical character recognition, and interfaces for text entry. In this paper, we present a systematic way to learn a similar type of probabilistic language model for hand drawings from a database of existing artwork by representing each stroke as a sequence of symbols. First, we propose a language in which the symbols are circular arcs with length fixed by a scale parameter and with curvature chosen from a fixed low-cardinality set. Then, we apply an algorithm based on dynamic programming to represent each stroke of the drawing as a sequence of symbols from our alphabet. Finally, we learn the probabilistic language model by constructing a Markov model. We compute the entropy of our language in a test set as measured by the expected number of bits required for each symbol. Our language model might be applied in future work to create a drawing interface for noisy and low-bandwidth input devices, for example an electroen-cephalograph (EEG) that admits one binary command per second. The results indicate that by leveraging our language model, the performance of such an interface would be enhanced by about 20 percent.
AB - Probabilistic language models are critical to applications in natural language processing that include speech recognition, optical character recognition, and interfaces for text entry. In this paper, we present a systematic way to learn a similar type of probabilistic language model for hand drawings from a database of existing artwork by representing each stroke as a sequence of symbols. First, we propose a language in which the symbols are circular arcs with length fixed by a scale parameter and with curvature chosen from a fixed low-cardinality set. Then, we apply an algorithm based on dynamic programming to represent each stroke of the drawing as a sequence of symbols from our alphabet. Finally, we learn the probabilistic language model by constructing a Markov model. We compute the entropy of our language in a test set as measured by the expected number of bits required for each symbol. Our language model might be applied in future work to create a drawing interface for noisy and low-bandwidth input devices, for example an electroen-cephalograph (EEG) that admits one binary command per second. The results indicate that by leveraging our language model, the performance of such an interface would be enhanced by about 20 percent.
UR - http://www.scopus.com/inward/record.url?scp=78149483613&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78149483613&partnerID=8YFLogxK
U2 - 10.1109/ICPR.2010.35
DO - 10.1109/ICPR.2010.35
M3 - Conference contribution
AN - SCOPUS:78149483613
SN - 9780769541099
T3 - Proceedings - International Conference on Pattern Recognition
SP - 109
EP - 112
BT - Proceedings - 2010 20th International Conference on Pattern Recognition, ICPR 2010
T2 - 2010 20th International Conference on Pattern Recognition, ICPR 2010
Y2 - 23 August 2010 through 26 August 2010
ER -