We investigate a method for selecting recordings of human face and head movements from a sign language corpus to serve as a basis for generating animations of novel sentences of American Sign Language (ASL). Drawing from a collection of recordings that have been categorized into various types of non-manual expressions (NMEs), we define a method for selecting an exemplar recording of a given type using a centroid-based selection procedure, using multivariate dynamic time warping (DTW) as the distance function. Through intra- and inter-signer methods of evaluation, we demonstrate the efficacy of this technique, and we note useful potential for the DTW visualizations generated in this study for linguistic researchers collecting and analyzing sign language corpora.
@inproceedings{kacorri:16007:sign-lang:lrec,
author = {Kacorri, Hernisa and Syed, Ali Raza and Huenerfauth, Matt and Neidle, Carol},
title = {Centroid-Based Exemplar Selection of {ASL} Non-Manual Expressions using Multidimensional Dynamic Time Warping and {MPEG4} Features},
pages = {105--110},
editor = {Efthimiou, Eleni and Fotinea, Stavroula-Evita and Hanke, Thomas and Hochgesang, Julie A. and Kristoffersen, Jette and Mesch, Johanna},
booktitle = {Proceedings of the {LREC2016} 7th Workshop on the Representation and Processing of Sign Languages: Corpus Mining},
maintitle = {10th International Conference on Language Resources and Evaluation ({LREC} 2016)},
publisher = {{European Language Resources Association (ELRA)}},
address = {Portoro{\v z}, Slovenia},
day = {28},
month = may,
year = {2016},
language = {english},
url = {https://www.sign-lang.uni-hamburg.de/lrec/pub/16007.pdf}
}