72
views
0
recommends
+1 Recommend
1 collections
    1
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Automatic motion tracking of lips using digital video and OpenFace 2.0

      research-article

      Read this article at

      ScienceOpenPublisher
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Because the lips are external organs, they can be easily observed by means that are non-invasive, including simple video recording. The current paper introduces a free-of-charge, highly portable, automatic solution for extracting oral posture from digital video, based on an existing face-tracking utility. We describe how this solution might benefit various lines of laboratory phonology research, including analysis of articulatory coordination and audiovisual speech, as well as phonetic fieldwork. We then provide a tutorial, framed as a simple experiment in metronomic speech. The tutorial describes how to configure the software to work with one’s digital camera of choice, how to extract relevant articulatory parameters from the face tracker, and how to approach statistical analysis. We have supplied pre-written Python scripts to aid the reader in following along with the experiment. The portability, ease, and affordability of the described solution have important implications for the accessibility of articulatory measurement, both in the lab and out in the field.

          Related collections

          Most cited references28

          • Record: found
          • Abstract: found
          • Article: not found

          Random effects structure for confirmatory hypothesis testing: Keep it maximal.

          Linear mixed-effects models (LMEMs) have become increasingly prominent in psycholinguistics and related areas. However, many researchers do not seem to appreciate how random effects structures affect the generalizability of an analysis. Here, we argue that researchers using LMEMs for confirmatory hypothesis testing should minimally adhere to the standards that have been in place for many decades. Through theoretical arguments and Monte Carlo simulation, we show that LMEMs generalize best when they include the maximal random effects structure justified by the design. The generalization performance of LMEMs including data-driven random effects structures strongly depends upon modeling criteria and sample size, yielding reasonable results on moderately-sized samples when conservative criteria are used, but with little or no power advantage over maximal models. Finally, random-intercepts-only LMEMs used on within-subjects and/or within-items data from populations where subjects and/or items vary in their sensitivity to experimental manipulations always generalize worse than separate F 1 and F 2 tests, and in many cases, even worse than F 1 alone. Maximal LMEMs should be the 'gold standard' for confirmatory hypothesis testing in psycholinguistics and beyond.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Evaluating significance in linear mixed-effects models in R

            Mixed-effects models are being used ever more frequently in the analysis of experimental data. However, in the lme4 package in R the standards for evaluating significance of fixed effects in these models (i.e., obtaining p-values) are somewhat vague. There are good reasons for this, but as researchers who are using these models are required in many cases to report p-values, some method for evaluating the significance of the model output is needed. This paper reports the results of simulations showing that the two most common methods for evaluating significance, using likelihood ratio tests and applying the z distribution to the Wald t values from the model output (t-as-z), are somewhat anti-conservative, especially for smaller sample sizes. Other methods for evaluating significance, including parametric bootstrapping and the Kenward-Roger and Satterthwaite approximations for degrees of freedom, were also evaluated. The results of these simulations suggest that Type 1 error rates are closest to .05 when models are fitted using REML and p-values are derived using the Kenward-Roger or Satterthwaite approximations, as these approximations both produced acceptable Type 1 error rates even for smaller samples.
              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              OpenFace 2.0: Facial Behavior Analysis Toolkit

                Bookmark

                Author and article information

                Contributors
                Journal
                1868-6354
                Laboratory Phonology: Journal of the Association for Laboratory Phonology
                Ubiquity Press
                1868-6354
                07 July 2020
                2020
                : 11
                : 1
                : 9
                Affiliations
                [1 ]Department of Psychology, University of California Santa Cruz, Santa Cruz, CA, US
                [2 ]Department of Psychology, California State University Channel Islands, Camarillo, CA, US
                Article
                10.5334/labphon.232
                09cd8f8e-a83b-479d-a172-295169c862ae
                Copyright: © 2020 The Author(s)

                This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See http://creativecommons.org/licenses/by/4.0/.

                History
                : 30 September 2019
                : 06 June 2020
                Categories
                Journal article

                Applied linguistics,General linguistics,Linguistics & Semiotics
                motion tracking,digital video,articulatory measurement,lip aperture

                Comments

                Comment on this article