60
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      DeepPoseKit, a software toolkit for fast and robust animal pose estimation using deep learning

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Quantitative behavioral measurements are important for answering questions across scientific disciplines—from neuroscience to ecology. State-of-the-art deep-learning methods offer major advances in data quality and detail by allowing researchers to automatically estimate locations of an animal’s body parts directly from images or videos. However, currently available animal pose estimation methods have limitations in speed and robustness. Here, we introduce a new easy-to-use software toolkit, DeepPoseKit, that addresses these problems using an efficient multi-scale deep-learning model, called Stacked DenseNet, and a fast GPU-based peak-detection algorithm for estimating keypoint locations with subpixel precision. These advances improve processing speed >2x with no loss in accuracy compared to currently available methods. We demonstrate the versatility of our methods with multiple challenging animal pose estimation tasks in laboratory and field settings—including groups of interacting individuals. Our work reduces barriers to using advanced tools for measuring behavior and has broad applicability across the behavioral sciences.

          eLife digest

          Studying animal behavior can reveal how animals make decisions based on what they sense in their environment, but measuring behavior can be difficult and time-consuming. Computer programs that measure and analyze animal movement have made these studies faster and easier to complete. These tools have also made more advanced behavioral experiments possible, which have yielded new insights about how the brain organizes behavior.

          Recently, scientists have started using new machine learning tools called deep neural networks to measure animal behavior. These tools learn to measure animal posture – the positions of an animal’s body parts in space – directly from real data, such as images or videos, without being explicitly programmed with instructions to perform the task. This allows deep learning algorithms to automatically track the locations of specific animal body parts in videos faster and more accurately than previous techniques. This ability to learn from images also removes the need to attach physical markers to animals, which may alter their natural behavior.

          Now, Graving et al. have created a new deep learning toolkit for measuring animal behavior that combines components from previous tools with the latest advances in computer science. Simple modifications to how the algorithms are trained can greatly improve their performance. For example, adding connections between layers, or ‘neurons’, in the deep neural network and training the algorithm to learn the full geometry of the body – by drawing lines between body parts – both enhance its accuracy. As a result of adding these changes, the new toolkit can measure an animal's pose from previously unseen images with high speed and accuracy, after being trained on just 100 examples. Graving et al. tested their model on videos of fruit flies, zebras and locusts, and found that, after training, it was able to accurately track the animals’ movements. The new toolkit has an easy-to-use software interface and is freely available for other scientists to use and build on.

          The new toolkit may help scientists in many fields including neuroscience and psychology, as well as other computer scientists. For example, companies like Google and Apple use similar algorithms to recognize gestures, so making those algorithms faster and more efficient may make them more suitable for mobile devices like smartphones or virtual-reality headsets. Other possible applications include diagnosing and tracking injuries, or movement-related diseases in humans and livestock.

          Related collections

          Most cited references82

          • Record: found
          • Abstract: found
          • Article: not found

          DeepLabCut: markerless pose estimation of user-defined body parts with deep learning

          Quantifying behavior is crucial for many applications in neuroscience. Videography provides easy methods for the observation and recording of animal behavior in diverse settings, yet extracting particular aspects of a behavior for further analysis can be highly time consuming. In motor control studies, humans or other animals are often marked with reflective markers to assist with computer-based tracking, but markers are intrusive, and the number and location of the markers must be determined a priori. Here we present an efficient method for markerless pose estimation based on transfer learning with deep neural networks that achieves excellent results with minimal training data. We demonstrate the versatility of this framework by tracking various body parts in multiple species across a broad collection of behaviors. Remarkably, even when only a small number of frames are labeled (~200), the algorithm achieves excellent tracking performance on test frames that is comparable to human accuracy.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Hybrid Monte Carlo

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

              State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet and Fast R-CNN have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features---using the recently popular terminology of neural networks with 'attention' mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model, our detection system has a frame rate of 5fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available. Extended tech report
                Bookmark

                Author and article information

                Contributors
                Role: Senior Editor
                Role: Reviewing Editor
                Journal
                eLife
                Elife
                eLife
                eLife
                eLife Sciences Publications, Ltd
                2050-084X
                01 October 2019
                2019
                : 8
                : e47994
                Affiliations
                [1 ]deptDepartment of Collective Behaviour Max Planck Institute of Animal Behavior KonstanzGermany
                [2 ]deptDepartment of Biology University of Konstanz KonstanzGermany
                [3 ]deptCentre for the Advanced Study of Collective Behaviour University of Konstanz KonstanzGermany
                [4 ]deptDepartment of Computer Science Princeton University PrincetonUnited States
                [5 ]deptChair for Computer Aided Medical Procedures Technische Universität München MunichGermany
                Max Planck Institute for Chemical Ecology Germany
                Princeton University United States
                Princeton University United States
                Princeton University United States
                Author information
                https://orcid.org/0000-0002-5826-467X
                https://orcid.org/0000-0002-2447-6295
                https://orcid.org/0000-0001-5291-788X
                https://orcid.org/0000-0001-8556-4558
                Article
                47994
                10.7554/eLife.47994
                6897514
                31570119
                64d723b2-882d-42a4-97d2-2c690ca86690
                © 2019, Graving et al

                This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.

                History
                : 26 April 2019
                : 18 September 2019
                Funding
                Funded by: FundRef http://dx.doi.org/10.13039/100000001, National Science Foundation;
                Award ID: IOS-1355061
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100000006, Office of Naval Research;
                Award ID: N00014-09-1-1074
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100000006, Office of Naval Research;
                Award ID: N00014-14-1-0635
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100000183, Army Research Office;
                Award ID: W911NG-11-1-0385
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100000183, Army Research Office;
                Award ID: W911NF14-1-0431
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/501100001659, Deutsche Forschungsgemeinschaft;
                Award ID: DFG Centre of Excellence 2117
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/501100010583, University of Konstanz;
                Award ID: Zukunftskolleg Investment Grant
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/501100003542, Ministry of Science, Research and Art Baden-Württemberg;
                Award ID: The Strukture-und Innovations fonds fur die Forschung of the State of Baden-Wurttemberg
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/501100004189, Max Planck Society;
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100010661, Horizon 2020 Framework Programme;
                Award ID: Marie Sklodowska-Curie grant agreement No. 748549
                Award Recipient :
                Funded by: FundRef http://dx.doi.org/10.13039/100007065, Nvidia;
                Award ID: GPU Grant
                Award Recipient :
                Funded by: Nvidia;
                Award ID: GPU Grant
                Award Recipient :
                The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
                Categories
                Tools and Resources
                Neuroscience
                Custom metadata
                A new deep-learning software toolkit with general-purpose methods for quickly and reliably measuring the full body posture of animals directly from images or videos without physical markers.

                Life sciences
                grévy's zebra,desert locust,d. melanogaster,equus grevyi,schistocerca gregaria,other
                Life sciences
                grévy's zebra, desert locust, d. melanogaster, equus grevyi, schistocerca gregaria, other

                Comments

                Comment on this article