32
views
0
recommends
+1 Recommend
1 collections
    0
    shares

      Submit your digital health research with an established publisher
      - celebrating 25 years of open access

      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Crowdsourcing a Normative Natural Language Dataset: A Comparison of Amazon Mechanical Turk and In-Lab Data Collection

      research-article
      , PhD 1 , , , PhD 2 , , PhD 1
      (Reviewer)
      Journal of Medical Internet Research
      JMIR Publications Inc.
      Internet, web, crowdsourcing, free recall

      Read this article at

      ScienceOpenPublisherPMC
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Background

          Crowdsourcing has become a valuable method for collecting medical research data. This approach, recruiting through open calls on the Web, is particularly useful for assembling large normative datasets. However, it is not known how natural language datasets collected over the Web differ from those collected under controlled laboratory conditions.

          Objective

          To compare the natural language responses obtained from a crowdsourced sample of participants with responses collected in a conventional laboratory setting from participants recruited according to specific age and gender criteria.

          Methods

          We collected natural language descriptions of 200 half-minute movie clips, from Amazon Mechanical Turk workers (crowdsourced) and 60 participants recruited from the community (lab-sourced). Crowdsourced participants responded to as many clips as they wanted and typed their responses, whereas lab-sourced participants gave spoken responses to 40 clips, and their responses were transcribed. The content of the responses was evaluated using a take-one-out procedure, which compared responses to other responses to the same clip and to other clips, with a comparison of the average number of shared words.

          Results

          In contrast to the 13 months of recruiting that was required to collect normative data from 60 lab-sourced participants (with specific demographic characteristics), only 34 days were needed to collect normative data from 99 crowdsourced participants (contributing a median of 22 responses). The majority of crowdsourced workers were female, and the median age was 35 years, lower than the lab-sourced median of 62 years but similar to the median age of the US population. The responses contributed by the crowdsourced participants were longer on average, that is, 33 words compared to 28 words ( P<.001), and they used a less varied vocabulary. However, there was strong similarity in the words used to describe a particular clip between the two datasets, as a cross-dataset count of shared words showed ( P<.001). Within both datasets, responses contained substantial relevant content, with more words in common with responses to the same clip than to other clips ( P<.001). There was evidence that responses from female and older crowdsourced participants had more shared words ( P=.004 and .01 respectively), whereas younger participants had higher numbers of shared words in the lab-sourced population ( P=.01).

          Conclusions

          Crowdsourcing is an effective approach to quickly and economically collect a large reliable dataset of normative natural language responses.

          Related collections

          Most cited references26

          • Record: found
          • Abstract: found
          • Article: not found

          Conducting behavioral research on Amazon's Mechanical Turk.

          Amazon's Mechanical Turk is an online labor market where requesters post jobs and workers choose which jobs to do for pay. The central purpose of this article is to demonstrate how to use this Web site for conducting behavioral research and to lower the barrier to entry for researchers who could benefit from this platform. We describe general techniques that apply to a variety of types of research and experiments across disciplines. We begin by discussing some of the advantages of doing experiments on Mechanical Turk, such as easy access to a large, stable, and diverse subject pool, the low cost of doing experiments, and faster iteration between developing theory and executing experiments. While other methods of conducting behavioral research may be comparable to or even better than Mechanical Turk on one or more of the axes outlined above, we will show that when taken as a whole Mechanical Turk can be a useful tool for many researchers. We will discuss how the behavior of workers compares with that of experts and laboratory subjects. Then we will illustrate the mechanics of putting a task on Mechanical Turk, including recruiting subjects, executing the task, and reviewing the work that was submitted. We also provide solutions to common problems that a researcher might face when executing their research on this platform, including techniques for conducting synchronous experiments, methods for ensuring high-quality work, how to keep data private, and how to maintain code security.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            The viability of crowdsourcing for survey research.

            Online contract labor portals (i.e., crowdsourcing) have recently emerged as attractive alternatives to university participant pools for the purposes of collecting survey data for behavioral research. However, prior research has not provided a thorough examination of crowdsourced data for organizational psychology research. We found that, as compared with a traditional university participant pool, crowdsourcing respondents were older, were more ethnically diverse, and had more work experience. Additionally, the reliability of the data from the crowdsourcing sample was as good as or better than the corresponding university sample. Moreover, measurement invariance generally held across these groups. We conclude that the use of these labor portals is an efficient and appropriate alternative to a university participant pool, despite small differences in personality and socially desirable responding across the samples. The risks and advantages of crowdsourcing are outlined, and an overview of practical and ethical guidelines is provided.
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              The rise of crowdsourcing

                Bookmark

                Author and article information

                Contributors
                Journal
                J Med Internet Res
                JMIR
                Journal of Medical Internet Research
                JMIR Publications Inc. (Toronto, Canada )
                1439-4456
                1438-8871
                May 2013
                20 May 2013
                : 15
                : 5
                : e100
                Affiliations
                [1] 1Schepens Eye Research Institute Boston, MAUnited States
                [2] 2Schepens Eye Research Institute, Massachusetts Eye and Ear Boston, MAUnited States
                Author notes
                Corresponding Author: Daniel R Saunders daniel_saunders@ 123456meei.harvard.edu
                Article
                v15i5e100
                10.2196/jmir.2620
                3668615
                23689038
                b39d7bc1-f252-4876-bc33-f31f9644dfc4
                ©Daniel R Saunders, Peter J Bex, Russell L Woods. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 20.05.2013.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License ( http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.

                History
                : 18 March 2013
                : 11 April 2013
                : 25 April 2013
                : 25 April 2013
                Categories
                Original Paper

                Medicine
                internet,web,crowdsourcing,free recall
                Medicine
                internet, web, crowdsourcing, free recall

                Comments

                Comment on this article