Blog
About

130
views
0
recommends
+1 Recommend
1 collections
    4
    shares
      • Record: found
      • Abstract: found
      • Conference Proceedings: found
      Is Open Access

      The Effect of Reasoning Strategies on Success in Early Learning of Programming: Lessons Learned from an External Experiment Replication

      , ,

      14th International Conference on Evaluation and Assessment in Software Engineering (EASE) (EASE)

      Evaluation and Assessment in Software Engineering

      12 - 13 April 2010

      Reasoning Strategies, Learning of Programming, Replication, Replicability, Lessons Learned

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Background. Literal or theoretical replications are important to evaluate and assess empirical results. However, there are still few replications in software engineering, and fewer external replications, i.e., developed by researchers other than the original ones. Aim. This paper discusses the difficulties found and the lessons learned from performing two literal replications of an experiment involving human subjects. Results. Our results apparently contradict the conclusions of the original experiment. However, several differences in context made it difficult to achieve valid comparability. Conclusion. Experiments involving human subjects should collect and report as many qualitative context information as possible, so the results can be related to the conditions under which the hypothesis were found to be true. Besides, given the difficulties found in this study, literal replication does not seem to be the best strategy for experiments involving human subjects in software engineering.

          Related collections

          Most cited references 4

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Reporting guidelines for controlled experiments in software engineering

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            On the difficulty of replicating human subjects studies in software engineering

              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              Using differences among replications of software engineering experiments to gain knowledge

                Bookmark

                Author and article information

                Contributors
                Conference
                April 2010
                April 2010
                : 1-10
                Affiliations
                Centro de Informática

                Universidade Federal

                de Pernambuco

                (CIn UFPE)

                Recife, PE, Brasil
                Article
                10.14236/ewic/EASE2010.11
                © A. César C. França et al. Published by BCS Learning and Development Ltd. 14th International Conference on Evaluation and Assessment in Software Engineering (EASE), Keele University, UK

                This work is licensed under a Creative Commons Attribution 4.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/

                14th International Conference on Evaluation and Assessment in Software Engineering (EASE)
                EASE
                Keele University, UK
                12 - 13 April 2010
                Electronic Workshops in Computing (eWiC)
                Evaluation and Assessment in Software Engineering
                Product
                Product Information: 1477-9358BCS Learning & Development
                Self URI (journal page): https://ewic.bcs.org/
                Categories
                Electronic Workshops in Computing

                Comments

                Comment on this article