26
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      Visual Analytics Methodology for Eye Movement Studies

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references16

          • Record: found
          • Abstract: found
          • Book Chapter: not found

          Eye Tracking in HCI and Usability Research

          Eye tracking is a technique whereby an individual’s eye movements are measured so that the researcher knows both where a person is looking at any given time and the sequence in which the person’s eyes are shifting from one location to another. Tracking people’s eye movements can help HCI researchers to understand visual and display-based information processing and the factors that may impact the usability of system interfaces. In this way, eye-movement recordings can provide an objective source of interface-evaluation data that can inform the design of improved interfaces. Eye movements also can be captured and used as control signals to enable people to interact with interfaces directly without the need for mouse or keyboard input, which can be a major advantage for certain populations of users, such as disabled individuals. We begin this article with an overview of eye-tracking technology and progress toward a detailed discussion of the use of eye tracking in HCI and usability research. A key element of this discussion is to provide a practical guide to inform researchers of the various eye-movement measures that can be taken and the way in which these metrics can address questions about system usability. We conclude by considering the future prospects for eye-tracking research in HCI and usability testing.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Spatial generalization and aggregation of massive movement data.

            Movement data (trajectories of moving agents) are hard to visualize: numerous intersections and overlapping between trajectories make the display heavily cluttered and illegible. It is necessary to use appropriate data abstraction methods. We suggest a method for spatial generalization and aggregation of movement data, which transforms trajectories into aggregate flows between areas. It is assumed that no predefined areas are given. We have devised a special method for partitioning the underlying territory into appropriate areas. The method is based on extracting significant points from the trajectories. The resulting abstraction conveys essential characteristics of the movement. The degree of abstraction can be controlled through the parameters of the method. We introduce local and global numeric measures of the quality of the generalization, and suggest an approach to improve the quality in selected parts of the territory where this is deemed necessary. The suggested method can be used in interactive visual exploration of movement data and for creating legible flow maps for presentation purposes.
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Algorithms for defining visual regions-of-interest: comparison with eye fixations

                Bookmark

                Author and article information

                Journal
                IEEE Transactions on Visualization and Computer Graphics
                IEEE Trans. Visual. Comput. Graphics
                Institute of Electrical and Electronics Engineers (IEEE)
                1077-2626
                December 2012
                December 2012
                : 18
                : 12
                : 2889-2898
                Article
                10.1109/TVCG.2012.276
                3783bd63-84f3-42b0-b977-b8e9432ebc63
                © 2012
                History

                Comments

                Comment on this article