Abstract
Until sufficient practical experience and documented examples are available, emerging technological methods to create educational content typically require in-depth knowledge from different fields in order to apply them in the most beneficial way. The method of creating photorealistic 3D models (PR3DMs) of real-world objects with the methods of photogrammetry is an example of such emerging technologies. This paper discusses the fundamental concepts of photogrammetry to provide an overview of the necessary basic knowledge for educators and researchers interested in the presentation of realistic virtual materials. Additionally, a deeper introduction to basic psychological processes while learning with photorealistic materials is presented. This provides deeper insights into potentials and issues, as well as an overview of the most crucial research gaps in learning with PR3DMs. Finally, current and future applications, as well as the potential for learning topics apart from factual knowledge, such as emotional or behavioral outcomes, are discussed.
Introduction
Technologies used to create realistic virtual environments have made tremendous improvements over the last decades. Throughout this evolution, several milestones can be identified, where visual quality has taken a substantial leap forward, ranging from advanced methods of texture mapping (Blinn, 1978) to the widespread use of 3D engines (e.g., Hall et al., 1992). Most recently, one of these developments resulted in enthusiastic outbursts among computer game critics, as they stated that a group of developers “[…] were able to import wholesale from the real world weathered and highly detailed models that actually feel and look real. […] a world that is incredibly beautiful, immersive and tangible.” referring to the experience as “beyond awe-inspiring” (Aidan, 2014). Such reviews gained researchers’ attention, as realistic virtual environments might be useful in educational settings as well.
The technology that enabled a small developer to elicit such a positive review was photogrammetry. Despite its primary use in the context of a modern videogame design, photogrammetry is not a recent invention. Specifically, the approach of photogrammetry was already mentioned in the 19th century (e.g., Meydenbauer, 1867). Since then, the term has been used in various applications and contexts. Therefore, this paper illuminates the topic with an introduction to the field and a brief discussion about the relevant technical foundations. Using different psychological perspectives, the underlying processes during learning with the resulting digital materials are discussed. This discussion emphasizes the benefits and the challenges of photogrammetry and photorealistic 3D models (PR3DMs) in educational settings. Some examples and applications are presented next. To conclude this comprehensive review, a glimpse into the future of photogrammetry in education and research is provided.
What Is Photogrammetry?
The term photogrammetry is a variation of the word photography (Reulke and Döring, 2006; Kraus, 2012). The word photography itself is a compound of photo (from the Greek term, meaning light) and graphy (from the Greek term, meaning recording or writing). For the final term photogrammetry, the word metron (from the Greek term for a device to measure something) is added. Hence, photogrammetry can most simply be interpreted as a device or a technology that measures something using the recording of light. The American Society for Photogrammetry and Remote Sensing defines the term even more broadly as “the art, science and technology of obtaining reliable information about physical objects and the environment through processes of recording measuring and interpreting images and patterns of electromagnetic radiant energy and other phenomena” (American Society for Photogrammetry and Remote Sensing, 2019). Following this interpretation, other signals besides light waves (e.g., X-rays, radio waves) can be used in the process. This broad definition opens the door for the wide applications of photogrammetry, including archeology (Drap et al., 2007; Howland et al., 2014), architecture (Debevec, 1997; Agnello et al., 2019), topographic mapping (Fonstad et al., 2013), mining (Abbaszadeh and Rastiveis, 2017), construction (Tuttas et al., 2014), autonomous driving (Nguyen et al., 2019), detecting underground nuclear tests (Wang et al., 2018) and harvesting (Walter et al., 2018). Because of this broad definition and the widespread use, several different technologies have been subsumed under the term photogrammetry. This paper focuses on the most relevant approaches for education and educational videogame design with regard to applicability, accessibility, and required resources. Specifically, it discusses the widely available digital photography as the tool for measurement and the creation of PR3DMs as the desired outcome. This perspective excludes other technologies, such as laser scanning or satellite positioning, as well as large-scale environments, such as landscapes.
Photorealistic 3D Models
A basic definition of PR3DMs is needed to provide a common ground for the following discussion. This is especially important as computer graphics and visual representations in educational materials have evolved tremendously (e.g., Houghton and Willows(eds), 1987), and what is referred to as realistic might have changed over the last decades. Thus, a straightforward approach from computer graphics research is used. In this field, computer-generated images are referred to as realistic if they are indistinguishable from photographs (Rademacher et al., 2001). Thus, such images might also be called photorealistic. Similarly, PR3DMs can be characterized as digital models of real-world objects that appear indistinguishable from their physical counterparts, from a human observer’s perspective. In addition to the previous focus on photography, this definition includes observations from different angles and of different displays. This addition demarcates PR3DMs from other forms of digitization. For example, methods such as 360° photography are frequently used to generate photorealistic environments (e.g., Chamilothori et al., 2019). In contrast to PR3DMs, the content cannot be observed from any other angle as the viewpoint is fixed.
Technical Foundation
For the purpose of digitizing physical objects, photogrammetry has to measure and record two distinct attributes. First, the position of each surface point on the targeted object needs to be measured; second, the nature of each point (e.g., color, transparency, reflectance) should be determined. To assess the position, a series of pictures (at least two but typically dozens up to a few hundred) from different angles is needed to gather spatial information. Algorithms compare different image segments and their relative movements to each other from picture to picture (Gherardi et al., 2012; Toldo, 2013). Summarized, this process can be compared to the human perception of depth using the motion parallax, enabling humans to assess depth through the amount of movement in their visual field, with far-away objects moving at a lower rate than objects nearby (Rogers and Graham, 1979). From this, the relative position of each image, respectively, the camera parameters, are estimated (Figure 1). The figures depict one potential use-case for photogrammetry, the digitization of historical artifacts for museums and educational purposes. They are not intended as an endorsement of phrenology.
FIGURE 1

A screenshot showing how a PR3DM is created using a series of photographs. The blue cones represent the camera positions of photographs of a physical object (left), preliminary coordinates are calculated (right), screenshots are taken in 3DF Zephyr Lite (3Dflow, 2019), and arrows are added to indicate camera orientations.
Those parameters can be used to determine the 3D coordinates of each point of the desired object, which can be depicted as point clouds (Figure 2). These clouds are further densified, and a polygon mesh is generated (e.g., Glencross et al., 2008; for a detailed introduction, see Kraus, 2012, 2013; Toldo, 2013).
FIGURE 2

A sparse point cloud (top left), a dense point cloud (top right), a polygon mesh (bottom left), and a textured mesh (bottom right); screenshots taken in 3DF Zephyr Lite (3Dflow, 2019).
The resulting digital replication can be considered an exact digital duplicate of the physical object, with its accuracy depending on several variables, such as the source image quality, the number of included vertices, or the utilized algorithms. Subsequently, the object’s entire appearance is digitized, where textures are generated from the source images and mapped to the surface of the 3D mesh (Figure 3). Modern graphic engines enable advanced textures with dynamic features, such as transparencies or reflections (Figure 4). However, these effects typically need to be applied manually as they are difficult to detect automatically using the methods of photogrammetry (but see Melendez et al., 2012, for an approach of optimizing materials from photographs).
FIGURE 3

Mapping of the texture (left) to the 3D mesh (right); screenshot taken in Blender (Blender Foundation, 2019).
FIGURE 4

Rendering changes to recreate the original appearance: without changes (left), additional reflections to mimic the appearance of ceramics (middle), and a fictional metallic appearance (right); screenshot taken in Blender (Blender Foundation, 2019).
This technological approach has achieved results comparable to those of the previously established method of laser scanning (Kersten et al., 2015; Rehany et al., 2017). Additionally, the workflow is more efficient than traditional 3D modeling (Esmaeili et al., 2017), and the degree of realism can be further enhanced through post-processing (e.g., antialiasing, shadows, or depth of field; Rademacher et al., 2001; Sugano et al., 2003; Kan and Kaufmann, 2012). However, the complex process is subject to several limitations and technical considerations. For example, featureless surfaces (Nicolae et al., 2014) or moving objects might complicate the measuring process as they might provide insufficient or conflicting data. In low-light conditions that require a large aperture, some segments of the image might be blurred. Thus, a pixel cannot be used to measure a specific point on the physical object. This limits the achieved detail in the digitized duplicate. A low-light condition might also be challenging during the generation of textures since artificial amplifications of the measured light might cause artifacts that do not correspond to reality. More fundamentally, the process cannot digitize what cannot be measured with the used instrument – digital photography. This could be the case if an object is partially obstructed (e.g., by vegetation; Azzola et al., 2019), the needed perspective cannot be realized (e.g., capturing the inside of a handle; Tucci et al., 2012), or no suitable physical object is available. In such instances, researchers managed to utilize different technical solutions, such as the use of endoscopy or computer tomography (Fritsch et al., 2018). Finally, the process is resource demanding in terms of the necessary calculations. However, the capacities of current personal computers are sufficient to calculate the necessary steps within hours (depending on the complexity), and software solutions offering the required calculations have become increasingly available (e.g., 3Dflow, 2019; Capturing Reality, 2019).
Photogrammetry From the Perspective of Psychology
As the technical foundations and the produced outcomes become clear, a discussion about the learners’ perceptual processes might support the understanding of PR3DMs in educational contexts. Additionally, several subtopics in educational psychology, such as emotional or social design, outline the potential benefits of such realistic materials.
Perception and Transfer
If the perception and the transfer of spatial information about virtual objects and environments are discussed, the concept of fidelity is frequently utilized (e.g., Christel, 1994; Mania et al., 2006; Mourkoussis et al., 2010). Similar to the definition of PR3DMs, fidelity denotes the extent to which experiences in virtual environments are indistinguishable from the experiences in the real world (Waller et al., 1998). In more detail, Waller and colleagues differentiate between environmental and interface fidelity. The first describes the extent to which virtual representations resemble the real world. The latter refers to the extent to which the developed mental model is identical to the virtual environment. Naturally, deviations that might occur during the two associated transfer processes (from the real world to the virtual environment; from the virtual environment to the mental model) might induce differences between the initial object and the resulting cognitive representation. This might be especially problematic in an educational setting, as the goal might be to transfer the information as accurately as possible. To address this issue, researchers have tried to investigate different interfaces (e.g., ranging from screens to head-mounted displays). However, this research stream mostly addresses the aspect of interface fidelity (e.g., Makransky et al., 2019). In contrast, photogrammetry might be used to enhance environmental fidelity, which might be essential to ensure that the digital learning material was an adequate representation in the first place. By using photogrammetry, creators could avoid deviations that could occur during manual remodeling of physical objects. With countless measurements in the process, a much higher level of detail could also be achieved without a costly investment of work hours. The achieved geometry quality might be especially important if the objects are not yet established in mental schemata (Mourkoussis et al., 2010).
The use of PR3DMs could mitigate some negative effects that occur during the perception of virtual objects and the transfer of the associated information. For example, it is not a trivial task to infer the correct material from rendered visualizations (Vangorp et al., 2007) or to estimate the exact sizes of objects (Ahn et al., 2019) and correct distances in virtual environments (Smallman and St John, 2005; Kunz et al., 2009; El Jamiy and Marsh, 2019). Highly detailed PR3DMs could provide additional references that might facilitate these fundamental tasks. Additionally, low-texture details might be confused with the texture perspective (Cutting and Millard, 1984), resulting in the misperception that such objects are farther away than they really are (Drascic and Milgram, 1996). This potential harmful effect could be minimized with high-resolution textures that could be created during the process of photogrammetry. Furthermore, post-processing effects, such as blur or color tint, might be used to mimic the atmospheric perspective (El Jamiy and Marsh, 2019), further correcting the perception of spatiality. The associated sharpness of the virtual objects might further amplify the perceived realism of the objects (Cowan et al., 2015). However, their impacts on different information processing mechanisms need further validation to explore the effects on learning outcomes.
Cognition
The use of 3D models created by photogrammetry can significantly alter the cognitive processing involved in a learning task. To comprehend the potential benefits and drawbacks of using realistic 3D models, it is helpful to consider the mental processing of realistic imagery first. Skulmowski and Rey (2018) explain the perception of realistic learning materials with Biederman,1985,1987geon theory. Geons are thought to be the basic perceptual units used in segmenting and processing visual stimuli (Biederman, 1987). Primitive 3D shapes, such as cylinders, cubes, and cones, have been identified as such units (Biederman, 1985). Based on the geon theory, the perception of 3D models can be considered a process where parts of the retinal image are grouped together as primitive 3D shapes to understand the spatial arrangement of the perceived object or scene (Biederman, 1985).
This perceptual process may be hindered by complex surfaces (Norman et al., 2004) and can thus be a potential source of cognitive load (i.e., information processed in the working memory; Sweller et al., 1998). Being required to mentally disassemble an object comprising hundreds of geons may induce more cognitive load than being presented with the object’s simplified version that features only a few primitive components. Based on this idea that realistic details induce cognitive load, some studies have shown that more details may interfere with learning processes (e.g., Scheiter et al., 2009). Altogether, realism can be an important aspect of cognitive processing in learning scenarios, but there may be disadvantages due to the higher perceptual complexity that it entails. Hence, future research is needed to find new ways of optimizing learning with realistic visualizations by reducing unnecessary cognitive strains. Stull et al. (2009) provide an example of such a method by demonstrating that the anatomy of a complex virtual bone model that can be freely rotated is more effectively learned if rotational axes (presented as thin cylindrical shapes next to the bone) are included as references to keep track of the object’s rotation. Another strategy to enhance learning with realistic visualizations is to use color cues that help segment complex surfaces (Skulmowski and Rey, 2018).
Learning Theories and Effects
Learning theories include cognitive frameworks as well. Both the cognitive load theory (CLT; Sweller et al., 1998; Sweller et al., 2019) and the cognitive theory of multimedia learning (CTML; Mayer, 2014) assume the working memory’s limitation in processing multimedia learning messages. These theories regard the working memory as the central bottleneck in information processing and according to the CLT, can be inundated by various types of cognitive loads. In addition to the intrinsic cognitive load (ICL), which is determined by the information complexity and the learner’s knowledge, the extraneous cognitive load (ECL) is determined by the information presentation and the learner’s requirements due to the instructional procedure (e.g., Sweller et al., 2019).
According to these two theories of cognitive psychology, realistic images should generally lead to an increase in ECL and consequently impair the learning performance. Empirically, these negative effects were demonstrated decades before the emergence of the CLT and the CTML. For example, Dwyer (1969) found that realistic images partially reduced the learning performance in comparison to less realistic representations. Even in more recent research, more realistic representations are sometimes associated with reduced learning performance (Scheiter et al., 2009), although the comparability between schematic and realistic images does not always seem to be given. Similar effects could be identified in virtual environments (Lokka et al., 2018). The influence of realism also appears to be moderated by the learners’ visual-spatial abilities. For example, Brucker et al. (2014) found that learners with lower visual-spatial abilities achieved better recognition performance in schematic visualizations, while learners with higher visual-spatial abilities achieved better recognition performance in realistic visualizations.
The use of realistic representations in the multimedia learning context can also be associated with the seductive detail effect, defined as interesting but irrelevant or unimportant additions to the actual learning topic (e.g., Harp and Mayer, 1998). Thus, the inclusion of decorative, realistic images in accordance with the seductive detail effect may lead to impaired learning performance (Rey, 2012). However, PR3DMs can also improve the learning process, among other things, given the attractiveness of the visual stimuli. The degree of realism of images can also be considered in the context of the signaling effect. The signaling effect implies that deeper understanding processes in multimedia learning occur when cues direct the learners’ attention to relevant information or highlight the organizational structure of the core content (e.g., Mayer and Fiorella, 2014; van Gog, 2014). Two recent meta-analyses (Richter et al., 2016; Schneider et al., 2018a) confirm the signaling effect. Similar to landmarks (Waller and Lippa, 2007), realistic details could serve as signaling elements. In another experiment (Skulmowski and Rey, 2020), realistic instructional visualizations served as signaling and improved learner retention in comparison to schematic visualizations. At the same time, however, realistic representations also partially increased the ECL in the study. The authors describe this result pattern as the realism paradox (Skulmowski and Rey, 2020).
The imagination effect could also help in a better estimation of the effect of realism in representations. Imagination is understood as the learning-enhancing effect through mental visualization or mental rehearsal of concepts or processes (e.g., Sweller et al., 2019). For example, two experiments conducted by Leopold and Mayer (2015) showed that students who were asked to form images corresponding to each of the nine paragraphs about the human respiratory system achieved better retention and transfer performance than students who only read the computer-based text about the topic. Additionally, schematic, less realistic may more likely help learners (further) visualize concepts or processes than more realistic images that presumably offer less scope for imaginative elaboration of the visualization.
Emotional Design
Apart from these well-established theories and effects, photogrammetry-based environments can provide new opportunities to foster learner performance and to establish constructive learning environments. The more realistic such environments are, the better people can immerse and engage in them (Walton, 1984; Dede, 2009). However, a higher degree of realism can also entail an increase in the viewers’ emotional response (Waltemate et al., 2018) or motivation (Makransky and Lilleholt, 2018). For example, watching a rather abstract map of a concentration camp is less emotionally demanding than a realistic version. Perceptions and affective feelings are inherently combined and mutually dependent (Wormwood et al., 2019). According to the integrated cognitive-affective model of multimedia learning (Plass and Kaplan, 2015) and the emotional design hypothesis (Um et al., 2012), the design of digital learning media can have a major impact on the learners’ affective states, while these states influence the selection, organization, and integration of new information into a coherent mental model stored in the long-term memory. Specifically, physically attractive stimuli in such environments, such as realistic colors, have a positive effect on the learners’ emotional response and learning performance (e.g., Dion and Walster, 1972; Tractinsky et al., 2000; Wolfson and Case, 2000). Several studies have replicated this hypothesis in terms of different colors and shapes or the inclusion of decorative elements in learning materials (e.g., Plass et al., 2014; Schneider et al., 2018b). Positive emotions, such as enjoyment of learning, elicited through realistic images, direct attention toward the learning task and allow the full use of cognitive resources to achieve the instructional objectives (Pekrun et al., 2002; Huk and Ludwigs, 2009). Moreover, elicited positive emotions may trigger an increase in cognitive processing compared to negative emotional episodes (e.g., Isen et al., 1987; Nadler et al., 2010). Eye-tracking data show that more detailed images attract more attention in both the initial learning phase and the overall learning processes than simplified illustrations (Lin et al., 2017). As suggested by Schneider et al. (2016), appealing pictures may contribute to an aesthetically pleasing design of multimedia learning materials, although learners’ cognitive load may increase (see also Plass and Kalyuga, 2019). This issue is of particular importance since an emotional overload may exceed the learners’ cognitive capacities and shift their attention away from the learning activities (Plass and Kaplan, 2015).
Consequently, the learning-diminishing effects of negative emotions elicited through highly realistic presentations might also occur. In a more realistic virtual environment tested by Slater et al. (2009), the participants reported not only a higher subjective presence but also an increased level of stress. According to the uncanny valley effect (Mori et al., 2012), the affinity to almost realistic representations of humans (e.g., a prosthetic hand) is lower than the affinity to pictures with a more simplified or perfectly realistic presentation. Instead of producing a more positive experience with increasing realism per se, strong adverse reactions are shown very frequently as well, particularly when users face highly elaborate digital entities. Specifically, digital technologies that appear to approach uniquely human qualities provoke ontological ambiguity (Moore, 2012), leading to uncertainty on how to behave toward and what to expect from them (Langer and König, 2018). This uncertainty also creates a sense of threat to physical health or human distinctiveness (Ferrari et al., 2016). These negative feelings can impair learning through aversive behaviors. In contrast, even negative emotional states can foster learning, for example, by promoting analytical-sequential, detailed, and rigid ways of processing information (Ashby et al., 1999; Pekrun et al., 2009) or by encouraging a more creative performance under specific conditions (George and Zhou, 2002). Consequently, it can be assumed that simplistic conceptions, which classify negative emotions as detrimental and positive emotions as beneficial, do not entirely apply and should therefore be avoided (Pekrun et al., 2002).
Social Learning
As one of the modern technical advantages, photogrammetry is especially suitable for creating social entities or the so-called personae (Hartmann et al., 2004) in digital environments. In an educational context, these entities can be classified as pedagogical agents (Veletsianos and Russell, 2014), which are computer-generated characters that serve instructional purposes (e.g., demonstrating, providing help, instructing; Veletsianos and Russell, 2014; Martha and Santoso, 2019). According to the media equation theory (Reeves and Nass, 1996) and the social agent theory (Moreno and Mayer, 2000), computer-based events are interpreted as social events. Social cues in instructional media environments (e.g., a photorealistic pedagogical agent) prime a social activation schema in the same way that the social processes of human-to-human communication are triggered. Thus, deep cognitive processes, such as selecting, organizing, and integrating relevant verbal and visual information into a coherent mental representation, are enhanced because learners interpret the virtual agent not only as an abstract source of information but also as a partner in social communication (Mayer, 2001). Here, the possibilities and the potential of photogrammetry become clear. The more realistic features that can be implemented, the more social cues can be displayed adequately. Empirical support can be found in recent publications in the field of virtual agents. Realistic 3D characters can be more effective for learning than simple 2D agents (Schroeder et al., 2013). According to Heidig and Clarebout (2011), the design of the agent can strengthen involvement in learning, as long as current technologies are used for a convincingly realistic creation. For example, Beege et al. (2017) point out that older participants reach higher retention scores when the information is provided by an old agent and vice versa. Photogrammetry can be used to display realistic wrinkles or smooth skin on agents’ faces and may thus play an important role in the generation of pedagogical agents. Future research could investigate whether this degree of realism reinforces the results, or simple social cues (e.g., a cartoon animation) trigger comparable effect sizes. Since the social role design of pedagogical agents enhances students’ motivation and learning behavior due to sociocognitive processes (e.g., Kim and Baylor, 2007), the agent’s clothing is relevant if the agent addresses the learner directly (Beege et al., 2019). Photogrammetry can be used to create realistic clothing in order to provide credible social role models. Additionally, photogrammetry can help designers overcome the uncanny valley effect (see section “Emotional Design”; Mori et al., 2012). In this case, a more compelling human-like value could increase involvement in learning (Schroeder and Traxler, 2017). The presence of realistic and visually credible social entities can thus unfold positive effects on learning, satisfaction, and engagement in online learning environments (Wang and Antonenko, 2017).
Applications
The comparably easy creation of PR3DMs with photogrammetry opens new avenues for various existing educational approaches. However, in the fields of education and psychology, photogrammetry is only slowly starting to gain traction. The following existing and conceivable applications might be promising starting points for applications, potentially benefiting from the previously discussed psychological perspectives.
Preservation and Museums
Apart from scientific research (e.g., computer science) and specialized fields (e.g., cartography), a wider range of potential uses can be identified in preservation and education in museums (e.g., Kersten et al., 2017; Kersten, 2018; Agnello et al., 2019). In the field of preservation or archeology, the measurement with photography provides special benefits that lead to its frequent application. For example, visual data can be gathered from almost everywhere, from underwater (Drap et al., 2007) to distant planets (Blasius, 1973). Thus, PR3DMs from almost every type of object can be created and subsequently preserved. As photography itself has been established even longer than photogrammetry, image archives can also be used as important data sources (Goesele et al., 2007). Thus, the creation of PR3DMs is feasible from objects that might not even exist anymore or are inaccessible (Kyriakaki et al., 2014; Vincent et al., 2015; Curry, 2019; The Arc/k Project, 2019). These benefits apply to museums in their tasks of preserving and presenting exhibits. Digitization also represents an important strategy to increase the exhibitions’ visibility and accessibility (e.g., museu4mpunkt0, 2019). In Germany alone, over 600 institutions already digitally share some of their materials (museum-digital, 2019). However, sharing digitized exhibits might be a challenging and resource-intensive task, especially for smaller institutions. Whereas larger museums might be capable of using advanced technologies, such as laser scanning to create digital content (e.g., Staatliche Museen zu Berlin, 2019), smaller institutions might need simpler approaches, such as photogrammetry with digital photography. The resulting PR3DMs can also be used to compare different exhibits (Dall’Asta et al., 2016), support reconstruction (Bujakiewicz et al., 2006), or produce duplicates via 3D printing (Ballarin et al., 2018).
Generative Learning in Classrooms
In addition to merely viewing 3D models or scenes generated using photogrammetry, the generative aspect of photogrammetry in the context of instruction could be a major focus of future research and applications. Numerous empirical investigations have shown that students embrace 3D-printed materials as learning resources (e.g., Li et al., 2018; Mogali et al., 2018), particularly when they are involved in creating these models (Backhouse et al., 2019). In addition, such printed models could be accompanied by digital applications, such as augmented reality (Dewitz et al., 2019). Similarly, physical props can be used as useful learning metaphors (Astrachan, 1998; Bragdon et al., 2010), supporting knowledge acquisition and skill transfer. Likewise, photogrammetry provides an opportunity for learners to engage with an object during the process of creating a digitized model as a form of generative learning (for an overview, see Fiorella and Mayer, 2016). Instead of passively viewing and learning from prefabricated materials, studies have provided evidence for the claim that letting learners create materials themselves can enhance their learning performance (Fiorella and Kuhlmann, 2019; Hoogerheide et al., 2019). However, some studies indicate that self-generating learning materials may produce unwanted cognitive load and can thus impede learning (Stull and Mayer, 2007). It is essential that photogrammetry applications become user friendly enough so that learners can generate their own 3D models with little or no supervision while being able to focus on the object and the model rather than on the demands imposed by the user interface. Additionally, pretraining could enhance learning outcomes, similar to the effect of such interventions on the learning gains using virtual reality (Meyer et al., 2019).
Research on Educational Videogames
Digital content creation can be a major challenge for videogame researchers because not every psychologist or educator is a skilled game designer and vice versa. This challenge holds true in the field of educational videogames. Researchers have to investigate learning mechanisms in a complex medium that potentially comprises a wide range of virtual environments and objects. Experimental research regularly follows a case-study approach with finished products to avoid complex content creation, as this significantly reduces the technical challenge. However, researchers state that a more detailed investigation with variations of individual facets is needed to validate different design approaches (Clark et al., 2016; Ke, 2016; Nebel, 2017). For workarounds, researchers have used games that are highly modifiable (Loh and Byun, 2009), include simplified production tools (Williams-Pierce, 2016), or focus on gameplay where the players have to create the content themselves (Nebel et al., 2016). These approaches reach their limits if the included assets are insufficient to realize the desired learning outcome or answer the research question. To fill this gap, the described process of photogrammetry could be used to create the necessary content, especially in combination with increasingly accessible game engines (e.g., Epic Games, 2018). This might empower the research on many topics that require modifications of the included virtual objects, particularly concerning the question of how realistic visualizations might influence the learning process. This topic is gaining importance because a substantial difference remains between PR3DMs and what has been considered realistic in meta-analytical investigations on educational videogames (Wouters et al., 2013; Clark et al., 2016).
Future of Photogrammetry in Education
The previous sections have highlighted the creation, processing, and applications of PR3DMs created via photogrammetry. However, to establish this approach, the research gaps need to be filled, technological skills should be acquired by instructors, and unique educational potentials ought to be discussed.
Lack of Research
A substantial lack of research suggests a major challenge for the use of photogrammetry in education. In contrast to other fields, such as cartography (Wilkening and Fabrikant, 2011), landscape visualization (Lange, 2001), virtual avatars (Latoschik et al., 2017), cognition (Lokka and Çöltekin, 2016), or the concept of presence (Cummings and Bailenson, 2016; van Gisbergen et al., 2019), research on realistic virtual objects in the field of instructional psychology is comparably scarce. Another challenge arises from the questionable transferability of the insights gained from related fields. Previous research on realism has most frequently used stimuli that noticeably differ from photorealistic materials, and these studies have often focused on fundamental processes, such as navigation or perception (e.g., Lokka and Çöltekin, 2017). Research that includes educational content is conducted less frequently. Therefore, it is not clear if research claiming positive effects on performance (Lessels and Ruddle, 2005; Slater et al., 1996) and on behavioral variables (Slater et al., 2009) or offering inconsistent insights (e.g., Hoffman et al., 2001; Zimmons and Panter, 2003; Mania et al., 2006; Yu et al., 2012; Lee et al., 2013) can be transferred to educational settings. Additionally, variables are often combined (e.g., shadows and textures; van Gisbergen et al., 2019), significantly increasing the challenge for researchers to infer specific design recommendations. Overall, this lack of dedicated empirical evidence limits the applications of PR3DMs in education because many considerations regarding the realism of virtual objects in educational settings remain speculative until empirically validated. This caution should be applied to the approaches and the considerations proposed in see section “Photogrammetry From the Perspective of Psychology,” such as emotional design or social aspects using PR3DMs.
Technical Developments and Accessability
As previously stated, the technical and mathematical foundations of photogrammetry are not entirely new, but the development in creating digitized duplicates has reached a crucial point, where even technical novices using non-professional hardware can accomplish tremendous accuracy (Figure 5). In particular, 3D creation processes, such as texture mapping, can efficiently compensate for insufficient geometric details (Rushmeier et al., 2000). The required software is also becoming more widespread (e.g., 3Dflow, 2019; Capturing Reality, 2019), and more educators can explore this approach without substantial technical or financial investments. Photogrammetry can even enable the creation of PR3DMs using a web-based image or video search (Weis, 2019).
FIGURE 5

A comparison of the physical object (left segment of the images) and the PR3DM (right segment of the images); no further enhancements (e.g., reflections) are used for this PR3DM rendering. Screenshots are taken in 3DF Zephyr Lite (3Dflow, 2019).
The process described in Section “Technical Foundation” indicates some subprocesses that need additional attention and are not yet fully automated in the available software. For example, the outlined phenomena that are difficult to measure with photography (e.g., reflections, transparencies) need to be applied with additional editing tools. Although available for free (e.g., Blender Foundation, 2019), such software needs considerable time to master. Similarly, the optimization of the mesh (e.g., simplifying, closing, and cleaning models) or the textures (e.g., color correction, manipulation of advanced features, such as normal mapping) can be achieved in dedicated software (e.g., Cignoni et al., 2008) that might require additional time to learn. However, it should be noted that these processes might not be necessary for creating typical PR3DMs, depending on the material and the complexity of the physical object. Altogether, further documentation of workflows (e.g., Fritsch, 2003; Orr et al., 2020) and deeper discussions about potentials, with a clear focus on educational settings, are needed.
Impact on Behavioral and Emotional Outcomes
Case studies focusing on the measuring aspect of the process might represent the majority of the existing experimental research using photogrammetry. If the learning material is used in virtual environments, factual knowledge is most commonly taught. This can be classified as basic knowledge (Bloom and Krathwohl, 1956) or as a low-level element of the information knowledge domain (Marzano and Kendall, 2007). For example, in medical studies, anatomical knowledge is tested (e.g., Nicholson et al., 2006). However, there is a lack of research on what perception of and interaction with PR3DMs might offer for different contexts. For example, learning settings relying on 3D models or immersive experiences created using photogrammetry may result in more realistic behavior (e.g., Dede, 2009) and could thus be more effective for a variety of training environments. Media psychology research has begun to investigate such effects of different types of virtual environments on emotional or behavioral variables (Zaalberg and Midden, 2010; Chirico and Gaggioli, 2019). In this vein, virtual environments are used for educational purposes that cannot be easily classified using established learning taxonomies, such as training in public speaking (Poeschl, 2017), controlling anxiety disorders (Krijn et al., 2004), treating post-traumatic stress disorder (Botella et al., 2015), managing fears and phobias (Hodges et al., 1995), or decreasing prejudice against specific groups (Stelzmann et al., 2019). It is not yet clear how PR3DMs could support such approaches. For example, would the impact of a virtual environment trying to illustrate the dangers of being a refugee be strengthened by using PR3DMs? Following the learning taxonomy of Merrill and Twitchell (1994), this could impact the emotional component as learners might react differently to related stimuli after interacting with the PR3DMs. This is in line with the discussion about naïve realism (Smallman and St John, 2005) and similar statements that realism has to serve a clear purpose (Paes and Irizarry, 2019), as even unrealistic representations can be sufficient to be correctly visually identified and processed (Mourkoussis et al., 2010). Therefore, the real potential of photogrammetry might not be observable if simple factual knowledge is the key component (see the following section for a discussion of Dwyer, 1969). Instead, in line with the presented psychological perspective, PR3DMs might be more useful if higher level learning outcomes (Bloom and Krathwohl, 1956; Marzano and Kendall, 2007) or emotional (Merrill and Twitchell, 1994) and behavioral outcomes are targeted. To shed more light on this, the creation of a taxonomy of learning gains through the use of PR3DMs might be an important task in the future, as soon as sufficient empirical studies are available.
Addressing Specific Learning Outcomes
We wish to emphasize that photogrammetry (and realism in general) is no cure-all for education. As we have summarized, there can be both positive and negative effects when using photorealistic visuals in educational settings. From older studies by Dwyer, we can learn that the benefits of realism depend on the type of knowledge in question. For instance, Dwyer (1969) found that more realistic visualizations in an anatomy learning task result in better results for learning tests that involve the visual recognition of anatomical elements, while the performance in learning tests that do not require fine-grained visual knowledge, such as vocabulary tests, can profit from the simplicity of line drawings (or even the absence of any visuals). As a result, more consideration should be given to the type of knowledge that educators wish to foster in learners. Although there are instances in which knowledge should be simplified to allow learners to grasp the essential information, there are several scenarios in which a high level of realism has been shown to be advantageous for learning as described in see section “Photogrammetry From the Perspective of Psychology.” For these scenarios, photogrammetry is a straight-forward, low-cost, and widely available technique that provides a resolution comparable to more expensive methods of digitization and thus empowers instructional designers to integrate cutting-edge visuals into their learning content. In summary, photorealism is not always desirable in learning materials, but if the exact reproduction of objects for learning is an important component of a learning task or might even enable new learning outcomes, photogrammetry currently is one of the best ways to generate PR3DMs.
Conclusion
Overall, this paper has illustrated a technological approach capable of producing photorealistic virtual objects and environments. Although not as established as other forms of learning with multimedia materials (e.g., images, Schneider et al., 2016; videos, Hoogerheide et al., 2019), the technology might reach a similar point in the near future. To foster this development, researchers should explore the potentials and the open questions discussed in this paper. Thus, the presented psychological perspective provides a crucial theoretical and empirical structure to explore the most beneficial applications of PR3DMs. This perspective could be extended by other perspectives, such as the broader field of computer graphics or skill transfer research, to gradually create a complete overview of the potential and challenges while working with PR3DMs. The outlined theoretical and practical considerations can be used to inspire more nuanced research, exploring the interaction between the effects on mental load through potentially more complex cognitive processing, potentially increased emotional or behavioral outcomes, and how such material could be connected to traditional or additional learning material (Coughenour et al., 2015). Research itself could benefit from the facilitated method of content production, and researchers addressing realistic digital materials might be capable of using PR3DMs as stimuli. Finally, this review has addressed over 50 years of research in diverse fields. The standard of realism achieved with photogrammetry was probably unimaginable in the earlier works. Thus, insights and theoretical considerations should be updated to reflect the current technical possibilities. Naturally, this presents a continuous (and stimulating) task for researchers.
Statements
Author contributions
SN wrote major parts of the manuscript. MB, SS, and GR wrote segments of the psychological perspective and proofreading. All authors contributed to the article and approved the submitted version.
Funding
The publication of this article was funded by Chemnitz University of Technology.
Acknowledgments
We want to thank Alexander Skulmowski for his input and criticism. This was very useful to us to improve the manuscript.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
1
3Dflow (2019). 3DF Zephyr. Available online at: https://www.3dflow.net/(accessed January 28, 2020).
2
AbbaszadehS.RastiveisH. (2017). “A comparison of close-range photogrammetry using a non-professional camerawith field surveying for vplume estimation,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-4/W4, (Heipke: International Society for Photogrammetry and Remote Sensing), 1–4. 10.5194/isprs-archives-XLII-4-W4-1-2017
3
AgnelloF.AvellaF.AgnelloS. (2019). “Virtual reality for historical architecture,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-2/W9, (Heipke: International Society for Photogrammetry and Remote Sensing), 9–16. 10.5194/isprs-archives-XLII-2-W9-9-2019
4
AhnJ.AhnE.MinS.ChoiH.KimH.KimG. J. (2019). “Size perception of augmented objects by different AR displays,” in HCI International 2019—Posters, Vol. 1033ed.StephanidisC. (London: Intech Open), 337–344. 10.1007/978-3-030-23528-4_46
5
AidanJ. (2014). Review: The Vanishing of Ethan Carter. Available online at: https://www.hardcoregamer.com/2014/09/26/review-the-vanishing-of-ethan-carter/107846/(accessed November 5, 2019)
6
American Society for Photogrammetry and Remote Sensing (2019). What is ASPRS?. Available online at: https://www.asprs.org/organization/what-is-asprs.html(accessed November 27, 2019)
7
AshbyF. G.IsenA. M.TurkenA. U. (1999). A neuropsychological theory of positive affect and its influence on cognition.Psychol. Rev.106529–550. 10.1037/0033-295x.106.3.529
8
AstrachanO. (1998). “Concrete teaching: hooks and props as instructional technology,” in Proceedings of the 6th Annual Conference on the Teaching of Computing and the 3rd Annual Conference on Integrating Technology into Computer Science Education: Changing the Delivery of Computer Science Education, (New York, NY: Association for Computing Machinery).
9
AzzolaP.CardaciA.Mirabella RobertiG.NanneiV. M. (2019). “UAV photogrammetry for cultural heritage preservation modeling and mapping venetian walls of bergamo,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-2/W9, ( Heipke: International Society for Photogrammetry and Remote Sensing), 45–50. 10.5194/isprs-archives-XLII-2-W9-45-2019
10
BackhouseS.TaylorD.ArmitageJ. A. (2019). Is this mine to keep? Three-dimensional printing enables active, personalized learning in anatomy.Anat. Sci. Educ.12518–528. 10.1002/ase.1840
11
BallarinM.BallettiC.VernierP. (2018). “Replicas in cultural heritage: 3d printing and the museum experience,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII–2, ( Heipke: International Society for Photogrammetry and Remote Sensing), 55–62. 10.5194/isprs-archives-XLII-2-55-2018
12
BeegeM.NebelS.SchneiderS.ReyG. D. (2019). Social entities in educational videos: combining the effects of addressing and professionalism.Comput. Hum. Behav.9340–52. 10.1016/j.chb.2018.11.051
13
BeegeM.SchneiderS.NebelS.MittangkJ.ReyG. D. (2017). Ageism – Age coherence within learning material fosters learning.Comput. Hum. Behav.75510–519. 10.1016/j.chb.2017.05.042
14
BiedermanI. (1985). Human image understanding: Recent research and a theory.Comput. Vis. Graph. Image Process.3229–73. 10.1016/0734-189X(85)90002-7
15
BiedermanI. (1987). Recognition-by-components: a theory of human image understanding.Psychol. Rev.94115–147. 10.1037/0033-295X.94.2.115
16
BlasiusK. R. (1973). A study of Martian topography by analytic photogrammetry.J. Geophys. Res.204411–4423. 10.1029/jb078i020p04411
17
Blender Foundation (2019). Blender (Version 2.81a). Available online at: https://www.blender.org/(accessed January 28, 2020).
18
BlinnJ. F. (1978). “Simulation of wrinkled surfaces,” in Proceedings of the 5th Annual Conference on Computer Graphics and Interactive Techniques, (New York, NY: Association for Computing Machinery), 286–292. 10.1145/800248.507101
19
BloomB. S.KrathwohlD. R. (1956). “Taxonomy of educational objectives,” in The Classification of Educational Goals, Handbook I: Cognitive domain, (Harlow: Longmans Green).
20
BotellaC.SerranoB.BañosR. M.Garcia-PalaciosA. (2015). Virtual reality exposure-based therapy for the treatment of post-traumatic stress disorder: a review of its efficacy, the adequacy of the treatment protocol, and its acceptability.Neuropsychiatr. Dis. Treat.112533–2545. 10.2147/NDT.S89542
21
BragdonA.UgurayA.WigdorD.AnagnostopoulosS.ZeleznikR.FemanR. (2010). “Gesture play: motivating online gesture learning with fun, positive reinforcement and physical metaphors,” in Proceedings of the ACM International Conference on Interactive Tabletops and Surfaces - ITS ’10, (New York, NY: Association for Computing Machinery), 10.1145/1936652.1936661
22
BruckerB.ScheiterK.GerjetsP. (2014). Learning with dynamic and static visualizations: realistic details only benefit learners with high visuospatial abilities.Comput. Hum. Behav.36330–339. 10.1016/j.chb.2014.03.077
23
BujakiewiczA.KowalczykM.PodlasiakP.ZawieskaD. (2006). 3D reconstruction and modelling of the contact surfaces for the archaeological small museum pieces.IAPRS Volume XXXVI2256–61. 10.1111/j.1477-9730.2007.00418.x
24
Capturing Reality (2019). RealityCapture. Available online at: https://www.capturingreality.com/(accessed January 28, 2020).
25
ChamilothoriK.WienoldJ.AndersenM. (2019). Adequacy of immersive virtual reality for the perception of daylit spaces: comparison of real and virtual environments.LEUKOS15203–226. 10.1080/15502724.2017.1404918
26
ChiricoA.GaggioliA. (2019). When virtual feels real: comparing emotional responses and presence in virtual and natural environments.Cyberpsychol. Behav. Soc. Netw.22220–226. 10.1089/cyber.2018.0393
27
ChristelM. G. (1994). The role of visual fidelity in computer-based instruction.Hum. Comput. Interact.2183–223. 10.1207/s15327051hci0902_2
28
CignoniP.CallieriM.CorsiniM.DellepianeM.GanovelliF.RanzugliaG. (2008). in MeshLab: An Open-Source Mesh Processing Tool, edsScaranoV.ChiaraR. D.ErraU.Aire-la-Ville: The Eurographics Association.
29
ClarkD. B.Tanner-SmithE. E.KillingsworthS. S. (2016). Digital games, design, and learning: a systematic review and meta-analysis.Rev. Educ. Res.8679–122. 10.3102/0034654315582065
30
CoughenourC. M.VincentM. L.de KramerM.SenecalS.FritschD.Flores GutirrezM.et al (2015). “Embedding knowledge in 3D data frameworks in cultural heritage,” in ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences, II-5/W3, ( Heipke: International Society for Photogrammetry and Remote Sensing), 47–52. 10.5194/isprsannals-II-5-W3-47-2015
31
CowanB.RojasD.KapralosB.MoussaF.DubrowskiA. (2015). Effects of sound on visual realism perception and task performance.Vis. Comput.311207–1216. 10.1007/s00371-014-1006-6
32
CummingsJ. J.BailensonJ. N. (2016). How immersive is enough? A meta-analysis of the effect of immersive technology on user presence.Media Psychol.19272–309. 10.1080/15213269.2015.1015740
33
CurryA. (2019). Virtual copy of ransacked museum comes to Mosul.Science363:573. 10.1126/science.363.6427.573
34
CuttingJ. E.MillardR. T. (1984). Three gradients and the perception of flat and curved surfaces.J. Exp. Psychol. Gen.113198–216. 10.1037/0096-3445.113.2.198
35
Dall’AstaE.BrunoN.BigliardiG.ZerbiA.RoncellaR. (2016). Photogrammetric Techniques For Promotion Of Archaeological Heritage: The Archaeological Museum Of Parma (ITALY), Vol. XLI-B5 (Prague: The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences), 12–19.
36
StelzmannD.TothR.SchieferdeckerD. (2019). “Encounters in VR: decreasing prejudice towards schizophrenia people,” in Proceedings of the Media Psychology Division, edsNebelS.PietschmannD.SchneiderS.KobanK.BeegeM.SkulmowskiA.et al (Chemnitz: University Press Chemnitz), 69–71.
37
DebevecP. E. (1997). “FACADE: modeling and rendering architecture from photographs and the campanile model,” in ACM SIGGRAPH 97 Visual Proceedings: The Art and Interdisciplinary Programs of SIGGRAPH’97, (New York, NY: Association for Computing Machinery), 254. 10.1145/259081.259366
38
DedeC. (2009). Immersive interfaces for engagement and learning.Science32366–69. 10.1126/science.1167311
39
DewitzL.KröberC.MessemerH.MaiwaldF.MünsterS.BruschkeJ.et al (2019). “Historical photos and visualizations: potential for research,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-2/W15, (Heipke: International Society for Photogrammetry and Remote Sensing), 405–412. 10.5194/isprs-archives-XLII-2-W15-405-2019
40
DionK.WalsterE. (1972). What is beautiful is good.J. Pers. Soc. Psychol.24:6. 10.1037/h0033731
41
DrapP.SeinturierJ.ScaradozziD.GambogiP.LongL.GauchF. (2007). Photogrammetry For Virtual Exploration Of Underwater Archeological Sites.Athens: Congenital Insensitivity to Pain with Anhidrosis, 7.
42
DrascicD.MilgramP. (1996). in Perceptual Issues in Augmented Reality, edsBolasM. T.FisherS. S.MerrittJ. O. (San Jose, CA: SPIE), 10.1117/12.237425
43
DwyerF. M. (1969). The effect of varying the amount of realistic detail in visual illustrations designed to complement programmed instruction.Program. Learn. Educ. Technol.6147–153. 10.1080/1355800690060301
44
El JamiyF.MarshR. (2019). Survey on depth perception in head mounted displays: distance estimation in virtual reality, augmented reality, and mixed reality.IET Image Process.13707–712. 10.1049/iet-ipr.2018.5920
45
Epic Games (2018). Unreal Engine 4. Available online at: https://www.unrealengine.com/(accessed January 28, 2020).
46
EsmaeiliH.ThwaitesH.WoodsP. C. (2017). “Workflows and challenges involved in creation of realistic immersive virtual museum, heritage, and tourism experiences: a comprehensive reference for 3D asset capturing,” in Proceedings of the 2017 13th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), (Piscataway, NJ: IEEE), 465–472. 10.1109/SITIS.2017.82
47
FerrariF.PaladinoM. P.JettenJ. (2016). Blurring human–machine distinctions: anthropomorphic appearance in social robots as a threat to human distinctiveness.Int. J. Soc. Robot.8287–302. 10.1007/s12369-016-0338-y
48
FiorellaL.MayerR. E. (2016). Eight ways to promote generative learning.Educ. Psychol. Rev.28717–741. 10.1007/s10648-015-9348-9
49
FiorellaL.KuhlmannS. (2019). Creating drawings enhances learning by teaching.J. Educ. Psychol.112811–822. 10.1037/edu0000392
50
FonstadM. A.DietrichJ. T.CourvilleB. C.JensenJ. L.CarbonneauP. E. (2013). Topographic structure from motion: a new development in photogrammetric measurement.Earth Surf. Process. Landforms38421–430. 10.1002/esp.3366
51
FritschD. (2003). “3D building visualisation – outdoor and indoor applications,” in Photogrammetric Week ’03, ed.GuganD. (Heidelberg: Wichmann Verlag), 281–290.
52
FritschD.WagnerJ. F.SimonS.CeranskiB.NiklausM.ZhanK.et al (2018). “Gyrolog—towards VR preservations of gyro instruments for historical and didactical research,” in 2018 Pacific Neighborhood Consortium Annual Conference and Joint Meetings (PNC), (Piscataway, NJ: IEEE), 1–7. 10.23919/PNC.2018.8579456
53
GeorgeJ. M.ZhouJ. (2002). Understanding when bad moods foster creativity and good ones don’t: The role of context and clarity of feelings.J. Appl. Psychol.87687–697. 10.1037/0021-9010.87.4.687
54
GherardiR.ToldoR.GarroV.FusielloA. (2012). “Automatic camera orientation and structure recovery with samantha,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XXXVIII-5/W16, (Heipke: International Society for Photogrammetry and Remote Sensing), 261–268. 10.5194/isprsarchives-XXXVIII-5-W16-261-2011
55
GlencrossM.WardG. J.MelendezF.JayC.LiuJ.HubboldR. (2008). A perceptually validated model for surface depth hallucination.ACM Trans. Graph. (TOG)271–8. 10.1145/1360612.1360658
56
GoeseleM.SnavelyN.CurlessB.HoppeH.SeitzS. M. (2007). “Multi-view stereo for community photo collections,” in Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, (Rio de Janeiro: IEEE), 1–8. 10.1109/ICCV.2007.4408933
57
HallT.RomeroJ.CarmackA.CarmackJ. (1992). Wolfenstein 3D. id Software. (Shreveport, Louisiana: Apogee Software, FormGen)
58
HarpS. F.MayerR. E. (1998). How seductive details do their damage: a theory of cognitive interest in science learning.J. Educ. Psychol.90414–434. 10.1037/0022-0663.90.3.414
59
HartmannT.SchrammH.KlimmtC. (2004). Personenorientierte medienrezeption: Ein Zwei-Ebenen-modell parasozialer Interaktionen.Publizistik49:25. 10.1007/s11616-004-0003-6
60
HeidigS.ClareboutG. (2011). Do pedagogical agents make a difference to student motivation and learning?Educ. Research Rev.627–54. 10.1016/j.edurev.2010.07.004
61
HodgesL. F.KooperR.MeyerT. C.RothbaumB. O.OpdykeD.de GraaffJ. J.et al (1995). Virtual environments for treating the fear of heights.Computer2827–34. 10.1109/2.391038
62
HoffmanH. G.Garcia-PalaciosA.ThomasA. K.SchmidtA. (2001). Virtual reality monitoring: phenomenal characteristics of real, virtual, and false memories.CyberPsychol. Behav.4565–572. 10.1089/109493101753235151
63
HoogerheideV.ViseeJ.LachnerA.van GogT. (2019). Generating an instructional video as homework activity is both effective and enjoyable.Learn. Instr.64:101226. 10.1016/j.learninstruc.2019.101226
64
HoughtonH. A.WillowsD. M. (eds) (1987). The Psychology of Illustration. Volume 2: Instructional Issues, Vol. 2. New York, NY: Springer-Verlag.
65
HowlandM. D.KuesterF.LevyT. E. (2014). Photogrammetry in the field: documenting, recording, and presenting archaeology.Mediter. Archaeol. Archaeom.4101–108.
66
HukT.LudwigsS. (2009). Combining cognitive and affective support in order to promote learning.Learn. Instr.19495–505. 10.1016/j.learninstruc.2008.09.001
67
IsenA. M.DaubmanK. A.NowickiG. P. (1987). Positive affect facilitates creative problem solving.J. Pers. Soc. Psychol.521122–1131. 10.1037/0022-3514.52.6.1122
68
KanP.KaufmannH. (2012). “High-quality reflections, refractions, and caustics in Augmented Reality and their contribution to visual coherence,” in Proceedings of the 2012 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), (Atlanta, GA: IEEE), 99–108. 10.1109/ISMAR.2012.6402546
69
KeF. (2016). Designing and integrating purposeful learning in game play: a systematic review.Educ. Technol. Res. Dev.64219–244. 10.1007/s11423-015-9418-1
70
KerstenT. P. (2018). “3D models and virtual tours for a museum exhibition of vietnamese cultural heritage exhibits and sites,” in Digital Heritage. Progress in Cultural Heritage: Documentation, Preservation, and Protection, Vol. 11196edsIoannidesM.FinkE.BrumanaR.PatiasP.DoulamisA.MartinsJ.et al (Cham: Springer), 528–538. 10.1007/978-3-030-01762-0_46
71
KerstenT. P.MechelkeK.MaziullL. (2015). “3D model of al zubarah fortress in qatar - terrestrial laser scanning vs. dense image matching,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XL-5/W4, (Heipke: International Society for Photogrammetry and Remote Sensing), 1–8. 10.5194/isprsarchives-XL-5-W4-1-2015
72
KerstenT. P.TschirschwitzF.DeggimS. (2017). “Development of a virtual museum including a 4d presentation of building history in virtual reality,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLII-2/W3, (Heipke: International Society for Photogrammetry and Remote Sensing), 361–367. 10.5194/isprs-archives-XLII-2-W3-361-2017
73
KimY.BaylorA. L. (2007). Pedagogical agents as social models to influence learner attitudes.Educ. Technol.4723–28.
74
KyriakakiG.DoulamisA.DoulamisN.IoannidesM.MakantasisK.ProtopapadakisE.et al (2014). 4D reconstruction of tangible cultural heritage objects from web-retrieved images.Int. J. Herit. Digital Era3431–451. 10.1260/2047-4970.3.2.431
75
KrausK. (2012). Photogrammetrie, Geometrische Informationen aus Photographien und Laserscanneraufnahmen, 7th Edn, Vol. 1. Berlin: De Gruyter, 10.1515/9783110908039
76
KrausK. (2013). Photogrammetrie, Verfeinerte Methoden und Anwendungen, 3rd Edn, Vol. 2. Berlin: De Gruyter, 10.1515/9783110906158
77
KrijnM.EmmelkampP. M. G.OlafssonR. P.BiemondR. (2004). Virtual reality exposure therapy of anxiety disorders: a review. Clin. Psychol. Rev.24259–281. 10.1016/j.cpr.2004.04.001
78
KunzB. R.WoutersL.SmithD.ThompsonW. B.Creem-RegehrS. H. (2009). Revisiting the effect of quality of graphics on distance judgments in virtual environments: a comparison of verbal reports and blind walking.Atten. Percept. Psychophys.711284–1293. 10.3758/APP.71.6.1284
79
LangeE. (2001). The limits of realism: perceptions of virtual landscapes.Landsc. Urban Plan.54163–182. 10.1016/S0169-2046(01)00134-7
80
LangerM.KönigC. J. (2018). Introducing and testing the creepiness of situation scale (CRoSS).Front. Psychol.9:2220. 10.3389/fpsyg.2018.02220
81
LatoschikM. E.RothD.GallD.AchenbachJ.WaltemateT.BotschM. (2017). “The effect of avatar realism in immersive social virtual realities,” in Proceedings of the 23rd ACM Symposium on Virtual Reality Software and Technology, Vol. 39 (New York, NY: Association for Computing Machinery), 1–10. 10.1145/3139131.3139156
82
LeeC.RinconG. A.MeyerG.HollererT.BowmanD. A. (2013). The effects of visual realism on search tasks in mixed reality simulation.IEEE Trans. Vis. Comput. Graph.19547–556. 10.1109/TVCG.2013.41
83
LeopoldC.MayerR. E. (2015). An imagination effect in learning from scientific text.J. Educ. Psychol.10747–63. 10.1037/a0037142
84
LesselsS.RuddleR. A. (2005). Movement around real and virtual cluttered environments.Presence14580–596. 10.1162/105474605774918778
85
LiF.LiuC.SongX.HuanY.GaoS.JiangZ. (2018). Production of accurate skeletal models of domestic animals using three-dimensional scanning and printing technology.Anat. Sci. Educ.1173–80. 10.1002/ase.1725
86
LinY. Y.HolmqvistK.MiyoshiK.AshidaH. (2017). Effects of detailed illustrations on science learning: an eye-tracking study.Instr. Sci.45557–581. 10.1007/s11251-017-9417-1
87
LohC. S.ByunJ. H. (2009). “Modding neverwinter nights into serious games,” in Digital Simulations for Improving Education: Learning Trough Artificial Teaching Environments, edsGibsonD.BaekY. (New York, NY: Information Science Reference), 408–426. 10.4018/978-1-60566-322-7.ch022
88
LokkaI. E.ÇöltekinA. (2016). “Simulating navigation with virtual 3d geovisualizations – a focus on memory related factors,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLI-B2, (Heipke: International Society for Photogrammetry and Remote Sensing), 671–673. 10.5194/isprsarchives-XLI-B2-671-2016
89
LokkaI. E.ÇöltekinA. (2017). Toward optimizing the design of virtual environments for route learning: Empirically assessing the effects of changing levels of realism on memory.Int. J. Digit. Earth12137–155. 10.1080/17538947.2017.1349842
90
LokkaI. E.ÇöltekinA.WienerJ.FabrikantS. I.RöckeC. (2018). Virtual environments as memory training devices in navigational tasks for older adults.Sci. Rep.8:10809. 10.1038/s41598-018-29029-x
91
MakranskyG.LilleholtL. (2018). A structural equation modeling investigation of the emotional value of immersive virtual reality in education.Educ. Technol. Res. Dev.661141–1164. 10.1007/s11423-018-9581-2
92
MakranskyG.TerkildsenT. S.MayerR. E. (2019). Adding immersive virtual reality to a science lab simulation causes more presence but less learning.Learn. Instr.60225–236. 10.1016/j.learninstruc.2017.12.007
93
ManiaK.WooldridgeD.CoxonM.RobinsonA. (2006). The effect of visual and interaction fidelity on spatial cognition in immersive virtual environments.IEEE Trans. Vis. Comput. Graph.12396–404. 10.1109/TVCG.2006.55
94
MarthaA. S. D.SantosoH. B. (2019). The design and impact of the pedagogical agent: a systematic literature review.J. Educ. Online16:15.
95
MarzanoR. J.KendallJ. S. (2007). The New Taxonomy of Educational Objectives.Thousand Oaks, CA: Corwin press.
96
MayerR. E. (2001). Multimedia Learning.Cambridge: Cambridge University Press, 10.1017/CBO9781139164603
97
MayerR. E. (2014). “Cognitive theory of multimedia learning,” in The Cambridge Handbook of Multimedia Learning, 2nd Edn, ed.MayerR. (Cambridge: Cambridge University Press), 43–71. 10.1017/CBO9781139547369.005
98
MayerR. E.FiorellaL. (2014). “Principles for reducing extraneous processing in multimedia learning: coherence, signaling, redundancy, spatial contiguity, and temporal contiguity principles,” in The Cambridge Handbook of Multimedia Learning, 2nd Edn, ed.MayerR. (Cambridge: Cambridge University Press), 279–315. 10.1017/CBO9781139547369.015
99
MelendezF.GlencrossM.StarckJ.WardG. J. (2012). “Transfer of albedo and local depth variation to photo-textures,” in Proceedings of the 9th European Conference on Visual Media Production, (New York, NY: Association for Computing Machinery), 40–48. 10.1145/2414688.2414694
100
MerrillM. D.TwitchellD. (1994). Instructional Design Theory.Englewood Cliffs, NJ: Educational Technology.
101
MeydenbauerA. (1867). Die Photometrographie. Wochenblatt Des Architektenvereins Zu Berlin1126–127.
102
MeyerO. A.OmdahlM. K.MakranskyG. (2019). Investigating the effect of pre-training when learning through immersive virtual reality and video: a media and methods experiment.Comput. Educ.140:103603. 10.1016/j.compedu.2019.103603
103
MogaliS. R.YeongW. Y.TanH. K. J.TanG. J. S.AbrahamsP. H.ZaryN.et al (2018). Evaluation by medical students of the educational value of multi-material and multi-colored three-dimensional printed models of the upper limb for anatomical education.Anat. Sci. Educ.1154–64. 10.1002/ase.1703
104
MooreR. K. (2012). A Bayesian explanation of the ‘Uncanny Valley’ effect and related psychological phenomena.Sci. Rep.2:864. 10.1038/srep00864
105
MorenoR.MayerR. E. (2000). Engaging students in active learning: The case for personalized multimedia messages.J. Educ. Psychol.92724–733. 10.1037/0022-0663.92.4.724
106
MoriM.MacDormanK. F.KagekiN. (2012). The uncanny valley [from the field].IEEE Robot. Automat. Mag.1998–100. 10.1109/mra.2012.2192811
107
MourkoussisN.RiveraF. M.TrosciankoT.DixonT.HawkesR.ManiaK. (2010). Quantifying fidelity for virtual environment simulations employing memory schema assumptions.ACM Trans. Appl. Percept.81–21. 10.1145/1857893.1857895
108
museum4punkt0 (2019). museum4punkt0—Digitale Strategien für das Museum der Zukunft. Available online at: https://www.museum4punkt0.de/(accessed February 21, 2019)
109
museum-digital (2019). Museum-digital. Available online at: https://www.museum-digital.de/(accessed February 21, 2019).
110
NadlerR. T.RabiR.MindaJ. P. (2010). Better mood and better performance. learning rule-described categories is enhanced by positive mood.Psychol. Sci.211770–1776. 10.1177/0956797610387441
111
NebelS. (2017). Investigating the Mechanisms of Competition within Educational Video Games—Die Mechanismen des Wettbewerbs in digitalen Lernspielen. Doctoral dissertation, Technische Universität Chemnitz, Chemnitz.
112
NebelS.SchneiderS.ReyG. D. (2016). Mining learning and crafting scientific experiments: a literature review on the use of minecraft in education and research.J. Educ. Technol. Soc.19355–366.
113
NguyenU.RottensteinerF.HeipkeC. (2019). “Confidence-aware pedestrian tracking using a stereo camera,” in ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences, IV-2/W5, ( Heipke: International Society for Photogrammetry and Remote Sensing), 53–60. 10.5194/isprs-annals-IV-2-W5-53-2019
114
NicholsonD. T.ChalkC.FunnellW. R. J.DanielS. J. (2006). Can virtual reality improve anatomy education? A randomised controlled study of a computer-generated three-dimensional anatomical ear model.Med. Educ.401081–1087. 10.1111/j.1365-2929.2006.02611.x
115
NicolaeC.NocerinoE.MennaF.RemondinoF. (2014). “Photogrammetry applied to problematic artefacts,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XL–5, (Heipke: International Society for Photogrammetry and Remote Sensing), 451–456. 10.5194/isprsarchives-XL-5-451-2014
116
NormanJ. F.ToddJ. T.OrbanG. A. (2004). Perception of three-dimensional shape from specular highlights, deformations of shading, and other types of visual information.Psychol. Sci.15565–570. 10.1111/j.0956-7976.2004.00720.x
117
OrrT. J.BellancaJ. L.NavoyskiJ.MacdonaldB.HelfrichW.DemichB. (2020). “Development of visual elements for accurate simulation,” in Advances in Human Factors and Simulation, Vol. 958ed.CassentiD. N. (Cham: Springer), 287–299. 10.1007/978-3-030-20148-7_26
118
PaesD.IrizarryJ. (2019). The relevance of visual cues in immersive environments: does pictorial realism matter?Comput. Civil Eng.201925–31. 10.1061/9780784482421.004
119
PekrunR.ElliotA. J.MaierM. A. (2009). Achievement goals and achievement emotions: testing a model of their joint relations with academic performance.J. Educ. Psychol.101115–135. 10.1037/a0013383
120
PekrunR.GoetzT.TitzW.PerryR. P. (2002). Academic emotions in students’ self-regulated learning and achievement: a program of qualitative and quantitative research.Educ. Psychol.3791–105. 10.1207/S15326985EP3702_4
121
PlassJ. L.HeidigS.HaywardE. O.HomerB. D.UmE. (2014). Emotional design in multimedia learning: effects of shape and color on affect and learning.Learn. Instr.29128–140. 10.1016/j.learninstruc.2013.02.006
122
PlassJ. L.KalyugaS. (2019). Four ways of considering emotion in cognitive load theory.Educ. Psychol. Rev.31339–359. 10.1007/s10648-019-09473-5
123
PlassJ. L.KaplanU. (2015). “Emotional design in digital media for learning,” in Emotions, Technology, Design, and Learning, edsTettegahS. Y.GartmeierM. (San Diego: Academic Press), 131–161. 10.1016/b978-0-12-801856-9.00007-4
124
PoeschlS. (2017). Virtual reality training for public speaking—A QUEST-VR framework validation.Front. ICT4:13. 10.3389/fict.2017.00013
125
RademacherP.LengyelJ.CutrellE.WhittedT. (2001). “Measuring the perception of visual realism in images,” in Rendering Techniques 2001, edsGortlerS. J.MyszkowskiK. (Berlin: Springer Science and Business Media), 235–247. 10.1007/978-3-7091-6242-2_22
126
ReevesB.NassC. (1996). The Media Equation.New York, NY: Cambridge University Press.
127
RehanyN.BarsiA.LovasT. (2017). “Capturing fine details involving low-cost sensors –a comparative study,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume XLII-2/W8, (Heipke: International Society for Photogrammetry and Remote Sensing), 213–220. 10.5194/isprs-archives-XLII-2-W8-213-2017
128
ReulkeR.DöringT. (2006). “Computer Vision im Kontext von Photographie und Photogrammetrie,” in Informatik, edsReisigW.FreytagJ.-C. (Berlin: Springer), 315–328. 10.1007/3-540-32743-6_13
129
ReyG. D. (2012). A review of research and a meta-analysis of the seductive detail effect.Educ. Res. Rev.7216–237. 10.1016/j.edurev.2012.05.003
130
RichterJ.ScheiterK.EitelA. (2016). Signaling text-picture relations in multimedia learning: a comprehensive meta-analysis.Educ. Res. Rev.1719–36. 10.1016/j.edurev.2015.12.003
131
RogersB.GrahamM. (1979). Motion parallax as an independent cue for depth perception.Perception8125–134. 10.1068/p080125
132
RushmeierH. E.RogowitzB. E.PiatkoC. (2000). “Perceptual issues in substituting texture for geometry,” in Proceedings of the Human Vision and Electronic Imaging V, edsRogowitzB. E.PappasT. N. (San Jose, CA: SPEI), 10.1117/12.387174
133
ScheiterK.GerjetsP.HukT.ImhofB.KammererY. (2009). The effects of realism in learning with dynamic visualizations.Learn. Instr.19481–494. 10.1016/j.learninstruc.2008.08.001
134
SchneiderS.BeegeM.NebelS.ReyG. D. (2018a). A meta-analysis of how signaling affects learning with media.Educ. Res. Rev.231–24. 10.1016/j.edurev.2017.11.001
135
SchneiderS.DyrnaJ.MeierL.BeegeM.ReyG. D. (2018b). How affective charge and text–picture connectedness moderate the impact of decorative pictures on multimedia learning.J. Educ. Psychol.110233–249. 10.1037/edu0000209
136
SchneiderS.NebelS.ReyG. D. (2016). Decorative pictures and emotional design in multimedia learning.Learn. Instr.4465–73. 10.1016/j.learninstruc.2016.03.002
137
SchroederN. L.AdesopeO. O.GilbertR. B. (2013). How effective are pedagogical agents for learning? A meta-analytic review.J. Educ. Comput. Res.491–39. 10.2190/EC.49.1.a
138
SchroederN. L.TraxlerA. L. (2017). Humanizing instructional videos in physics: when less is more.J. Sci. Educ. Technol.26269–278. 10.1007/s10956-016-9677-6
139
SkulmowskiA.ReyG. D. (2018). Realistic details in visualizations require color cues to foster retention.Comput. Educ.12223–31. 10.1016/j.compedu.2018.03.012
140
SkulmowskiA.ReyG. D. (2020). The realism paradox: realism can act as a form of signaling despite being associated with cognitive load.Hum. Behav. Emerg. Technol.10.1002/hbe2.190[Epub ahead of print].
141
SlaterM.KhannaP.MortensenJ.YuI. (2009). Visual realism enhances realistic response in an immersive virtual environment.IEEE Comput. Graph. Appl.2976–84. 10.1109/MCG.2009.55
142
SlaterM.LinakisV.UsohM.KooperR. (1996). “Immersion, presence, and performance in virtual environments: an experiment with tri-dimensional chess,” in Proceedings of the ACM Symposium on Virtual Reality Software and Technology, (Hong Kong: Association for Computing Machinery), 163–172.
143
SmallmanH. S.St JohnM. (2005). Naive realism: misplaced faith in realistic displays.Ergon. Design Q. Hum. Fact. Appl.136–13. 10.1177/106480460501300303
144
Staatliche Museen zu Berlin (2019). Der Pergamon Altar. Available online at: http://3d.smb.museum/pergamonaltar/(accessed February 18, 2019)
145
StullA. T.HegartyM.MayerR. E. (2009). Getting a handle on learning anatomy with interactive three-dimensional graphics.J. Educ. Psychol.101803–816. 10.1037/a0016849
146
StullA. T.MayerR. E. (2007). Learning by doing versus learning by viewing: three experimental comparisons of learner-generated versus author-provided graphic organizers.J. Educ. Psychol.99808–820. 10.1037/0022-0663.99.4.808
147
SuganoN.KatoH.TachibanaK. (2003). “The effects of shadow representation of virtual objects in augmented reality,” in The Second IEEE and ACM International Symposium on Mixed and Augmented Reality, 2003. Proceedings, (New York, NY: Association for Computing Machinery), 76–83. 10.1109/ISMAR.2003.1240690
148
SwellerJ.MerriënboerJ. J. G. V.PaasF. G. W. C. (1998). Cognitive architecture and instructional design.Educ. Psychol. Rev.10251–296. 10.1023/A:1022193728205
149
SwellerJ.van MerriënboerJ. J. G.PaasF. (2019). Cognitive architecture and instructional design: 20 years later.Educ. Psychol. Rev.31261–292. 10.1007/s10648-019-09465-5
150
The Arc/k Project (2019). The Arc/k Project. Available online at: https://arck-project.org/(accessed February 21, 2019).
151
ToldoR. (2013). Towards Automatic Acquisition of High-Level 3D Models from Images.Verona: Universit‘a degli Studi di Verona.
152
TractinskyN.KatzA. S.IkarD. (2000). What is beautiful is usable.Interact. Comput.13127–145. 10.1016/S0953-5438(00)00031-X
153
TucciG.CiniD.NobileA. (2012). “Effective 3D digitization of archaeological artifacts for interactive virtual museum,” in ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XXXVIII-5/W16, ( Heipke: International Society for Photogrammetry and Remote Sensing), 413–420. 10.5194/isprsarchives-XXXVIII-5-W16-413-2011
154
TuttasS.BraunA.BormannA.StillaU. (2014). “Konzept zur automatischen Baufortschrittskontrolle durch Integration eines Building Information Models und photogrammetrisch erzeugten Punktwolken,” in Proceedings GPF Tagungsband 23/2014 (Hamburg: DGfK, DGPF, GfGI, and GiN).
155
UmE. R.PlassJ. L.HaywardE. O.HomerB. D. (2012). Emotional design in multimedia learning.J. Educ. Psychol.104485–498.
156
van GisbergenM.KovacsM.CamposF.van der HeeftM.VugtsV. (2019). “What we don’t know. The effect of realism in virtual reality on experience and behaviour,” in Augmented Reality and Virtual Reality, edsTom DieckM. C.JungT. (Cham: Springer), 45–57. 10.1007/978-3-030-06246-0_4
157
van GogT. (2014). “The Signaling (or Cueing) principle in multimedia learning,” in The Cambridge Handbook of Multimedia Learning, 2nd Edn, ed.MayerR. (Cambridge: Cambridge University Press), 263–278. 10.1017/CBO9781139547369.014
158
VangorpP.LaurijssenJ.DutréP. (2007). The influence of shape on the perception of material reflectance.ACM Trans. Graphics26:77. 10.1145/1239451.1239528
159
VeletsianosG.RussellG. S. (2014). “Pedagogical Agents,” in Handbook of Research on Educational Communications and Technology, edsSpectorJ. M.MerrillM. D.ElenJ.BishopM. J. (Berlin: Springer Science and Business Media), 759–769. 10.1007/978-1-4614-3185-5_61
160
VincentM. L.GutierrezM. F.CoughenourC.ManuelV.BendichoL.-M.RemondinoF.et al (2015). “Crowd-sourcing the 3D digital reconstructions of lost cultural heritage,” in Proceedings of the 2015 Digital Heritage, (Granada: IEEE), 171–172. 10.1109/DigitalHeritage.2015.7413863
161
WallerD.HuntE.KnappD. (1998). The transfer of spatial knowledge in virtual environment training.Presence Teleoperat. Virtual Environ.7129–143. 10.1162/105474698565631
162
WallerD.LippaY. (2007). Landmarks as beacons and associative cues: their role in route learning.Mem. Cogn.35910–924. 10.3758/BF03193465
163
WaltemateT.GallD.RothD.BotschM.LatoschikM. E. (2018). The impact of avatar personalization and immersion on virtual body ownership, presence, and emotional response.IEEE Trans. Vis. Comput. Graph.241643–1652. 10.1109/TVCG.2018.2794629
164
WalterJ.EdwardsJ.McDonaldG.KuchelH. (2018). Photogrammetry for the estimation of wheat biomass and harvest index.Field Crops Res.216165–174. 10.1016/j.fcr.2017.11.024
165
WaltonK. L. (1984). Transparent pictures: on the nature of photographic realism.Crit. Inquiry11246–277. 10.1086/448287
166
WangJ.AntonenkoP. D. (2017). Instructor presence in instructional video: effects on visual attention, recall, and perceived learning.Comput. Hum. Behav.7179–89. 10.1016/j.chb.2017.01.049
167
WangT.ShiQ.NikkhooM.WeiS.BarbotS.DregerD.et al (2018). The rise, collapse, and compaction of Mt. Mantap from the 3 September 2017 North Korean nuclear test.Science361166–170. 10.1126/science.aar7230
168
WeisS. (2019). How I 3D scanned the Treasury at Petra Without Leaving Home. Available online at: https://packet39.com/blog/2019/01/09/how-i-3d-scanned-the-treasury-at-petra-without-leaving-home/(accessed December 17, 2019).
169
WilkeningJ.FabrikantS. I. (2011). “How do decision time and realism affect map-based decision making?,” in Spatial Information Theory, Vol. 6899edsEgenhoferM.GiudiceN.MoratzR.WorboysM. (Berlin: Springer Science and Business Media), 1–19. 10.1007/978-3-642-23196-4_1
170
Williams-PierceC. C.-M. (2016). Multiplying Fractions and “Little Big Planet 2”: Playing to Learn.Ann Arbor, MI: ProQuest LLC.
171
WolfsonS.CaseG. (2000). The effects of sound and colour on responses to a computer game.Interact. Comput.13183–192. 10.1016/S0953-5438(00)00037-0
172
WormwoodJ. B.SiegelE. H.KopecJ.QuigleyK. S.BarrettL. F. (2019). You are what I feel: a test of the affective realism hypothesis.Emotion19788–798. 10.1037/emo0000484
173
WoutersP.van NimwegenC.van OostendorpH.van der SpekE. D. (2013). A meta-analysis of the cognitive and motivational effects of serious games.J. Educ. Psychol.105249–265. 10.1037/a0031311
174
YuI.MortensenJ.KhannaP.SpanlangB.SlaterM. (2012). Visual realism enhances realistic response in an immersive virtual environment—part 2.IEEE Comput. Graph. Appl.3236–45. 10.1109/MCG.2012.121
175
ZaalbergR.MiddenC. (2010). “Enhancing human responses to climate change risks through simulated flooding experiences,” in Persuasive Technology, Vol. 6137edsPlougT.HasleP.Oinas-KukkonenH. (Berlin: Springer), 205–210. 10.1007/978-3-642-13226-1_21
176
ZimmonsP.PanterA. (2003). “The influence of rendering quality on presence and task performance in a virtual environment,” in IEEE Virtual Reality, 2003. Proceedings, (Los Angeles, CA: IEEE), 293–294. 10.1109/VR.2003.1191170
Summary
Keywords
photorealistic 3D models, photogrammetry, educational psychology, virtual reality, realism, educational technology
Citation
Nebel S, Beege M, Schneider S and Rey GD (2020) A Review of Photogrammetry and Photorealistic 3D Models in Education From a Psychological Perspective. Front. Educ. 5:144. doi: 10.3389/feduc.2020.00144
Received
28 January 2020
Accepted
21 July 2020
Published
07 August 2020
Volume
5 - 2020
Edited by
Shashidhar Venkatesh Murthy, James Cook University, Australia
Reviewed by
Mashhuda Glencross, The University of Queensland, Australia; Dieter Fritsch, University of Stuttgart, Germany
Updates
Copyright
© 2020 Nebel, Beege, Schneider and Rey.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Steve Nebel, steve.nebel@phil.tu-chemnitz.de
This article was submitted to Digital Education, a section of the journal Frontiers in Education
Disclaimer
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.