Hostname: page-component-669899f699-swprf Total loading time: 0 Render date: 2025-04-28T12:07:46.429Z Has data issue: false hasContentIssue false

Egocentric and allocentric memory recall strategies moderate transfer action sentence recognition

Published online by Cambridge University Press:  21 April 2025

Cosimo Tuena
Affiliation:
Applied Technology for Neuro-Psychology Lab, IRCCS Istituto Auxologico Italiano, Milan, Italy Department of Theoretical and Applied Sciences, eCampus University, Novedrate, Italy
Daniele Di Lernia
Affiliation:
Department of Theoretical and Applied Sciences, eCampus University, Novedrate, Italy
Giuseppe Riva
Affiliation:
Applied Technology for Neuro-Psychology Lab, IRCCS Istituto Auxologico Italiano, Milan, Italy Humane Technology Lab, Università Cattolica del Sacro Cuore, Milan, Italy
Claudia Repetto*
Affiliation:
Department of Psychology, Università Cattolica del Sacro Cuore, Milan, Italy
Silvia Serino
Affiliation:
Department of Psychology, Università degli Studi Milano – Bicocca, Milan Italy
*
Corresponding author: Claudia Repetto; Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Embodied cognition theory proposes that spatial cognition preferences facilitate the simulation of action language. Importantly, spatial cognition relies on either egocentric (body-dependent) or allocentric (body-independent) representations. Research demonstrates that spatial representation proclivity influences the simulation of non-transfer action sentences. However, the impact of individual spatial cognition preferences on transfer action sentence simulation remains unexplored. We administered an egocentric and allocentric memory task and an action sentence recognition task to 37 participants. We used an egocentric–allocentric recall strategy proclivity index to classify participants and employed this metric as a moderator between the transfer perspective (first-person perspective, 1PP vs. third-person perspective, 3PP) and the transfer type (concrete vs. abstract). We found that spatial preferences do not moderate 1PP transfer action sentence recognition. Importantly, we found that egocentric proclivity improves 3PP transfer action sentence recognition and that allocentric proclivity hampers 3PP transfer action sentence recognition. No moderation was found for the transfer type. The study suggests that recognition memory for sentences describing others’ actions is related to body-dependent spatial representations, suggesting a possible link between spatial memory proclivity and action language simulation.

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial licence (http://creativecommons.org/licenses/by-nc/4.0), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use.
Copyright
© The Author(s), 2025. Published by Cambridge University Press

1. Introduction

Embodied cognition (EC) theories propose a fundamental shift away from cognitivist assumptions, arguing that the sensorimotor system plays a crucial role in cognitive representations. This reconceptualization of cognition has influenced research examining various cognitive domains, including memory and language (Borghi, Reference Borghi2024; Dove, Reference Dove2023; Ianì, Reference Ianì2019). EC suggests that language comprehension engages sensorimotor information to mentally simulate sentence content (Glenberg & Gallese, Reference Glenberg and Gallese2012; Meteyard et al., Reference Meteyard, Cuadrado, Bahrami and Vigliocco2012). This is especially evident with concrete action language (triggered by concrete verbs such as ‘to kick’), although evidence also indicates that abstract action language (triggered by figurative verbs such as ‘to plan’) is grounded in the sensorimotor system (Ghandhari et al., Reference Ghandhari, Fini, Da Rold and Borghi2020; Glenberg et al., Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008; Sakreida et al., Reference Sakreida, Scorolli, Menz, Heim, Borghi and Binkofski2013). Furthermore, language simulation processes have been documented in tasks examining embodiment through both ‘online’ (e.g., verification tasks) and ‘offline’ (e.g., recognition memory task) paradigms (e.g., Brunyé et al., Reference Brunyé, Ditman, Mahoney, Augustyn and Taylor2009; Reference Brunyé, Ditman, Giles, Holmes and Taylor2016; Díez-Álamo et al., Reference Díez-Álamo, Glenberg, Díez, Alonso and Fernandez2020; Ditman et al., Reference Ditman, Brunyé, Mahoney and Taylor2010; Tuena, Di Lernia, Riva, et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023; Tuena, Di Lernia, Rodella, et al., Reference Tuena, Di Lernia, Riva, Serino and Repetto2023). The fact that the body plays an important role in language comprehension is further demonstrated by the observation that individuals can simulate action language embodying the spatial perspectives (point of view) of the characters involved (Beveridge & Pickering, Reference Beveridge and Pickering2013; Meteyard et al., Reference Meteyard, Cuadrado, Bahrami and Vigliocco2012; Zwaan, Reference Zwaan2016). Indeed, growing evidence suggests a link between spatial cognition and language simulation (Beveridge & Pickering, Reference Beveridge and Pickering2013; Ibáñez et al., Reference Ibáñez, Kühne, Miklashevsky, Monaco, Muraki, Ranzini and Tuena2023; Majid et al., Reference Majid, Bowerman, Kita, Haun and Levinson2004; Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017). The spatial grounding hypothesis (Beveridge & Pickering, Reference Beveridge and Pickering2013) states that for the action simulation to occur, readers must represent the spatial context (i.e., the spatial relationships among entities within a given text by means of a situation model; Zwaan & Radvansky, Reference Zwaan and Radvansky1998) for the sentence. While agent-perspective simulations appear to be the most readily accessible, action simulations from other perspectives (e.g., patient or observer) are also possible. These non-agent simulations are facilitated by self-referential pronouns that occupy thematic roles other than the agent. This process is referred to as action perspective-taking. Additionally, this hypothesis highlights a second aspect: the spatial perspective (e.g., agent, patient or observer) from which the situational model incorporating the action is mentally represented. This latter process is designated as spatial perspective-taking.

A particular case of action sentences is when two characters are involved, one with the role of the agent and the other with the role of the patient/receiver (namely transfer action sentences). These types of stimuli have been extensively investigated using the action-sentence compatibility effect (ACE) (e.g., Glenberg et al., Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008; Glenberg & Kaschak, Reference Glenberg and Kaschak2002; Ibáñez et al., Reference Ibáñez, Cardona, Dos Santos, Blenkmann, Aravena, Roca and Bekinschtein2013; Papesh, Reference Papesh2015). The standard ACE paradigm requires participants to judge whether the sentences make sense by pressing a button located closer to or farther away from their body, relative to a starting position. These sentences could be both concrete (e.g., “Courtney handed you the notebook”) and abstract (e.g., “You confessed your secret to Dan”). The key manipulation involves the compatibility between the action described in the sentence (away from or toward the participant’s body) and the action required to respond (pushing a button far or close to the body). In a congruent condition, the physical action performed by the participant is compatible with the action described in the sentence. In the incongruent condition, the required physical action is incompatible with the action described in the sentence. Faster reaction times were observed in the congruent compared to incongruent conditions for both concrete and abstract action sentences (e.g., Borreggine & Kaschak, Reference Borreggine and Kaschak2006; Glenberg et al., Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008; Glenberg & Kaschak, Reference Glenberg and Kaschak2002) or only for concrete action sentences (e.g., Diefenbach et al., Reference Diefenbach, Rieger, Massen and Prinz2013). Despite the interesting results, Papesh (Reference Papesh2015) reported weak evidence for ACE in her review. More recently, using the ACE procedure, one study (Díez-Álamo et al., Reference Díez-Álamo, Glenberg, Díez, Alonso and Fernandez2020) showed that sentences that describe concrete or abstract transfer toward the self are processed faster and remembered better than concrete and abstract transfer sentences describing motion away from the reader. The authors did not find the ACE and named the result the linguistic looming effect. These findings were replicated by a recent pre-registered multi-lab study on the ACE (Morey et al., Reference Morey, Kaschak, Díez-Álamo, Glenberg, Zwaan, Lakens and Ziv-Crispel2022). This is somehow different from what Ditman et al. (Reference Ditman, Brunyé, Mahoney and Taylor2010) found for non-transfer (single-character) concrete action language. In this case, action language is remembered better when sentences depict an action from a 1PP (e.g., “You slice the tomato”) than from a 3PP (e.g., “He slices the tomato”); this finding demonstrates that simulation could occur also during recognition memory tasks and in a similar way to the so-called enactment effect (see, Engelkamp, Reference Engelkamp1998). Conversely, Díez-Álamo et al. (Reference Díez-Álamo, Glenberg, Díez, Alonso and Fernandez2020) suggested that language-induced visual perceptual looming bias grabs attention and facilitates a quick adaptative response. However, according to the spatial grounding hypothesis, when a spatial context is unavailable for a transfer action sentence, the processes of action language simulation and perspective-taking may be impaired. (Beveridge & Pickering, Reference Beveridge and Pickering2013; Gianelli et al., Reference Gianelli, Farnè, Salemme, Jeannerod and Roy2011). Moreover, in a recent consensus paper, embodied cognition researchers proposed that when studying action language, it is crucial to evaluate individual preferences, particularly those related to spatial cognition, as they may influence simulation processes. (Ibáñez et al., Reference Ibáñez, Kühne, Miklashevsky, Monaco, Muraki, Ranzini and Tuena2023).

Spatial memory is a fundamental aspect of human spatial cognition (Burgess, Reference Burgess2008). In particular, spatial memory is the capacity to encode, store and remember the position and location of objects or places and includes orientation and distance (VandenBos, Reference VandenBos2007). This information can be represented using two spatial frames of reference: egocentric and allocentric (Burgess, Reference Burgess2008). In the former, the environment is represented in a body-dependent way (e.g., “to reach the church, go left at the bar”), whereas in the latter, it is represented independently from our body and position (e.g., “the church is close to the bar”). These two representations can be used flexibly by switching from one to another (Ekstrom et al., Reference Ekstrom, Arnold and Iaria2014). In particular, research (Goeke et al., Reference Goeke, König and Gramann2013; Gramann et al., Reference Gramann, Müller, Eick and Schönebeck2005) showed that individuals have a preference for either an egocentric or allocentric frame of reference for solving spatial orientation tasks. In these tasks, participants watched on the computer screen short virtual paths (e.g., 12 seconds), without environmental landmarks, with different angle deviations from the fixed starting point. At the end of the path, participants were asked to indicate using arrows presented on the screen the heading direction of the initial position. Some arrows implied no rotation from the initial direction (i.e., participants kept the starting heading direction fixed during the path; allocentric preference), whereas other arrows denoted rotation from the starting direction (i.e., participants rotated the initial heading direction according to the path direction shown in the video; egocentric preference). Crucially, these tasks require ‘online’ spatial computations with an immediate retrieval of the navigated path and participants have no motor control over the navigated path (i.e., joypad/keyboard for navigation). However, egocentric and allocentric preferences are not used during ‘online’ spatial cognition; indeed, individuals could rely on egocentric or allocentric memory strategies also during spatial tasks that require information to be stored in memory (Chersi & Burgess, Reference Chersi and Burgess2015). Using landmark-based virtual reality (VR) navigation tasks, it is possible to assess the ability to encode or recall object locations in an egocentric or allocentric representation by manipulating the virtual environmental cues (Burgess, Reference Burgess2008; Guderian et al., Reference Guderian, Dzieciol, Gadian, Jentschke, Doeller, Burgess and Vargha-Khadem2015). Environmental cues in navigation can be categorized into two types: boundaries and landmarks (Chersi & Burgess, Reference Chersi and Burgess2015; Doeller et al., Reference Doeller, King and Burgess2008; Schuck et al., Reference Schuck, Doeller, Polk, Lindenberger and Li2015). Boundaries are extended obstacles, such as walls, that define the limits of an environment, whereas landmarks are discrete objects within the space, such as monuments, that serve as reference points. Boundary-based navigation is predominantly supported by the hippocampus and typically employs an allocentric strategy forming a cognitive map of the environment independent of one’s position. Landmark-based navigation, on the other hand, predominantly engages the dorsal striatum and typically involves an egocentric strategy by encoding spatial information from the navigator’s perspective with respect to discrete (also known as local) cues.

Regarding the interaction between action language simulation and spatial cognition, previous research has demonstrated that in non-transfer action sentences where participants act as agents, a first-person perspective (1PP; egocentric) is activated, while sentences featuring another person as the agent trigger a third-person perspective (3PP; allocentric) (Beveridge & Pickering, Reference Beveridge and Pickering2013; Brunyé et al., Reference Brunyé, Ditman, Mahoney, Augustyn and Taylor2009; Reference Brunyé, Ditman, Giles, Holmes and Taylor2016). Crucially, spatial preferences might affect ‘online’ simulation processes. One study (Vukovic & Williams, Reference Vukovic and Williams2015) explored the impact of egocentric/allocentric frames of reference proclivity on concrete (action) sentence simulation through an action sentence-picture verification task. Participants listened to ‘You’ (“You are opening a bottle” – the participant is the agent) and ‘I’ (“I am opening a bottle” – someone else is the agent) action concrete sentences and then were asked to match a photo of that action from the 1PP or the 3PP. By using a spatial orientation task (Gramann et al., Reference Gramann, Müller, Eick and Schönebeck2005), the authors categorized participants as having an egocentric or allocentric proclivity. The first experiment reported in the study found that when listening to ‘You’ and ‘I’ sentences, participants with an egocentric preference were faster to verify 1PP photos compared to 3PP photos. No effect was found for allocentric participants. In the second experiment, the authors demonstrated that allocentric participants have differences in action sentence simulation when contextual information is manipulated (experimental instructions and stimuli; i.e., participants were asked to ‘spy’ a conversation of two individuals shown on the PC screen). In this case, allocentric participants had a faster response time to sentence-photo pairs from a 3PP perspective. Similarly to the first sentence-picture matching experiment procedure, a study (Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017) used instead a continuous numeric index to rate spatial frames of reference preference computed from the same computerized orientation task (Gramann et al., Reference Gramann, Müller, Eick and Schönebeck2005). They found that egocentric participants are facilitated when the sentences match the perspective of the photo shown (1PP and 3PP).

To examine the intricate interaction between transfer sentences and frames of reference proclivity, a cross-cultural study (Tuena et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023) examined Italian and American English participants. Cultural background can influence the use of spatial frames of reference at the population level. North Americans predominantly favor an allocentric frame of reference orientation, whereas Europeans demonstrate no specific preference (Goeke et al., Reference Goeke, Kornpetpanee, Köster, Fernández-Revelles, Gramann and König2015). In Tuena and colleagues’ study, participants read concrete transfer sentences in which they could be either the agent or the patient. They were then asked to verify whether a photo from a first-person perspective (1PP) or third-person perspective (3PP) matched the previously shown sentence. The researchers found no differences between the two samples, suggesting shared embodied mechanisms: when readers were the patients in action sentences, they demonstrated faster response times in the photo-matching task. This suggests that participants simulated the agent’s action (performed by another person) while maintaining the patient’s point of view (their own) from a third-person perspective. Consequently, action and spatial perspective sentence simulation may operate independently in the context of concrete transfer action sentence phrases. Crucially, participants demonstrated faster responses when they were the agents of the action sentence and the picture depicted the action from their first-person perspective (1PP), suggesting that action and spatial perspective-taking converge with the participant’s perspective in these conditions.

In summary, for non-transfer action sentences (Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017; Vukovic & Williams, Reference Vukovic and Williams2015), egocentric participants simulate pictures more rapidly from a first-person perspective (1PP) compared to a third-person perspective (3PP) when acting as agents. Conversely, allocentric participants simulate others’ actions more quickly from a 3PP compared to 1PP, but only when contextual information is manipulated. However, these patterns shift with transfer sentences, as readers may adopt varying action and spatial perspectives. The findings from Tuena et al. (Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023) challenge the spatial grounding hypothesis (Beveridge & Pickering, Reference Beveridge and Pickering2013): regardless of spatial cultural bias, action perspective-taking consistently occurs from the agent’s perspective (whether the participant or another person), while spatial perspective-taking remains fixed to the reader’s (participant’s) viewpoint, which can be either 1PP or 3PP.

From the studies mentioned above, it is still unclear how spatial cognition preferences and language simulation are bound and possibly affected by contextual aspects, like the type of spatial task (e.g., online vs. memory).

In the present study, we investigated whether participants’ predisposition toward egocentric or allocentric memory modulates their recognition of transfer action sentences (both concrete and abstract) from different perspectives. To pursue these aims, we employed a VR spatial memory task to generate an egocentric–allocentric recall strategy proclivity index to predict action sentence recognition performance.

Regardless of the type (concrete or abstract) of transfer, we speculate that participants with an egocentric recall preference will have better recognition of 1PP sentences, conversely, participants with an allocentric recall preference will have better 3PP sentences recognition.

2. Methods

2.1. Participants

We recruited 37 Italian young adults for this study (Mage = 24.05, SDage = 2.57; Meducation = 15.56, SDeducation = 2.32; males = 18; right-handed = 31). Participants were recruited at psychology courses at the Catholic University of Milan. Inclusion criteria were: speaking Italian as a native language and age 18–30 years. Exclusion criteria were self-reported history of memory and/or language disorders. We computed the sample size by considering that all the participants were allocated to each level of the independent variables of the sentence recognition task and the number of stimuli used (stimuli-within-condition design). With a medium Cohen’s d of 0.5, a power of 0.8, and 80 target stimuli, the power analysis for a mixed-effects model (Westfall et al., Reference Westfall, Kenny and Judd2014) required a minimum of 26 participants. Moreover, considering that our design included a third continuous predictor in the linear model, at least 10 observations were added (VanVoorhis & Morgan, Reference VanVoorhis and Morgan2007). The study was approved by the Ethics Committee of the Catholic University of Milan. Participants gave their written consent to participate.

2.2. Egocentric and allocentric spatial memory task

We employed a VR landmark-based navigation task to test participants’ egocentric and allocentric spatial memory strategies, as in a similar study (Guderian et al., Reference Guderian, Dzieciol, Gadian, Jentschke, Doeller, Burgess and Vargha-Khadem2015). The virtual environment consisted of a circular arena surrounded by a wall, with an obelisk inside the arena and some fixed distal cues (i.e., mountains and clouds) visible throughout the experiment to provide orientation. In the encoding phase, participants navigated the virtual arena (50 virtual meters in diameter) using arrow keys and a mouse. They had to collect eight objects and memorize their locations in the arena. The eight objects (supra-span performance; Jeneson & Squire, Reference Jeneson and Squire2011; Spinnler & Tognoni, Reference Spinnler and Tognoni1987) were balanced for living and non-living categories. The objects had predetermined locations and appeared one at a time in random order across the trials. The participants had to navigate to the exact item location to collect the item and see the following. Once they reached it, the object disappeared, and the participant had to find the next one. Object locations could be encoded in this phase using the boundaries of the arena (i.e., wall – allocentric cue) and a local landmark (i.e., obelisk – egocentric cue). Figure 1 shows the environment at encoding. Each object was collected four times in random order.

Figure 1. Virtual environment of the arena.

Note: During the encoding phase, both a local landmark (obelisk – egocentric) and a boundary (wall – allocentric) were used to encode the item location. During the recall phase, either the local landmark or the wall was shown to assess egocentric or allocentric memory recall.

During the recall phase, either the wall or the obelisk was removed in random order across recall trials. This forced the use of an allocentric (i.e., obelisk removed) or egocentric (i.e., wall removed) strategy during spatial memory recall (Chersi & Burgess, Reference Chersi and Burgess2015). At each trial, participants were shown each item one at a time at the bottom of the PC screen and had to navigate to the location where the item had previously been collected. Once there, they had to press the spacebar if they were satisfied with the remembered location. Each object was tested four times for 32 recall trials (16 trials for each allocentric and egocentric recall strategy). The response variable was the Euclidean distance between the recalled and actual location of each object trial at recall (distance expressed as virtual meters of the virtual environment between the recalled and the actual location at encoding). The task was created with Unity software.

2.3. Transfer sentence recognition memory task

We developed a sentence recognition memory task with Gorilla software (Anwyl-Irvine et al., Reference Anwyl-Irvine, Massonnié, Flitton, Kirkham and Evershed2020) with the old-new paradigm (Squire et al., Reference Squire, Wixted and Clark2007). A set of 160 sentences was created involving two characters (the pronoun ‘You’/the participant and Gianni or Maria). We used ‘You’ sentences because this yielded the most consistent results in studies on sentence simulation (Brunyé et al., Reference Brunyé, Ditman, Mahoney, Augustyn and Taylor2009; Reference Brunyé, Ditman, Giles, Holmes and Taylor2016; Ditman et al., Reference Ditman, Brunyé, Mahoney and Taylor2010; Vukovic & Williams, Reference Vukovic and Williams2015). The set of stimuli was divided into four action sentence conditions: 1PP concrete transfer sentences (e.g., “You shoot the rubber band at Gianni”), 3PP concrete transfer sentences (e.g., “Maria shoots the rubber band at you”), 1PP abstract transfer sentences (e.g., “You give some time to Maria”); and lastly, 3PP abstract transfer sentences (e.g., “Maria gives some time to you”). In the 1PP, the reader is the agent, while in 3PP, the reader is the grammatical patient (receiver). Forty action sentences were taken from a previous study (Tuena et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023), while the remaining were created ad hoc.

As in Díez-Álamo et al. (Reference Díez-Álamo, Glenberg, Díez, Alonso and Fernandez2020), we created a main list of 80 sentences (20 sentences for each condition) and then we created a reversed list (i.e., the sentence “You shoot the rubber band at Gianni” in the main list was reversed to “Gianni shoots the rubber band at you”). Following the procedure of Díez-Álamo and colleagues, participants in the encoding phase learned the sentences of the main list (old items) during the recognition phase, the sentences of the reversed list were shown as new items in addition to old items. We counterbalanced this aspect across participants so that the participant that learned “You shoot the rubber band at Gianni” had as a new item “Gianni shoots the rubber band at you,” the following participant had as target “Gianni shoots the rubber band at you” and as new item “You shoot the rubber band at Gianni.” The participants should discriminate the correct agent/patient for the old from the new sentences. Stimuli were also balanced by gender (Maria/Gianni) so that half of the concrete and abstract sentences involved a male and half a female. Additionally, we analyzed the sentence lengths (letters plus spaces) for old stimuli, comparing concrete and abstract sentences between first-person singular (1PP) and third-person singular (3PP) perspectives in both main and reversed lists. For concrete sentences in the main lists, the comparison between 1PP and 3PP yielded t37 = 1.88, p = 0.068. For abstract sentences in the main lists, the 1PP versus 3PP comparison resulted in t37 = 1.70, p = 0.099. In the reversed lists, the 1PP-3PP comparison for concrete sentences showed t37 = 1.98, p = 0.057, while for abstract sentences, it resulted in t37 = 1.37, p = 0.180. None of these comparisons revealed statistically significant differences in sentence length between 1PP and 3PP perspectives. In addition, sentence lengths (letters plus spaces) for old stimuli were not statistically different for concrete and abstract sentences, in the main (concrete, t37 = 1.88, p = 0.068; abstract, t37 = 1.70, p = 0.099) and reversed (concrete, t37 = 1.98, p = 0.057; abstract, t37 = 1.37, p = 0.180) lists, between 1PP and 3PP.

To reduce the effort to learn 80 sentences in only one block and then recognize 160 items, we divided the recognition task into two identical parts. In this way, in the first block, participants learned 40 target items and recognized 80 (old + new) sentences; then, a second block was presented, where the remaining 40 target sentences were learned, followed by the recognition part. To reduce any potential order-of-presentation effect, the four blocks order was counterbalanced across participants (sentence length lists at encoding: p = 0.526). In this way, old-new (1PP vs. 3PP) items and presentation blocks were counterbalanced during the task, yielding four possible lists to be administered. See Supplementary Material 1 for an example of one of the four lists presented and its English translation.

In the encoding phase, sentences were presented at the center of the screen for 5 seconds, followed by a 500ms fixation cross. In the recognition phase, the old-new sentences were presented at the center of the screen with no time limit to respond, once responded a 500ms fixation cross appeared. To respond, the participants used the mouse by clicking the two relevant (old-new) screen sections placed in the middle of the bottom part of the PC screen.

2.4. Procedure

The participants were welcomed in a quiet room at the Catholic University of Milan and read and signed the consent form. The PC for this study was a VR-ready Dell G5 15.6inch. Before the tasks, participants were required to provide some demographic information (age, sex, education and handedness). Then, the participants were invited to take a seat 50cm from the PC screen. The spatial memory and the sentence recognition tasks were administered in a counterbalanced order across participants.

The instructions for the intentional spatial memory task at encoding were: “Now you will be in a circular virtual arena and your task is to collect some objects and memorize their locations because you will then be asked to remember them later. You will see one object at a time. You will see each object four times in the same position so that you can remember better its location. To help you to memorize the location you can use the obelisk, the wall, the mountain range, and the fixed clouds as references. You can navigate within the arena with the arrow keys and the mouse. To collect the object, go exactly over it. It will disappear and you will be presented with the next one.” After this phase, the recall instructions were read “You will be asked to put each object in the position where you collected it. However, in random order, either the obelisk or the wall will be removed. Once you are in the location you think is correct, press the space bar to release the object and proceed with the next one. You will be asked to replace each object several times regardless of the correctness of your answer.

The intentional sentence recognition memory task was divided into two blocks (encoding block one and recognition block one, encoding block two and recognition block two). Instructions were displayed before each encoding and recognition phase. Then, the encoding instructions of the first block appeared. “Now you will see some sentences. Your job is to read the sentences carefully and memorize them because you will then be asked to remember them later. Stay focused because the phrases will only be shown for a few seconds on the screen and will change automatically.” The instructions of the recognition phase were “Now you will see some sentences. Your task is to evaluate, using the appropriate buttons on the screen, if the sentence is NEW or OLD. Press OLD with the mouse if you believe that you have seen the sentence among those you memorized a little while ago. Press NEW if you think you have not seen it among those shown before.

2.5. Statistical analyses

R studio (v. 3.6.3) was used to perform the analyses. The design is a full within-subjects design (Transfer Type, 2 levels: abstract vs. concrete; Transfer Perspective, 2 levels: 1PP vs. 3PP) with a continuous moderator (egocentric–allocentric Proclivity index). The egocentric–allocentric Proclivity index was computed as the delta between the average recall error allocentric strategy minus the average recall error egocentric strategy (i.e., the negative index represents greater accuracy for the allocentric strategy and a positive one greater accuracy for the egocentric strategy). We also computed an accuracy-weighted Proclivity index, which is the aforementioned delta multiplied by the sum of the average allocentric and egocentric error. The former is a measure of preference and the latter is an index of preference weighted by accuracy.

The dependent variable was a-prime (A’). A’ is a non-parametric measure of the signal-detection theory in the case of non-normal distributions of hit and false alarm rates (Pastore et al., Reference Pastore, Crawley, Berens and Skelly2003). An A’ equal to 0.5 represents chance responses and an A’ of 1 is the best signal-noise discrimination performance. A’ computation requires collapsing each trial (sentence ID) to compute the average hit and false alarm rate (psycho package in R).

We used the linear mixed-effects model ANOVA (Luke, Reference Luke2017) and Bonferroni corrected simple slope post-hoc tests (Lenth, Reference Lenth2022). Participants were put as a random effect with random intercept to account for inter-individual variability, whereas for RT also sentence ID was also added as a random effect. R formulas were lmer(A’ ~ type×perspective×Proclivity index+(1|ID), REML = T, control = lmerControl(optimizer = “bobyqa”)). Assumptions for the linear method were met by looking at diagnostic plots. Regarding pre-processing of the data, we checked regardless of the experimental conditions any random guessing patterns of A’ and old-new response bias (A’ = 0, bias = 1; one participant removed). Analyses were carried out on 36/37 participants. Then, we used the interquartile range detection method to identify abnormal trials on sentence (377/5920) and spatial (41/1216) memory tasks and outliers were replaced as missing responses, which can be handled by the linear mixed-effects model (Brown, Reference Brown2021). Effect size (η2p) was interpreted according to small = 0.01, medium = 0.06, and large = 0.14 (Richardson, Reference Richardson2011). The alpha level was set to 0.05 for all analyses. Data are fully available at OSF (https://osf.io/yvhra/). The study was not pre-registered.

3. Results

The average hit rate of the sentence recognition task, regardless of the conditions, was 64.02 (SD = 17.36). Participants were better at recalling item spatial locations during the allocentric (wall) compared to the egocentric (obelisk) condition, the average egocentric error (expressed as virtual meters) was 12.56 (SD = 4.49) and the average allocentric error was 11.21 (SD = 4.54) and this difference was significant (p < 0.001). Concerning the spatial task, we did not find any main effect of testing effect across trials (p = 0.28) or testing effect interaction with egocentric and allocentric conditions (p = 0.76). Additionally, any main block order effect (p = 0.08) or block order effect interaction with egocentric and allocentric conditions (p = 0.21) was found. See Table 1 for a detailed summary of the transfer sentence recognition task.

Table 1. Transfer sentence recognition task performance

Note: Median and inter-quartile range are reported. CR: correct rejection; FA: false alarms; RT: reaction times; A′: A prime; 1PP: first-person perspective; 3PP: third-person perspective.

3.1. Egocentric–allocentric proclivity and recognition accuracy

To study the impact of egocentric and allocentric recall strategy proclivity on transfer action sentence recognition, we used a within-subjects linear mixed-effects ANOVA with Proclivity index as moderator and Transfer Perspective and Transfer Type as categorical variables.

Concerning the main effects, we did not find any significant results (Transfer Perspective, p = 0.070; Transfer Type, p = 0.172; P, p = 0.175). These two-term interactions were not significant (Transfer Perspective by Transfer Type, p = 0.331; P by Transfer Type, p = 0.594). The three-term interaction was not significant (Proclivity index by Transfer Perspective by Transfer Type, p = 0.592).

Importantly, we found a two-term significant Proclivity index by Transfer Perspective interaction (F1 = 24.19, p < 0.001, η2p = 0.19, 95%CI [0.09, 1.00]). Simple slope analysis showed that A’ changed as a function of the Proclivity index trend for 3PP sentences (β = 0.04, SE = 0.01, p = 0.009) but not for 1PP (p = 1). A positive Proclivity index (egocentric proclivity) predicts higher A’ scores for 3PP sentences, whereas a negative Proclivity (allocentric proclivity) predicts lower A’ for 3PP sentences; furthermore, the contrast between the slopes was significant (est. diff. = −0.04, SE = 0.01, p < 0.001), indicating different patterns of sentence recognition as a function of spatial frame of reference proclivity and transfer perspective. Conversely, no impact emerged of Proclivity index on 1PP phrase recognition. Figure 2 shows the Proclivity index results. No other significant result was found. The results were replicated using the accuracy-weighted Proclivity index.

Figure 2. Main results of the experiment.

Note: Negative Proclivity index values represent participants with better accuracy in the allocentric than the egocentric recall strategy condition, positive Proclivity index values represent participants with better accuracy during the egocentric than the allocentric recall strategy condition. A’: A prime; 1PP: first-person perspective; 3PP: third-person perspective transfer.

4. Discussion

This study aimed to investigate whether individual preferences for egocentric and allocentric spatial memory influenced action language memory, particularly focusing on the recognition of concrete and abstract transfer sentences where either the participant or another person serves as the agent. Participants’ spatial memory performance was assessed using a task that involved both egocentric and allocentric recall strategies. Using a continuous index (Proclivity index, a delta between egocentric and allocentric average performance), participants were evaluated as being more accurate with the egocentric strategy compared to the allocentric one or vice versa. Memory for transfer action sentences was evaluated with an old-new recognition paradigm using a signal detection parameter.

First, we did not find any impact of proclivity on 1PP phrases for either egocentric or allocentric participants. Second, we found that the greater the egocentric proclivity in the participants, the greater the recognition accuracy for 3PP sentences, regardless of the transfer type. Consequently, the greater the allocentric proclivity, the lower the recognition accuracy for 3PP sentences, regardless of the transfer type. Lastly, the results did not show any differences in transfer type.

According to our predictions, we expected better recognition for 1PP sentences for egocentric participants and better recognition for 3PP sentences for allocentric participants. In contrast, we found that egocentric preference facilitates 3PP sentence recognition compared to an allocentric preference while neither egocentric nor allocentric proclivity had an impact on 1PP sentence recognition. Lastly, we did not predict differences concerning the type of action sentence (abstract vs. concrete); this hypothesis was confirmed instead.

Regarding the first finding, a previous study (Tuena et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023) showed that when two-character sentences depict a concrete transfer and point of view of the reader as the agent, simulation is facilitated and occurs in a ‘default’ manner with transfer action sentence-picture verification task (see, also Brunyé et al., Reference Brunyé, Ditman, Mahoney, Augustyn and Taylor2009, Reference Brunyé, Ditman, Giles, Holmes and Taylor2016). Such facilitation also arises for recognition memory with concrete sentences without transfer (Ditman et al., Reference Ditman, Brunyé, Mahoney and Taylor2010). This might explain why no impact of egocentric or allocentric preference for 1PP sentences was found: additional support of spatial cognition processes is not required during 1PP sentence simulation. Nonetheless, a study (Vukovic & Williams, Reference Vukovic and Williams2015) showed that egocentric participants simulate faster transfer action sentence-picture pairs where the participant is the agent, rather than when the agent is someone else. Yet in their study, the authors used a dichotomous proclivity classification rather than a continuous index, sentences without transfer, and a spatial orientation rather than a spatial memory task. These differences might account for dissimilar results. However, consistent with this study, we found no effect of allocentric proclivity on 1PP sentences. While our data did not reveal a significant relationship between egocentric–allocentric proclivity and 1PP transfer sentence simulation from memory, further research would be needed to draw definitive conclusions about the absence of such effects. Our findings suggest the possibility that 1PP transfer sentence simulation from memory may operate independently of egocentric–allocentric proclivity, though this interpretation should be considered preliminary

Regarding the second finding, we found that spatial memory preferences and ability come into play when non-self-related simulation processes are required. Tuena et al. (Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023) found that the participant can simulate concrete transfer phrases when (s)he is not the agent of the action while retaining the spatial perspective of the patient (the participant’s point of view); however, the simulation of the action occurs at higher processing cost (slower reaction times) for the individual compared to a ‘default’ 1PP transfer action sentence-picture matching condition. This is also in line with a study (Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017) that showed that the greater the egocentric index, the greater the magnitude of the transfer action sentence-picture verification of concrete sentences.

Our results suggest that egocentric ability is a key process to simulate from memory sentences when the agent is someone else. Indeed, egocentric spatial representations are rooted in the sensorimotor states of the body (Chersi & Burgess, Reference Chersi and Burgess2015). This is in line with studies on spatial perspective taking that suggest that to assume someone else’s perspective, an egocentric translocation of our body coordinates is required (Vogeley et al., Reference Vogeley, May, Ritzl, Falkai, Zilles and Fink2004; Vogeley & Fink, Reference Vogeley and Fink2003). Assuming the perspective of someone else should not be defined as a pure allocentric process, rather it requires egocentric mechanisms, such that the translation from 1PP to 3PP encompasses an egocentric to allocentric switch first and then an allocentric to egocentric (new egocentric 3PP) switch (Vogeley et al., Reference Vogeley, May, Ritzl, Falkai, Zilles and Fink2004; Vogeley & Fink, Reference Vogeley and Fink2003). Consequently, egocentric memory appears to be a crucial aspect to simulate sentences when the reader is not the agent of an action sentence and participants simulate sensorimotor states of someone else (see Tuena et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023). However, our results are not in line with a previous study (Vukovic & Williams, Reference Vukovic and Williams2015) that showed that allocentric participants do not show any facilitation or inhibition of action language simulation. On the contrary, we showed that allocentric participants are not facilitated in 3PP sentence simulation. An allocentric representation that is not body-based cannot help with sentence simulation of someone else’s actions. Moreover, we did not provide allocentric participants with specific task instructions to manipulate their spatial model of the phrase (Vukovic & Williams, Reference Vukovic and Williams2015) but rather just asked them to remember and then recognize the stimuli presented. We extended previous research by demonstrating that participants who prefer to use an egocentric memory strategy are more accurate (compared to allocentric participants) in recognizing 3PP transfer sentences. In other words, simulation (Ditman et al., Reference Ditman, Brunyé, Mahoney and Taylor2010) of 3PP transfer sentences from memory is facilitated in individuals that have an egocentric memory strategy preference.

Lastly, we found no difference in recognition performance based on spatial frame of reference proclivity, regardless of the type of transfer. This finding is in line with strong embodiment positions that state that sensorimotor states support both concrete and abstract language representations (Glenberg et al., Reference Glenberg, Sato, Cattaneo, Riggio, Palumbo and Buccino2008; Meteyard et al., Reference Meteyard, Cuadrado, Bahrami and Vigliocco2012; Sakreida et al., Reference Sakreida, Scorolli, Menz, Heim, Borghi and Binkofski2013). We extended previous findings by showing that egocentric–allocentric proclivity has no impact on concrete and abstract two-character sentences when simulated from memory.

A significant departure from previous literature (Tuena et al., Reference Tuena, Di Lernia, Rodella, Bellinzona, Riva, Costello and Repetto2023; Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017; Vukovic & Williams, Reference Vukovic and Williams2015) is the use of a VR spatial memory task and a sentence recognition memory task (rather than an ‘online’ verification task) with transfer sentences to further investigate action language and spatial cognition preferences. A second key aspect is the use of VR, this is a promising technology that can be used to study spatial cognition mechanisms in a standardized way while retaining realism depending on technological features and task characteristics. In contrast to standard cognitive task stimuli (e.g., pre-recorded videos), VR can provide combined visual and bodily (e.g., motor commands) information close to real-world spatial cognition even when used in its 2D (e.g., PC monitor and joystick) setup (Chen et al., Reference Chen, King, Burgess and O’Keefe2013; Tuena et al., Reference Tuena, Serino, Pedroli, Stramba-Badiale, Riva and Repetto2021). Conversely, pre-recorded videos used in previous research on this topic (e.g., Vukovic & Shtyrov, Reference Vukovic and Shtyrov2017; Vukovic & Williams, Reference Vukovic and Williams2015), while retaining visual information and optic flow, do not allow the participants to actively control navigation with motor commands and route decision-making, which is what happens as we navigate. Such aspects make VR navigation closer to cognitive processes that occur during real-world navigation (Tuena et al., Reference Tuena, Serino, Dutriaux, Riva and Piolino2019).

It is important to note that embodied language is influenced by both contextual and individual differences (Ibáñez et al., Reference Ibáñez, Kühne, Miklashevsky, Monaco, Muraki, Ranzini and Tuena2023). We demonstrated that spatial memory preferences affect simulation from memory of both abstract and concrete transfer sentences involving an agent and a patient. Careful consideration of metrics, sentence types, characters in the sentence, individual preferences, and task procedure/instructions is mandatory in the context of embodied language.

However, we must acknowledge some limitations of our study. First, we did not control for the action or spatial perspective from which the participants encoded and/or recognized the transfer action sentences; therefore, it was not possible to draw any conclusions regarding the spatial grounding hypothesis. Second, overall recognition accuracy is not high, probably due to the intrinsic task difficulty that might have affected our results. Third, our study focused on young adults, employing an age range (18–30) commonly used in spatial navigation studies using the same VR landmark-based navigation task (Doeller et al., Reference Doeller, King and Burgess2008) and in studies comparing allocentric and egocentric abilities across age groups (e.g., Rosenbaum et al., Reference Rosenbaum, Winocur, Binns and Moscovitch2012). Therefore, we acknowledge that our results cannot be generalized to other age groups, given the well-known differences between young and older adults in spatial performance (Lester et al., Reference Lester, Moffat, Wiener, Barnes and Wolbers2017). Future research should investigate how individual spatial cognition preferences influence transfer action sentence simulation across different age groups.

To conclude, this study suggests that recognition memory performance for transfer action sentences with a third-person agent is associated with participants’ egocentric spatial memory abilities. The observed pattern aligns with perspectives on sensorimotor simulation from memory, though the precise mechanisms require further investigation. While these findings contribute to our understanding of how individual differences in spatial cognition may relate to sentence processing and sensorimotor representations, further research is needed to establish direct connections to action language simulation processes.

Supplementary material

The supplementary material for this article can be found at http://doi.org/10.1017/langcog.2025.17.

Data availability statement

Data are fully available at OSF https://osf.io/yvhra/

Author contribution

Conceptualization: C.T., S.S.; Writing – first draft: C.T.; Editing: C.T., S.S., C.R., D.D.L.; Data collection: C.T.; Task: C.T.; Study method: C.T., C.R.; Statistical analyses: C.T.; Supervision: G.R., S.S., C.R. All the authors approved the final version of this manuscript.

Acknowledgments

The work was partially supported by the Italian Ministry of Health.

Funding statement

This work was financed by the Italian Ministry of Education, University, and Research (MIUR) through PRIN-2017 “INSPECT” (Project 2017JPMW4F).

Competing interests

The authors declare none

Footnotes

C.R. and S.S. have contributed equally.

References

Anwyl-Irvine, A., Massonnié, J., Flitton, A., Kirkham, N., & Evershed, J. (2020). Gorilla in our MIDST: An online behavioral experiment builder. Behavior Research Methods, 52, 388407. https://doi.org/10.1101/438242CrossRefGoogle ScholarPubMed
Beveridge, M. E. L., & Pickering, M. J. (2013). Perspective taking in language: Integrating the spatial and action domains. Frontiers in Human Neuroscience, 7, 577. https://doi.org/10.3389/fnhum.2013.00577CrossRefGoogle ScholarPubMed
Borghi, A. M. (2024). Mechanistic explanation and the integration between language and action. Cognitive Neuroscience, 15, 9899.CrossRefGoogle ScholarPubMed
Borreggine, K. L., & Kaschak, M. P. (2006). The action-sentence compatibility effect: It’s all in the timing. Cognitive Science, 30, 10971112. https://doi.org/10.1207/s15516709cog0000_91CrossRefGoogle ScholarPubMed
Brown, V. A. (2021). An introduction to linear mixed-effects modeling in R. Advances in Methods and Practices in Psychological Science, 4(1). https://doi.org/10.1177/2515245920960351CrossRefGoogle Scholar
Brunyé, T. T., Ditman, T., Giles, G. E., Holmes, A., & Taylor, H. A. (2016). Mentally simulating narrative perspective is not universal or necessary for language comprehension. Journal of Experimental Psychology: Learning, Memory, and Cognition, 42, 15921605.Google ScholarPubMed
Brunyé, T. T., Ditman, T., Mahoney, C. R., Augustyn, J. S., & Taylor, H. A. (2009). When you and I share perspectives: Pronouns modulate perspective taking during narrative comprehension. Psychological Science, 20, 2732. https://doi.org/10.1111/j.1467-9280.2008.02249.xCrossRefGoogle ScholarPubMed
Burgess, N. (2008). Spatial cognition and the brain. Annals of the New York Academy of Sciences, 1124, 7797. https://doi.org/10.1196/annals.1440.002CrossRefGoogle ScholarPubMed
Chen, G., King, J. A., Burgess, N., & O’Keefe, J. (2013). How vision and movement combine in the hippocampal place code. Proceedings of the National Academy of Sciences of America, 110, 378383. https://doi.org/10.1073/pnas.1215834110CrossRefGoogle ScholarPubMed
Chersi, F., & Burgess, N. (2015). The cognitive architecture of spatial navigation: hippocampal and striatal contributions. Neuron, 88, 6477. https://doi.org/10.1016/j.neuron.2015.09.021CrossRefGoogle ScholarPubMed
Diefenbach, C., Rieger, M., Massen, C., & Prinz, W. (2013). Action-sentence compatibility: The role of action effects and timing. Frontiers in Psychology, 4, 113. https://doi.org/10.3389/fpsyg.2013.00272CrossRefGoogle ScholarPubMed
Díez-Álamo, A. M., Glenberg, A. M., Díez, E., Alonso, M. A., & Fernandez, A. (2020). The linguistic looming effect. Journal of Memory and Language, 114, 104147. https://doi.org/10.1016/j.jml.2020.104147CrossRefGoogle Scholar
Ditman, T., Brunyé, T. T., Mahoney, C. R., & Taylor, H. A. (2010). Simulating an enactment effect: Pronouns guide action simulation during narrative comprehension. Cognition, 115, 172178. https://doi.org/10.1016/j.cognition.2009.10.014CrossRefGoogle ScholarPubMed
Doeller, C. F., King, J. A., & Burgess, N. (2008). Parallel striatal and hippocampal systems for landmarks and boundaries in spatial memory. Proceedings of the National Academy of Sciences of the United States of America, 105, 59155920.CrossRefGoogle ScholarPubMed
Dove, G. O. (2023). Rethinking the role of language in embodied cognition. Philosophical Transactions of the Royal Society B, 378, 20210375.CrossRefGoogle ScholarPubMed
Ekstrom, A. D., Arnold, A. E. G. F., & Iaria, G. (2014). A critical review of the allocentric spatial representation and its neural underpinnings: Toward a network-based perspective. Frontiers in Human Neuroscience, 8, 803. https://doi.org/10.3389/fnhum.2014.00803CrossRefGoogle Scholar
Engelkamp, J. (1998). Memory for actions. Psychology Press/Taylor & Francis (UK).Google Scholar
Ghandhari, M., Fini, C., Da Rold, F., & Borghi, A. M. (2020). Different kinds of embodied language: A comparison between Italian and Persian languages. Brain and Cognition, 142, 105581. https://doi.org/10.1016/j.bandc.2020.105581CrossRefGoogle Scholar
Gianelli, C., Farnè, A., Salemme, R., Jeannerod, M., & Roy, A. C. (2011). The agent is right: When motor embodied cognition is space-dependent. PLoS ONE, 6(9), e25036. https://doi.org/10.1371/journal.pone.0025036CrossRefGoogle ScholarPubMed
Glenberg, A. M., & Gallese, V. (2012). Action-based language: A theory of language acquisition, comprehension, and production. Cortex, 48, 905922. https://doi.org/10.1016/j.cortex.2011.04.010CrossRefGoogle Scholar
Glenberg, A. M., & Kaschak, M. P. (2002). Grounding language in action. Psychonomic Bulletin and Review, 9, 558565. https://doi.org/10.1109/TAMD.2011.2140890CrossRefGoogle ScholarPubMed
Glenberg, A. M., Sato, M., Cattaneo, L., Riggio, L., Palumbo, D., & Buccino, G. (2008). Processing abstract language modulates motor system activity. Quarterly Journal of Experimental Psychology, 61, 905919. https://doi.org/10.1080/17470210701625550CrossRefGoogle ScholarPubMed
Goeke, C., König, P., & Gramann, K. (2013). Different strategies for spatial updating in yaw and pitch path integration. Frontiers in Behavioral Neuroscience, 7, 5. https://doi.org/10.3389/fnbeh.2013.00005CrossRefGoogle ScholarPubMed
Goeke, C., Kornpetpanee, S., Köster, M., Fernández-Revelles, A. B., Gramann, K., & König, P. (2015). Cultural background shapes spatial reference frame proclivity. Scientific Reports, 5, 113. https://doi.org/10.1038/srep11426CrossRefGoogle ScholarPubMed
Gramann, K., Müller, H. J., Eick, E. M., & Schönebeck, B. (2005). Evidence of separable spatial representations in a virtual navigation task. Journal of Experimental Psychology: Human Perception and Performance, 31, 11991223. https://doi.org/10.1037/0096-1523.31.6.1199Google Scholar
Guderian, S., Dzieciol, X. A. M., Gadian, D. G., Jentschke, X. S., Doeller, C. F., Burgess, N., … Vargha-Khadem, F. (2015). hippocampal volume reduction in humans predicts impaired allocentric spatial memory in virtual-reality navigation. The Journal of Neuroscience, 35, 1412314131. https://doi.org/10.1523/JNEUROSCI.0801-15.2015CrossRefGoogle ScholarPubMed
Ianì, F. (2019). Embodied memories: Reviewing the role of the body in memory processes. Psychonomic Bulletin and Review, 26, 17471766. https://doi.org/10.3758/s13423-019-01674-xCrossRefGoogle ScholarPubMed
Ibáñez, A., Cardona, J. F., Dos Santos, Y. V, Blenkmann, A., Aravena, P., Roca, M., … Bekinschtein, T (2013). Motor-language coupling: Direct evidence from early Parkinson’s disease and intracranial cortical recordings. Cortex, 49, 968984. https://doi.org/10.1016/j.cortex.2012.02.014CrossRefGoogle ScholarPubMed
Ibáñez, A., Kühne, K., Miklashevsky, A., Monaco, E., Muraki, E. J., Ranzini, M., … Tuena, C. (2023). Ecological meanings: A consensus paper on individual differences and contextual influences in embodied language. Journal of Cognition, 6(1), 59. https://doi.org/10.31219/osf.io/ej5y3CrossRefGoogle ScholarPubMed
Jeneson, A., & Squire, L. R. (2011). Working memory, long-term memory, and medial temporal lobe functionial temporal lobe function. Learning & Memory, 19, 1525.CrossRefGoogle Scholar
Lenth, R. V. (2022). Emmeans: Estimated marginal means, aka least-squares means. (R package version 1.8.1-1). https://CRAN.R-project.org/package=emmeansGoogle Scholar
Lester, A. W., Moffat, S. D., Wiener, J. M., Barnes, C. A., & Wolbers, T. (2017). The aging navigational system. Neuron, 95, 10191035. https://doi.org/10.1016/j.neuron.2017.06.037CrossRefGoogle ScholarPubMed
Luke, S. G. (2017). Evaluating significance in linear mixed-effects models in R. Behavioural Research, 49, 14941502. https://doi.org/10.3758/s13428-016-0809-yCrossRefGoogle ScholarPubMed
Majid, A., Bowerman, M., Kita, S., Haun, D. B. M., & Levinson, S. C. (2004). Can language restructure cognition? The case for space. Trends in Cognitive Sciences, 8, 108114. https://doi.org/10.1016/j.tics.2004.01.003CrossRefGoogle ScholarPubMed
Meteyard, L., Cuadrado, S. R., Bahrami, B., & Vigliocco, G. (2012). Coming of age: A review of embodiment and the neuroscience of semantics. Cortex, 48, 788804. https://doi.org/10.1016/j.cortex.2010.11.002CrossRefGoogle Scholar
Morey, R. D., Kaschak, M. P., Díez-Álamo, A. M., Glenberg, A. M., Zwaan, R. A., Lakens, D., … Ziv-Crispel, N. (2022). A pre-registered, multi-lab non-replication of the action-sentence compatibility effect (ACE). Psychonomic Bulletin & Review, 29, 613626.CrossRefGoogle ScholarPubMed
Papesh, M. H. (2015). Just out of reach: On the reliability of the action-sentence compatibility effect. Journal of Experimental Psychology: General, 144, e116e141. https://doi.org/10.1037/xge0000125CrossRefGoogle ScholarPubMed
Pastore, R. E., Crawley, E. J., Berens, M. S., & Skelly, M. A. (2003). “Nonparametric” A′ and other modern misconceptions about signal detection theory. Psychonomic Bulletin and Review, 10, 556569. https://doi.org/10.3758/BF03196517CrossRefGoogle ScholarPubMed
Richardson, J. T. E. (2011). Eta squared and partial eta squared as measures of effect size in educational research. Educational Research Review, 6, 135147. https://doi.org/10.1016/j.edurev.2010.12.001CrossRefGoogle Scholar
Rosenbaum, R. S., Winocur, G., Binns, M. A., & Moscovitch, M. (2012). Remote spatial memory in aging: All is not lost. Frontiers in Aging Neuroscience, 4, 25. https://doi.org/10.3389/fnagi.2012.00025CrossRefGoogle Scholar
Sakreida, K., Scorolli, C., Menz, M. M., Heim, S., Borghi, A. M., & Binkofski, F. (2013). Are abstract action words embodied? an fMRI investigation at the interface between language and motor cognition. Frontiers in Human Neuroscience, 7, 125. https://doi.org/10.3389/fnhum.2013.00125CrossRefGoogle ScholarPubMed
Schuck, N. W., Doeller, C. F., Polk, T. A., Lindenberger, U., & Li, S. C. (2015). Human aging alters the neural computation and representation of space. NeuroImage, 117, 141150. https://doi.org/10.1016/j.neuroimage.2015.05.031CrossRefGoogle ScholarPubMed
Spinnler, H., & Tognoni, G. (1987). Standardizzazione e taratura italiana di test neuropsicologici. Italian Journal of Neurological Sciences, 8, 1120.Google Scholar
Squire, L. R., Wixted, J., & Clark, R. E. (2007). Recognition memory and the medial temporal lobe: A new perspective. Nature Reviews. Neuroscience, 8, 872883. https://doi.org/10.1038/nrn2154CrossRefGoogle ScholarPubMed
Tuena, C., Di Lernia, D., Riva, G., Serino, S., & Repetto, C. (2023). Phonological similarity between words is represented in declarative memory as spatial distance. Psychological Research, 87(8), 24992510. https://doi.org/10.1007/s00426-023-01830-yCrossRefGoogle ScholarPubMed
Tuena, C., Di Lernia, D., Rodella, C., Bellinzona, F., Riva, G., Costello, M. C., & Repetto, C. (2023). The interaction between motor simulation and spatial perspective-taking in action language: A cross-cultural study. Memory and Cognition. https://doi.org/10.3758/s13421-023-01427-1CrossRefGoogle Scholar
Tuena, C., Serino, S., Dutriaux, L., Riva, G., & Piolino, P. (2019). Virtual enactment effect on memory in young and aged populations: A systematic review. Journal of Clinical Medicine, 8, 620. https://doi.org/10.3390/jcm8050620CrossRefGoogle Scholar
Tuena, C., Serino, S., Pedroli, E., Stramba-Badiale, M., Riva, G., & Repetto, C. (2021). Building embodied spaces for spatial memory neurorehabilitation with virtual reality in normal and pathological aging. Brain Sciences, 11, 1067.CrossRefGoogle ScholarPubMed
VandenBos, G. R. (2007). APA dictionary of psychology. American Psychological Association. https://doi.org/10.5860/choice.44-3029Google Scholar
VanVoorhis, C. W., & Morgan, B. L. (2007). Understanding power and rules of thumb for determining sample sizes. Tutorials in Quantitative Methods for Psychology, 3, 4350.CrossRefGoogle Scholar
Vogeley, K., & Fink, G. R. (2003). Neural correlates of the first-person-perspective. Trends in Cognitive Sciences, 7, 3842. https://doi.org/10.1016/S1364-6613(02)00003-7CrossRefGoogle ScholarPubMed
Vogeley, K., May, M., Ritzl, A, Falkai, P., Zilles, K., & Fink, G. R. (2004). Neural correlates of first-person perspective as one constituent of human self-consciousness. Journal of Cognitive Neuroscience, 16, 817827. https://doi.org/10.1162/089892904970799CrossRefGoogle ScholarPubMed
Vukovic, N., & Shtyrov, Y. (2017). Cortical networks for reference-frame processing are shared by language and spatial navigation systems. NeuroImage, 161, 120133. https://doi.org/10.1016/j.neuroimage.2017.08.041CrossRefGoogle ScholarPubMed
Vukovic, N., & Williams, J. N. (2015). Individual differences in spatial cognition influence mental simulation of language. Cognition, 142, 110122. https://doi.org/10.1016/j.cognition.2015.05.017CrossRefGoogle ScholarPubMed
Westfall, J., Kenny, D. A., & Judd, C. M. (2014). Statistical power and optimal design in experiments in which samples of participants respond to samples of stimuli. Journal of Experimental Psychology: General, 143, 20202045.CrossRefGoogle ScholarPubMed
Zwaan, R. A. (2016). Situation models, mental simulations, and abstract concepts in discourse comprehension. Psychonomic Bulletin and Review, 23, 10281034. https://doi.org/10.3758/s13423-015-0864-xCrossRefGoogle ScholarPubMed
Zwaan, R. A., & Radvansky, G. A. (1998). Situation models in language comprehension and memory. Psychological Bulletin, 123, 162185.CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. Virtual environment of the arena.Note: During the encoding phase, both a local landmark (obelisk – egocentric) and a boundary (wall – allocentric) were used to encode the item location. During the recall phase, either the local landmark or the wall was shown to assess egocentric or allocentric memory recall.

Figure 1

Table 1. Transfer sentence recognition task performance

Figure 2

Figure 2. Main results of the experiment.Note: Negative Proclivity index values represent participants with better accuracy in the allocentric than the egocentric recall strategy condition, positive Proclivity index values represent participants with better accuracy during the egocentric than the allocentric recall strategy condition. A’: A prime; 1PP: first-person perspective; 3PP: third-person perspective transfer.

Supplementary material: File

Tuena et al. supplementary material

Tuena et al. supplementary material
Download Tuena et al. supplementary material(File)
File 36 KB