Embodied language processing
![]() | This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these messages)
|
Embodied cognition occurs when an organism’s sensorimotor capacities (ability of the body to respond to its senses with movement), body and environment play an important role in thinking. The way in which a person’s body and their surroundings interacts also allows for specific brain functions to develop and in the future to be able to act.[1] This means that not only does the mind influence the body’s movements, but the body also influences the abilities of the mind. There are three generalizations that are assumed to be true relating to embodied cognition. A person's motor system (that controls movement of the body) is activated when (1) they observe manipulable objects, (2) process action verbs, and (3) observe another individual's movements.[2]
In order to create movement of the body, a person usually thinks (or the brain subconsciously functions) about the movement it would like to accomplish. Embodied language processing asserts that there can also be an opposite influence. This means that moving your body in a certain way will impact how you comprehend, as well as process, language – whether it is an individual word or a complete phrase or sentence. Embodied language processing suggests that the brain resources that are used for perception, action, and emotion are also used during language comprehension.[3] Studies have found that participants are faster at comprehending a sentence when the picture that goes along with it matches the actions described in the sentence.[4] Action and language about action have been found to be connected because the areas of the brain that control them overlap [3] It has been found that action can influence how a person understands a word, phrase, or sentence, but language can also impact a person's actions.
Language Processing
Language comprehension involves three processes that overlap. [citation needed] First, words and phrases are associated with certain objects in the environment or perceived symbols. Secondly, a person must determine how they can interact with the objects; these are called affordances. For example, a chair affords sitting for a human, but not for a tree. Or a pen affords writing for a human but not for a hamster. And finally, a person must look at the different parts of the sentence to determine how the affordances combine in order to accomplish a goal. This process is called the Indexical Hypothesis. [5]
To summarize this hypothesis, an understanding of language results from a simulation of the actions that are implied by the meaning of the sentence. Put more simply, meaning is based on action.[6]
Actions can also influence a person's reaction time even when it does not make sense with the actual meaning of the sentence. A study conducted asked participants to perform a list of actions. They were then asked to make judgments if a sentence was logical or not. Participants were faster to respond that a sentence was logical if they had performed that action previously. For example, participants were faster to determine that the metaphorical phrase "toss out some ideas" was a logical phrase after they had performed a tossing motion, even though you can't literally toss a plan.[7]
Connection between language and emotion
Some research has claimed that simulating emotions can play an important role in understanding and processing language. In one study, the researchers measured the time needed for a person to comprehend a sentence written with an implied emotion while they were in either a matching or opposite emotional state. The participants who were forced to smile responded with higher humor ratings of cartoons than participants who read these cartoons while frowning. These researchers also found that pleasant sentences were read faster when participants were smiling than when they were frowning. Unpleasant sentences were read slower when participants were smiling than when frowning. These results demonstrate that emotional cues created by the body help in language comprehension. Further research has shown that the same areas in the brain that process emotions also function during language comprehension that has an emotional connection.[8]
Other research has shown that words with a positive connotation, for example: love, are identified faster when the identification process includes an approach action (flexing an arm) than when the action is a withdrawal action (arm extension).[8]
Emotional stimulation affects language comprehension within a sentence or a phrase, but has not been shown to facilitate language comprehension on a single word basis. Whereas a person's motor (action) system influences understanding for both individual word as well as entire sentences [8]
Experiential Trace Hypothesis
Experiential Trace Hypothesis states that each time we interact with the world, traces of that particular experience are left in our brain.[9] These traces can be accessed again when a person thinks of words or sentences that remind them of that experience. Additionally, these traces in our brain are linked to the action that they are related to.[9] Words and sentences become those cues that retrieve these traces from our mind. Researchers have studied if the previous experience with a word, such as its location (up or down) in space, affects how people understand and then respond to that word.[10] In one experiment, researchers hypothesized that if reading an object word also activates a location that is linked to that noun, then the following action response should be compatible with that association.[10] They found that participants were faster to push a button higher than another button when the word was associated with being "up" or "above" than when the button was lower than the other for words associated with "up" and "above". The results of this study displayed that participants were faster to respond when the location of the word and the action they had to perform were similar. This demonstrates that language processing and action are connected. This research also found that the location information of a word is automatically activated after seeing the word.[10] In a similar study, it was discovered that participants were equally as fast at responding to words that were associated with either an upward or downward location when the buttons to respond to these words were horizontal – meaning that the experiential trace effect was ruled out when the responding action did not link to either of the locations that were activated.[4]
Experiential-Simulation Theory of Language Understanding
Some theorists have proposed an experiential-simulation approach of language understanding. They argue that previous experiential traces related to a word may be reactivated at a later stage when accessing the meaning of the same word. This has been highlighted through the example of encountering the word ‘airplane’ in a situation where someone points to an airplane in the sky, thus making one look upwards. These experiential traces, e.g. ‘looking upwards’ are later reactivated when accessing the meaning of the word ‘airplane'. Similarly, another example might be when a person accesses the meaning of the word ‘snail’ , they might also access experiential traces associated with this word, e.g. ‘looking downwards’ (likely towards the ground). [11]
Language Comprehension and Motor Systems Involved in Action
Concrete verbs
As a result of previous experience to certain words, several studies have found that the action associated with a certain word is also activated in the motor cortices when processing that same word. For example, using event-related functional magnetic resonance imaging (fMRI), it was discovered that exposure to concrete action verbs referring to face, arm, or leg actions (e.g., to lick, pick, kick) activated motor regions that are stimulated when making actions with the foot, hand, or mouth.[12]
Abstract verbs
However, findings are not as clear cut when abstract verbs are involved. Embodied theories of language comprehension assume that abstract concepts, as well as concrete ones, are grounded in the sensorimotor system (Jirak et al., 2010).[13] Some studies have investigated the activation of motor cortices using abstract and also concrete verbs, examining the stimulation of the motor cortices when comprehending literal action verbs (concrete) vs. the metaphorical usage of the same action verbs (abstract). One such study used fMRI to study participants whilst they viewed actions performed by the mouth, hand or foot, and read literal and metaphorical sentences related to the mouth hand or foot. This study found activation in the premotor cortex for literal action (e.g. “grasping the scissors”) but not for metaphorical usage (e.g. “grasping the idea”).[14] These findings suggest that the assumption of embodied theories that abstract concepts, as well as concrete ones, are grounded in the sensorimotor system may not be true.
However in contrast, other research has found motor cortex activation for the metaphorical usage of action verbs. One such study investigated cortical activation during comprehension of literal and idiomatic sentences using Magnetoencephalography (MEG). During a silent reading task, participants were presented with stimuli which included both literal and metaphorical arm-related action verbs, e.g. “Mary caught the fish” versus “Mary caught the sun”, and also literal and metaphorical leg-related action verbs, e.g. “Pablo jumped on the armchair” versus “Pablo jumped on the bandwagon”. This study found that processing of abstract verbs (idioms in this case) did indeed activate motor regions of the brain, activating anterior fronto-temporal activity very early compared to literal verbs.[15]
Concrete and abstract sentences
Other studies have investigated activation of the motor system during comprehension of concrete and abstract sentences. Using Transcranial Magnetic Stimulation (TMS) and a behavioural paradigm, one study investigated whether listening to action-related sentences activated activity within the motor cortices. This was investigated using Motor Evoked Potentials (MEPs) from the TMS which were recorded from hand muscles when stimulating the hand motor area, and from foot and leg muscles when stimulating the foot motor area. Participants were presented with sentences relating to hand or foot actions. As control, participants listened to sentences containing abstract content. The study found there was indeed activation of the motor cortices whilst listening to sentences expressing foot/leg and hand/arm actions. This activation specifically concerned the areas of the motor system ‘where the effector involved in the processed sentence is motorically represented’ (pp. 360). Specifically, the results showed that listening to hand-action-related sentences prompted a decrease of MEP amplitude recorded from hand muscles and listening to foot-action-related sentences prompted a decrease of MEP amplitude recorded from foot muscle.[16]
Action-Sentence Compatibility Effect (ACE)
Sentence processing can activate motor neurons based on the actions referred to in the sentence. In one study, researchers asked participants to make judgments on whether a sentence was sensible or not. For example, "You handed Courtney the notebook" versus "Courtney handed you the notebook". They asked participants in one condition to push a button farther away from their body if the sentence was logical and a button close to their body when it wasn't logical. The results of this study demonstrated that participants were faster at pushing the "sentence is logical" button when the action in the sentence matched the action required by them to push the correct button.[8] This means if the sentence read "you handed Courtney the notebook", the participants were faster to push the button that was farther away from them when this button meant the sentence was logical. The depicted motion in these sentences affected the amount of time required to understand the sentences that described the motion that is in the same direction. This effect has been shown to apply to sentences that describe concrete actions (putting a book on a shelf) as well as more abstract actions (you told the story to the policeman).[6]
Other studies have tried to understand the ACE phenomenon by examining the modulation of motor resonance during language comprehension. In one study participants were asked to read sentences containing a frame of between one and three words. The participants had to rotate a knob, in one direction for half of the experiment and in the other direction for the other half. Each 5° of rotation induced the presentation of a new frame. Each of the sentences described actions involving manual rotation. In these, the rotation direction would or would not match the rotation direction implied by the sentence. Earlier studies, such as that by Glenberg & Kaschak (2002), examined motor resonance in responses to sentences presumably given after the sentence was read. In contrast, results of this study revealed that motor resonance had dissipated before the end of the sentence, with motor resonance occurring on the verb. This study made use of comprehension questions rather than sensibility sentences. The researchers have argued that this created a more naturalistic reading situation, so it could be argued that the results of this study are deemed more suitable because they are in regards to more naturalistic language. Overall, the researchers have concluded that motor resonance is quite immediate and short-lived and that duration of the effect is modified by linguistic context.[17]
Neurophysiological evidence has also been presented to prove an ACE. This research used a behavioural paradigm as well as Event-Related Potential (ERP) to record brain activity, allowing the researchers to explore the neural brain markers of the ACE paradigm in semantic processing and motor responses. ERP was particularly beneficial in helping the researchers to investigate the bi-directional hypothesis of action-sentence comprehension, which proposes that language processing facilitates movement and movement also facilitates language comprehension. In the study participants listened to sentences describing an action that involved an open hand, a closed hand or no manual action. They were then required to press a button to indicate their understanding of the sentence. Each participant was assigned a hand-shape, either closed or open, which was required to activate the button. As well as two groups (closed or open hand-shapes), there were three different categories relating to hand-shape: compatible, incompatible and neutral. Behavioural results from the study showed that participants responded quicker when the hand-shape required to press the response-button was compatible with the hand-shape inferred by the sentence. ERP results provided evidence to support the bi-directional hypothesis, showing that cortical markers of motor processes were affected by sentence meaning, therefore providing evidence for a semantics-to-motor effect. ERP results also demonstrated a motor-to-semantics effect as brain markers of comprehension were modified by motor effects.[18]
The Action-Compatibility Effect also states that the brain resources used to plan and carry out actions are also used in language comprehension; therefore, if an action implied in a sentence is different from the suggested response, there is interference within these brain resources.[6]
Word Activation
Other studies have demonstrated that reading an object name interferes with how a person plans on grasping that object.[19] It was also found that similar words can prime similar actions. Playing the piano and using a typewriter both utilize similar motor actions; these words prime each other in a word decision task.[19] These studies have concluded that activation of motor decisions occur automatically when exposed to action to action-related words.[19]
Actions emphasize meaning
Many studies have shown how body movements and speech can be combined to emphasize meaning (often called gesturing). A person can observe the actions of another to help them comprehend what that person is saying.[7] For example, if a person is pointing repeatedly, it helps the listener to understand that the direction being inferred is very important; whereas if it was a casual point in the general direction, the location of the object may not be as necessary to comprehend what the speaker is saying. Another example may be the stomping of one’s foot. This can help the listener to understand the anger and frustration being conveyed by the speaker.[citation needed]
Implications
Many studies have demonstrated that people’s understanding of words and sentences can influence their movements and actions as well as the opposite – peoples’ actions can influence how quickly they can comprehend a word or sentence. This knowledge is important for many reasons. One study looked at the impact of embodied cognition in a classroom setting to facilitate and enhance language learning. For a child, there is a difference between oral language learning and reading. In oral language learning, the mapping between a symbol (word) and the object is common – often brought about by gesturing to the object.[20] However, when a child is learning to read, they focus on the letter-sound combinations and the correct pronunciation of the words. Usually, the object the words are referring to, are not immediately connected with the word so an association between the word and object isn't immediately made.[20] The researchers of this study suggest the Moved by Reading intervention which consists of two parts – Physical Manipulation stage and an Imagined Manipulation stage.[20] In physical manipulation, the child reads a sentence and then is instructed to act out that sentence with available toys.[20] This forces the child to connect words with objects and their actions. In the imagined manipulation stage, the child reads the sentence and is then asked to imagine how they would interact with toys to act out the sentence.[20] They studied this further and discovered that it is possible for these children to still benefit from the effects of embodied cognition when they manipulate objects on a computer screen.[20] This embodied cognition software can help children facilitate language comprehension.[citation needed]
References
- ^ Cowart, M. (2005, Jul. 8 ). Embodied Cognition. http://www.iep.utm.edu/embodcog/
- ^ Mahon, B. Z, and A. Caramazza. (2008). A critical look at the embodied cognition hypothesis and a new proposal. Journal of Physiology - Paris, 102 pp. 59–70.
- ^ a b Rueschemeyer, S. et al. (2010). Effects of Intentional Motor Actions on Embodied Language Processing. Experimental Psychology, 57 (4), pp. 260−66. doi:10.1027/1618-3169/a000031
- ^ a b Zwann, R. A. (2002). Language Comprehenders Mentally Represent the Shape of Objects. Psychological Science, 13 (2), pp. 168–71
- ^ Glenberg, A. M. (2002). The Indexical hypothesis: Meaning from language, world, and image. Words and Images: Working Together-Working Differently, Albex.
- ^ a b c Glenberg, A. M, and M. Kaschak. (2002). Grounding Language in Action. Psychnomic Bulletin & Review, 9 (3), pp. 558–65.
- ^ a b Gibbs, R. W. (2006). Language and Communication. Embodiement and Cognitive Science, pp. 158–207.
- ^ a b c d Havas, D. A. et al. (2007). Emotion simulation during language comprehension. Psychonomic Bulletin & Review, 14 (3), pp. 436−41.
- ^ a b Zwaan R.A. & Pecher, D. The grounding of cognition: The role of perception and action in memory, language, and thinking. Cambridge, UK: Cambridge University Press.
- ^ a b c Lachmair, M. et al. (2011). Root versus roof: automatic activation of location information during word processing. Psychonomic Bulletin & Review, 18 pp. 1180–88. doi:10.3758/s13423-001-0158-x
- ^ Zwaan, R. A., & Madden, C. J. (2005). Embodied sentence comprehension. Grounding cognition: The role of perception and action in memory, language, and thinking, 224-245.
- ^ Hauk, O., Johnsrude, I., & Pulvermüller, F. (2004). Somatotopic representation of action words in human motor and premotor cortex. Neuron, 41(2), 301-307.
- ^ Jirak, D., Menz, M. M., Buccino, G., Borghi, A. M., & Binkofski, F. (2010). Grasping language–A short story on embodiment. Consciousness and cognition, 19(3), 711-720.
- ^ Aziz-Zadeh, L., Wilson, S. M., Rizzolatti, G., & Iacoboni, M. (2006). Congruent embodied representations for visually presented actions and linguistic phrases describing actions. Current Biology, 16(18), 1818-1823.
- ^ Boulenger, V., Shtyrov, Y., & Pulvermüller, F. (2012). When do you grasp the idea? MEG evidence for instantaneous idiom understanding. Neuroimage, 59(4), 3502-3513.
- ^ Buccino, G., Riggio, L., Melli, G., Binkofski, F., Gallese, V., & Rizzolatti, G. (2005). Listening to action-related sentences modulates the activity of the motor system: a combined TMS and behavioral study. Cognitive Brain Research, 24(3), 355-363.
- ^ Zwaan, R. A., & Taylor, L. J. (2006). Seeing, acting, understanding: motor resonance in language comprehension. Journal of Experimental Psychology: General, 135(1), 1.
- ^ Aravena, P., Hurtado, E., Riveros, R., Cardona, J. F., Manes, F., & Ibáñez, A. (2010). Applauding with closed hands: neural signature of action-sentence compatibility effects. PLoS One, 5(7), e11751.
- ^ a b c Fischer, M., & Zwann, R. (2008). Embodied language: A review of the role of the motor system in language comprehension. The Quarterly Journal of Experimental Psychology, 61(6), 825–50. doi:10.1080/17470210701623605
- ^ a b c d e f Glenberg, A. M, and A. Goldberg. (2011). Improving early reading comprehension using embodied CAI. Instructional Science, 39 pp. 27–39. doi:10.1007/s11251-009-9096-7