Mu wave suppression during the perception of meaningless syllables: EEG evidence of motor...

6
UNCORRECTED PROOF Please cite this article in press as: Crawcour, S., et al. Mu wave suppression during the perception of meaningless syllables: EEG evidence of motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neuropsychologia.2009.05.001 ARTICLE IN PRESS G Model NSY 3299 1–6 Neuropsychologia xxx (2009) xxx–xxx Contents lists available at ScienceDirect Neuropsychologia journal homepage: www.elsevier.com/locate/neuropsychologia Mu wave suppression during the perception of meaningless syllables: EEG evidence of motor recruitment 1 2 Stephen Crawcour, Andrew Bowers, Ashley Harkrider, Tim Saltuklaroglu 3 Department of Audiology and Speech Language Pathology, University of Tennessee, 553 S. Stadium Hall, Knoxville, TN 37996-0740, United States 4 5 article info 6 7 Article history: 8 Received 11 December 2008 9 Received in revised form 16 April 2009 10 Accepted 3 May 2009 11 Available online xxx 12 13 Keywords: 14 Mirror neuron 15 Speech perception 16 Electroencephalography 17 Syllables 18 Sensory 19 abstract Motor involvement in speech perception has been recently studied using a variety of techniques. In the current study, EEG measurements from Cz, C3 and C4 electrodes were used to examine the relative power of the mu rhythm (i.e., 8–13Hz) in response to various audio-visual speech and non-speech stimuli, as suppression of these rhythms is considered an index of ‘mirror neuron’ (i.e., motor) activity. Fourteen adult native English speaking females watched and listened to nine audio-video stimuli clips assembled from three different auditory stimuli (speech, noise, and pure tone) combined with three different video stimuli (speech, noise, and kaleidoscope—made from scrambling an image from the visual speech). Rel- ative to the noise–noise (baseline condition), all visual speech conditions resulted in significant levels of suppression, a finding that is consistent with previous reports of mirror activity to visual speech and mu suppression to ‘biological’ stimuli. None of the non-speech conditions or conditions in which speech was presented via audition only resulted in any significant suppression of the mu rhythm in this population. Thus, visual speech perception appears to be more closely associated with motor activity than acoustic speech perception. It is postulated that in this study, the processing demands incurred by the task were insufficient for inducing significant mu suppression via acoustic speech only. The findings are discussed in theoretical contexts of speech perception and the mirror system. We suggest that this technique may offer a cost-efficient, non-invasive technique for measuring motor activity during speech perception. © 2009 Published by Elsevier Ltd. 1. Introduction 20 The processes underlying human speech perception have been 21 widely examined and debated over the last six decades or so. By 22 some accounts, speech is perceived as a function of its acoustic con- 23 stituents and their impact on the auditory system (Klatt, 1979; Kuhl 24 & Miller, 1975; Massaro & Cohen, 1990; Ohala, 1996; Stevens, 1981; 25 Sussman, 1989). Additionally, these acoustic theories generally hold 26 that speech perception and production are distinct processes. Crit- 27 icisms of these perspectives were propagated by studies showing 28 the lack of acoustic invariance in similar speech percepts (Liberman, 29 Cooper, Shankweiler, & Studdert-Kennedy, 1967), the influence of 30 visual stimuli on speech percepts (e.g., McGurk effect; McGurk & 31 MacDonald, 1976), the phenomenon of categorical perception (e.g., 32 Mann & Liberman, 1983), and the limited temporal resolution of 33 the auditory system for processing rapidly changing acoustic stim- 34 uli (Liberman, 1957; Liberman, Delattre, & Cooper, 1952). As such, 35 Liberman and Mattingly (1985) proposed an alternative viewpoint. 36 They suggested that speech is perceptually coded as a sequence of 37 dynamic ‘gestures’, that are representative of the manner in which 38 Corresponding author. E-mail address: [email protected] (T. Saltuklaroglu). they are produced and co-articulated in the vocal tract. It is these 39 gestures that are thought to form the invariants for both perception 40 and production, linking the two processes and allowing them to 41 operate efficiently and effectively, together as one, in a specialized 42 linguistic manner. As such, under this ‘motor theory’, the dynamic 43 architecture of the mechanism employed to produce speech plays 44 an essential role in its perception. 45 Conceptually, motor theory appeared to answer many questions 46 regarding the nature of speech, yet one of its downfalls was a rel- 47 ative dearth of physiological evidence for the neural connectivity 48 between speech perception and production. The discovery of mir- 49 ror neurons in the ventral premotor cortex (area F5) of the macaque 50 monkey (Di Pellegrino, Fadiga, Fogassi, Gallese, & Rizzolatti, 1992; 51 Rizzolatti, Fadiga, Gallese, & Fogassi, 1996) provides compelling 52 evidence for motor involvement in sensory processes and there- 53 fore, a central linking of perception and production. Neurons in this 54 motor region, which is considered to be a homolog of Broca’s area 55 in humans (Rizzolatti & Craighero, 2004), were found to fire both 56 when monkeys performed or observed goal directed actions (e.g., 57 grasping). The location and firing patterns of these ‘mirror neurons’ 58 helped support the notion of motor involvement in speech per- 59 ception (Liberman & Whalen, 2000) and theoretical perspectives 60 of human communication evolving from this observation/action 61 matching system. This neural matching system is thought to 62 0028-3932/$ – see front matter © 2009 Published by Elsevier Ltd. doi:10.1016/j.neuropsychologia.2009.05.001

Transcript of Mu wave suppression during the perception of meaningless syllables: EEG evidence of motor...

N

ME

1

2

S3

D4

5

a6

7

A8

R9

R10

A11

A12

13

K14

M15

S16

E17

S18

S19

120

21

w22

s23

s24

&25

S26

t27

i28

t29

C30

v31

M32

M33

t34

u35

L36

T37

d38

0d

PR

OO

F

ARTICLE IN PRESSG ModelSY 3299 1–6

Neuropsychologia xxx (2009) xxx–xxx

Contents lists available at ScienceDirect

Neuropsychologia

journa l homepage: www.e lsev ier .com/ locate /neuropsychologia

u wave suppression during the perception of meaningless syllables:EG evidence of motor recruitment

tephen Crawcour, Andrew Bowers, Ashley Harkrider, Tim Saltuklaroglu ∗

epartment of Audiology and Speech Language Pathology, University of Tennessee, 553 S. Stadium Hall, Knoxville, TN 37996-0740, United States

r t i c l e i n f o

rticle history:eceived 11 December 2008eceived in revised form 16 April 2009ccepted 3 May 2009vailable online xxx

eywords:irror neuron

peech perceptionlectroencephalography

a b s t r a c t

Motor involvement in speech perception has been recently studied using a variety of techniques. In thecurrent study, EEG measurements from Cz, C3 and C4 electrodes were used to examine the relative powerof the mu rhythm (i.e., 8–13 Hz) in response to various audio-visual speech and non-speech stimuli, assuppression of these rhythms is considered an index of ‘mirror neuron’ (i.e., motor) activity. Fourteenadult native English speaking females watched and listened to nine audio-video stimuli clips assembledfrom three different auditory stimuli (speech, noise, and pure tone) combined with three different videostimuli (speech, noise, and kaleidoscope—made from scrambling an image from the visual speech). Rel-ative to the noise–noise (baseline condition), all visual speech conditions resulted in significant levels ofsuppression, a finding that is consistent with previous reports of mirror activity to visual speech and mu

D yllables

ensorysuppression to ‘biological’ stimuli. None of the non-speech conditions or conditions in which speech waspresented via audition only resulted in any significant suppression of the mu rhythm in this population.Thus, visual speech perception appears to be more closely associated with motor activity than acousticspeech perception. It is postulated that in this study, the processing demands incurred by the task wereinsufficient for inducing significant mu suppression via acoustic speech only. The findings are discussedin theoretical contexts of speech perception and the mirror system. We suggest that this technique may

-inva

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

OR

RE

CTE

offer a cost-efficient, non

. Introduction

The processes underlying human speech perception have beenidely examined and debated over the last six decades or so. By

ome accounts, speech is perceived as a function of its acoustic con-tituents and their impact on the auditory system (Klatt, 1979; KuhlMiller, 1975; Massaro & Cohen, 1990; Ohala, 1996; Stevens, 1981;

ussman, 1989). Additionally, these acoustic theories generally holdhat speech perception and production are distinct processes. Crit-cisms of these perspectives were propagated by studies showinghe lack of acoustic invariance in similar speech percepts (Liberman,ooper, Shankweiler, & Studdert-Kennedy, 1967), the influence ofisual stimuli on speech percepts (e.g., McGurk effect; McGurk &acDonald, 1976), the phenomenon of categorical perception (e.g.,ann & Liberman, 1983), and the limited temporal resolution of

he auditory system for processing rapidly changing acoustic stim-

UN

C

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

li (Liberman, 1957; Liberman, Delattre, & Cooper, 1952). As such,iberman and Mattingly (1985) proposed an alternative viewpoint.hey suggested that speech is perceptually coded as a sequence ofynamic ‘gestures’, that are representative of the manner in which

∗ Corresponding author.E-mail address: [email protected] (T. Saltuklaroglu).

55

56

57

028-3932/$ – see front matter © 2009 Published by Elsevier Ltd.oi:10.1016/j.neuropsychologia.2009.05.001

sive technique for measuring motor activity during speech perception.© 2009 Published by Elsevier Ltd.

they are produced and co-articulated in the vocal tract. It is thesegestures that are thought to form the invariants for both perceptionand production, linking the two processes and allowing them tooperate efficiently and effectively, together as one, in a specializedlinguistic manner. As such, under this ‘motor theory’, the dynamicarchitecture of the mechanism employed to produce speech playsan essential role in its perception.

Conceptually, motor theory appeared to answer many questionsregarding the nature of speech, yet one of its downfalls was a rel-ative dearth of physiological evidence for the neural connectivitybetween speech perception and production. The discovery of mir-ror neurons in the ventral premotor cortex (area F5) of the macaquemonkey (Di Pellegrino, Fadiga, Fogassi, Gallese, & Rizzolatti, 1992;Rizzolatti, Fadiga, Gallese, & Fogassi, 1996) provides compellingevidence for motor involvement in sensory processes and there-fore, a central linking of perception and production. Neurons in thismotor region, which is considered to be a homolog of Broca’s areain humans (Rizzolatti & Craighero, 2004), were found to fire bothwhen monkeys performed or observed goal directed actions (e.g.,

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

grasping). The location and firing patterns of these ‘mirror neurons’ 58

helped support the notion of motor involvement in speech per- 59

ception (Liberman & Whalen, 2000) and theoretical perspectives 60

of human communication evolving from this observation/action 61

matching system. This neural matching system is thought to 62

E

INN

2 sycho

c63

g64

(65

t66

r67

w68

o69

g70

a71

i72

b73

74

s75

t76

b77

i78

t79

e80

s81

M82

n83

I84

l85

v86

a87

a88

r89

t90

(91

i92

c93

f94

295

(96

i97

s98

T99

o100

m101

m102

103

i104

p105

v106

t107

t108

p109

n110

S111

i112

v113

d114

t115

y116

i117

s118

t119

S120

f121

o122

t123

p124

125

p126

p127

e128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

166

167

168

169

170

171

172

173

174

175

176

177

178

179

180

181

182

183

184

185

186

187

188

189

NC

OR

RE

CT

ARTICLEG ModelSY 3299 1–6

S. Crawcour et al. / Neurop

reate a biological link between senders and receivers of gesturaloals, which may have served in the evolution of communicationRizzolatti & Arbib, 1998). These notions were further bolstered byhe discovery of a subclass of mirror neurons that fired not only inesponse to seeing an action, but also to hearing sounds associatedith a specific action such as paper ripping (Kohler et al., 2002)

r peanut breaking (Keysers et al., 2003). Hence, the authors sug-ested that mirror neurons code the intended goal of an action inn abstract amodal manner rather than a specific action itself, elic-ting strong parallels to the nature of speech ‘gestures’ as describedy motor theorists.

A growing body of research suggests that in humans, the mirrorystem may be involved in action recognition, imitation, empa-hy and theory of mind. Its role in speech perception also haseen scrutinized using various measures. In the auditory modal-

ty, transcranial magnetic stimulation (TMS) has been used to showhat listening to lingual speech sounds could evoke stronger motorvoked potentials (MEPs) in the tongue relative to non-speechounds (Fadiga, Craighero, Buccino, & Rizzolatti, 2002) and strongerEPs in lip muscles when listening to speech while watching white

oise (Watkins & Paus, 2004; Watkins, Strafella, & Paus, 2003).n addition, using functional magnetic resonance imaging (fMRI),istening to meaningless speech has been found to bilaterally acti-ate portions of the ventral premotor cortex (though not Broca’srea), portions of the motor cortex, and the supplementary motorrea (Wilson, Saygin, Sereno, & Iacoboni, 2004) and motor corticalegions in a somatotopic manner (Pulvermüller et al., 2006) rela-ive to non-speech stimuli. In the visual modality, Nishitani and Hari2002) used magnetoecepahlography (MEG) to reveal that observ-ng still pictures of lips could activate Broca’s area and the motorortex. Similar bilateral motor activation patterns have also beenound to silent speech lip movements using fMRI (Campbell et al.,001) and stilled speech (Calvert & Campbell, 2003). Buccino et al.2004) also used fMRI to discover significantly higher activationsn portions of the left inferior frontal gyrus in response to viewingpeech reading and lip smacking, but not to viewing a dog barking.hey suggested that activation of one’s own motor system via actionbservation occurs when the action in question is part of one’s ownotor repertoire, again suggesting a biological underpinning for theirror system.Speech can be perceived unimodally via either audition or vision

n isolation, or bimodally (audio-visually). Though the studies aboverovide evidence that speech perception through either audition orision can activate the human mirror system, they did not examinehe relative strength of each modality for inducing mirror activa-ion or their relative strengths compared to audio-visual speecherception. Because the mirror system is thought to have close con-ections to the somatosensory (SI) system, Möttönen, Järveläinen,ams, and Hari (2004) used MEG to examine how viewing and hear-ng speech modulated activity in the left SI mouth cortex. Whereasiewing speech induced significant SI modulation, hearing speechid not. Similarly Sundara, Namasivayam, and Chen (2001) foundhat visual and audio-visual presentations of the syllable/ba/bothielded significant increases in MEP amplitudes, whereas the MEPncrease produced from auditory perception alone did not reachignificance. Activation levels of motor areas during speech percep-ion have also been examined using fMRI. Skipper, Nusbaum, andmall (2005) found that audio-visual speech activated the inferiorrontal gyrus and premotor cortex to a greater extent than audior visual speech alone. They also found that the activation level ofhe premotor cortex was modulated by a number of phonemes that

U

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

articipants could visually identify.It has been suggested that the visually perceived gestures may

lay the stronger role in activating the motor system during speecherception (Skipper, von Wassenhove, Nusbaum, & Small, 2007),specially in conditions in which auditory speech is absent or

D P

RO

OF

PRESSlogia xxx (2009) xxx–xxx

degraded (Callan et al., 2003). In addition, it is not clear how brainregions involved in speech production are differentially activatedduring various perception tasks. For example, it has been sug-gested that motor recruitment might be influenced by the degreeof linguistic processing necessary to a task (Callan, Jones, Callan, &Akahane-Yamada, 2004; Ojanen et al., 2005; Wilson & Iacoboni,2006). Thus, the numerous measures of motor recruitment inspeech perception and the variety of stimuli employed, combinedwith the diversity of findings, make it difficult to reconcile the dis-crepancies within the current body of research and explain theextent to which motor recruitment may be necessary in speechperception. As such, further investigation in this area is warranted.

Electroencephalography (EEG) has been suggested as a promis-ing, cost-efficient and non-invasive means of indirectly examiningthe mirror neuron activity in humans. In particular, measurementsof oscillation amplitudes in the mu frequencies (8–13 Hz) mea-sured across the sensorimotor cortices acquired via surface levelelectrodes are thought to provide a valid index of mirror activ-ity (Altschuler, Vankov, Wang, Ramachandran, & Pineda, 1997).Mu ‘rhythms’ are influenced by both motor activity and atten-tion (see Pineda, 2005 for full review). When a person is at rest,amplitudes of waves in this band are highest because sensorimotorneurons responsible for generating these waves fire synchronously.Conversely, when a person performs an action, the pattern of fir-ing is asynchronized, resulting in suppression of the mu waveand smaller amplitudes. However, a number of studies also havefound that mu waves are suppressed when normal adults observehuman hand movements (Muthukumaraswamy & Johnson, 2004;Muthukumaraswamy, Johnson, & McNair, 2004; Oberman et al.,2005; Virji-Babul et al., 2008) and implied point-light human bio-logical animation (Saygin, Wilson, Hagler, Bates, & Sereno, 2004;Ulloa & Pineda, 2007), and even when participants imagine biolog-ical motion (Pineda, Allison, & Vankov, 2000).

Because mu suppression can occur in these passive observa-tion/imagination conditions in the absence of motor activity, thelevel of suppression is thought to provide an index of mirror neuronactivity. When employing these paradigms, recordings from Cz, C3and C4 are thought to be indirect measures of cortical activity in thesupplementary motor areas and left and right sensorimotor cortices(S1-M1; Babiloni et al., 1999), respectively. Hence, EEG recordingsfrom these electrodes are considered to be measuring “downstreammodulation of sensorimotor areas by mirror neurons” (Oberman etal., 2005, p. 191). As the recordings are made from the scalp, it isdifficult to map the sources of suppression to cortical landmarks.However, Nishitani and Hari (2000), in a study using MEG, foundthat the sources of mirror activity may be further ‘upstream’ inthe primary motor cortex and in the inferior frontal cortex (e.g.,BA 44). As these regions are often activated in speech perceptiontasks, it seems plausible that EEG recordings of mu rhythms at elec-trode sites Cz, C3 and C4 might also be suppressed when speech isperceived.

Though Muthukumaraswamy, Johnson, Gaetz, and Cheyne(2004) examined mu suppression to oro-facial movements (i.e.,teeth-baring, blowing), to our knowledge, EEG has not yet beenused to examine differential levels of mirror neuron activity tothe perception of speech and non-speech stimuli. Previous stud-ies that have identified selective mirror neuron functioning inresponse to observed dynamic biological stimuli have employedvisual noise and ‘non-biological’ conditions as bases for compari-son. Oberman et al. (2005) used a bouncing ball, whereas Ulloa andPineda (2007) used scrambled versions of their point-light biolog-

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

ical animations. As speech can be conveyed through both auditory 190

and visual channels, in order to differentiate the effects of speech 191

from noise and non-biological stimuli, it seems logical in the cur- 192

rent study to employ all three types of stimulus (noise, speech, 193

and non-biological) conditions for both input modalities. Based 194

D

INN

sycho

u195

p196

p197

a198

t199

s200

e201

o202

t203

s204

2205

2206

207

S208

o209

b210

s211

fi212

i213

n214

i215

t216

U217

2218

219

(220

v221

v222

i223

a224

t225

c226

r227

o228

u229

n230

o231

r232

w233

M234

s235

a236

237

238

239

240

241

242

243

244

245

246

247

248

249

250

251

252

253

254

255

256

257

258

259

260

261

262

263

264

265

266

267

268

269

270

271

272

273

274

275

ARTICLEG ModelSY 3299 1–6

S. Crawcour et al. / Neurop

pon previous findings of motor activation as a function of speecherception, it is hypothesized that conditions in which speech isresented visually will result in the highest levels of mirror neuronctivation and therefore, the most robust mu suppression relativeo when noise is presented through both modalities. When non-peech stimulus conditions are presented, mu suppression is notxpected. These data will also provide some preliminary findingsf how auditory speech may suppress mu waves and contribute tohe body of research investigating how auditory and visual speechignals interact in motor recruitment.

. Methods

.1. Participants

Fourteen right-handed female adults aged 20–51 (mean age = 28.03 years;D = 8.24 years) who were native English speakers and had no diagnosed historyf communicative, cognitive or attentional disorder. A female cohort was chosenecause of the recent finding of stronger mu suppression in response to biologicaltimuli in females than in males (Cheng et al., 2008). As, to our knowledge this is therst study examining mu suppression to speech, it seemed appropriate to first exam-

ne the gender in which the responses might be strongest. In addition, in order toegate any potential ‘processing’ differences that second language acquisition may

ncur, native English speakers were chosen because an English speaker producedhe speech stimuli. Prior to the experiment, informed consent (approved by Theniversity of Tennessee Institutional Review Board) was obtained for all participants.

.2. Stimuli and experimental conditions

The stimuli consisted of audio-visual presentations of (1) meaningless speech,2) noise and (3) a second set of non-biological stimuli. Audio, video and audio-isual speech stimuli were constructed from recordings (using a Sony DCRHC30ideo camera) of the mouth (Fig. 1) of a native English speaking adult male produc-ng a continuous stream of co-articulated meaningless CV syllables (e.g., /da/, /bi/,nd /ga/). A male voice was chosen for the stimuli as males have lower fundamen-al frequencies and their voices are richer in harmonic content. The audio-visuallips were edited using Apple iMovie (version 5.0.2), which allowed for the sepa-ation of the audio and video tracks where necessary. The noise stimuli consistedf auditory and visual white noise. In the same vein that Ulloa and Pineda (2007)sed scrambled point-light stimuli as analogs of their point-light stimuli, the visualon-biological analog of speech movement was constructed by taking a still frame

UN

CO

RR

EC

TE

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

f the mouth recordings and converting it into a kaleidoscope with five symmet-ical portions of the mouth centered at 72◦ to each other using software from theebsite (http://www.krazydad.com/kaleido). Video recordings were made (Screenovie Recorder 2.6) as undulating kaleidoscopic motion was applied in a random

ymmetrical and concentric manner (Fig. 2). The auditory non-biological analog was1000 Hz pure tone. To help ensure uniform intensity levels, all audio stimuli were

Fig. 1. Three examples of three still frames extracted from the dyna

Fig. 2. Three examples of three still frames extracted from the dynami

PR

OO

F

PRESSlogia xxx (2009) xxx–xxx 3

normalized using Adobe Audition (version 1.5) software. Hence, 90 s audio-visualstimuli clips were constructed for nine different audio-visual conditions: (1) NN:noise (auditory)–noise (visual), (2) NS: noise–speech, (3) NK: noise–kaleidoscope,(4) SN: speech–noise, (5) SS: speech–speech, (6) SK: speech–kaleidoscope, (7) TN:tone–noise, (8) TS: tone–speech, and (9) TK: tone–kaleidoscope. The speech stimulicontained 61–69 syllables. Each clip was recorded onto a DVD using Apple iDVD(version 5.0.1).

2.3. Procedure

The experiment was conducted in an electronically and magnetically shielded,double-walled, sound-treated booth. Participants were seated in a comfortablereclining armchair with their heads and necks well supported. After the electrodeswere placed, they were instructed to sit quietly with their eyes open and attend tothe audio and visual stimuli. The stimuli were played from the DVD with a PhillipsHDMI DVD player (model DVP5892). The visual output was displayed on a 32-in.Phillips LCD monitor (model 32PFL5332D). The audio output was routed througha Mackie DFX-6 mixer and was delivered binaurally to participants’ ears at theirpreferred intensity level using Ear Tone model ER-1-14A insert earphones. In orderto help ensure that participants were paying attention, they were asked to silentlycount the number of times they heard or saw a syllable initiated by the /b/ phoneme.All experimental conditions were presented to participants twice in separate ran-dom sequences (i.e., two blocks). Experimental conditions were separated by 30 sbreaks, during which participants were asked to report the number of /b/ initiatedsyllables they were able to recognize from the previous condition. All participantsreported the hearing or seeing /b/ initiated syllables in the speech conditions and noparticipant reported hearing or seeing speech in any of the non-speech conditions.

2.4. EEG data acquisition and analysis

Thirteen electrodes were used to acquire EEG data based on the international10–20 method of electrode placement (Jasper, 1958) using a 32-channel, unlinked,sintered NeuroScan Quik Cap. Non-inverting electrodes included Cz, C3, C4, Pz, P3,P4, Fz, F3, F4, O1, O2, M1, and M2. The inverting electrode was placed on the nasionand the ground electrode was at Fpz. The electro-oculogram (EOG) was recorded byelectrodes placed on the left superior orbit and the left interior orbit. The impedancesof all electrodes were measured at 30 Hz before, during, and after testing and werenever greater than 10 k�.

EEG was collected and analyzed using Compumedics NeuroScan Scan 4.3.3software and Synamps 2 system. Data were obtained for approximately 180 s percondition (90 s for each of two runs), filtered (0.15–100 Hz), and digitized via a 24-bitanalog-to-digital converter at a sampling rate of 500 Hz. It has been reported thatEEG in the 8–13 Hz frequency band recorded from electrodes in the region of the

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

occipital cortex are confounded by states of expectancy and awareness (Klimesch, 276

Doppelmayr, Russegger, Pachinger, & Schwaiger, 1998). In addition, there are shared 277

frequencies between the mu rhythm and posterior alpha bands and the activity in 278

the posterior alpha bands may be stronger than the mu rhythms. For this reason, 279

researchers like Oberman et al. (2005) have suggested that EEG obtained from Cz, 280

C3, and C4 might be confounded by posterior activity. However, this factor is present 281

mic visual speech stimuli. The original stimuli were in color.

c visual kaleidoscopic stimuli. The original stimuli were in color.

INN

4 sycho

a282

w283

f284

r285

w286

t287

i288

w289

a290

e291

292

i293

c294

m295

s296

O297

n298

a299

3300

301

s302

t303

H304

o305

r306

a307

a308

m309

w310

r311

312

p313

e314

e315

m316

a317

c318

Faattn(co

319

320

321

322

323

324

325

326

327

328

329

330

331

332

333

334

335

336

337

338

339

340

341

342

343

344

345

346

ARTICLEG ModelSY 3299 1–6

S. Crawcour et al. / Neurop

cross conditions as they all they involve visual stimuli presentation to participantsith their eyes open and hence does not affect any condition to a greater extent. To

urther combat this, the first and last 10 s of each of the two blocks of 90 s data cor-esponding to one experimental condition was removed. The remaining 70 s of dataere used in the analysis. As data for each condition was collected in two blocks, a

otal 140 s of data were collected per condition. Eye blink and eye movements weredentified in the EOG channel and EEG at all other channels during these intervalsas removed. Then, the integrated power in the 8–13 Hz range was calculated usingFast Fourier Transformation (1024 points) on the data, which were divided into 2 spochs. A cosine window was used to minimize artifacts due to data splicing.

Mu suppression was calculated as the ratio of the power during the eight exper-mental conditions (NS, NT, KS, SS, SN, ST, KN, and KT) relative to the power in the NNondition, which was considered the baseline. Data are presented as ratios to mini-ize individual variability in absolute mu power that may result from differences in

calp thickness or electrode impedance, as opposed to mirror neuron activity (e.g.,berman et al., 2005). A log transformation was applied to each ratio because of theon-normal distribution that ratio data yield. Thus, negative and positive log ratiosre indicative of mu suppression and enhancement, respectively.

. Results

Suppression of mu rhythms relative to baseline were mea-ured from Cz, C3 and C4 as these are the sites coinciding withhe mu wave source and most robust measurements (Pineda &echt, 2009). Because alpha bands were not suppressed in anyther electrode site, we can safely rule out the possibility thatecordings from these sites were influenced by posterior alphactivity. Mu suppression across three electrode sites (Cz, C3, and C4)nd eight conditions were analyzed using a two-factor repeated-easures ANOVA. Greenhouse-Geisser corrections were appliedhere sphericity assumptions were violated and Bonferroni cor-

ections were applied to multiple post hoc comparisons.A significant main effect was found for condition [F(7,91) = 6.367,

= 0.001]. However, no significant main effect was found for the

UN

CO

RR

EC

TE

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

lectrode site [F(2,26) = 0.105, p = 0.901] or the interaction betweenlectrode site and condition, [F(14,182) = 0.992, p = 0.463]. Hence,u suppression data were collapsed across the three electrodes and

re displayed as function of experimental condition in Fig. 3. Theollapsed data were used in post hoc t-tests that examined how

ig. 3. Mu suppression across experimental condition consisting of combinations ofuditory and visual stimuli. Labels on the X-axis represent the following conditions:udio speech and visual speech (SS), audio noise and visual speech (NS), audio pureone and visual speech (TS), audio speech and visual white noise (SN), audio pureone and visual white noise (TN), audio speech and visual kaleidoscope (SK), audiooise and visual kaleidoscope (NK), and audio pure tone and visual kaleidoscopeTK). Bars represent mean log ratio of mu power relative to baseline (NN condition)ollapsed across Cz, C3 and C4 electrodes. Error bars represent one standard errorf the mean.

347

348

349

350

351

352

353

354

355

356

357

358

359

360

361

362

363

364

365

366

367

368

369

370

371

372

373

374

375

376

D P

RO

OF

PRESSlogia xxx (2009) xxx–xxx

each condition suppressed the mu wave relative to the baseline(Oberman et al., 2005; Oberman, Ramachandran, & Pineda, 2008).The conditions in which visual speech were presented were theonly ones to induce significant and robust levels of mu suppres-sion [SS: t(13) = −4.223, p = .008; NS: t(13) = −3.904, p = .016; TS:t(13) = −4.760, p < .001].

4. Discussion

In accord with our first hypothesis, significant suppression ofthe mu wave was found across electrodes Cz, C3 and C4 when par-ticipants visually perceived streams of meaningless co-articulatedspeech gestures (i.e., viewed continuous lip movements). All threevisual speech conditions tested (SS, NS, and TS) produced simi-lar levels of suppression, suggesting that visual speech perceptionrecruited the motor system irrespective of the auditory stim-ulus with which it was paired. These findings are consistentboth with studies showing similar patterns of mu suppression toobserved human movements (Muthukumaraswamy, Johnson, &McNair, 2004; Oberman et al., 2005; Virji-Babul et al., 2008) andwith studies using MEG or fMRI showing bilateral motor activity tovisual speech perception (Calvert & Campbell, 2003; Campbell et al.,2001; Nishitani & Hari, 2002; Skipper et al., 2005). Furthermore, inaccord with our second hypothesis, the non-speech conditions (TN,TK, and NK) failed to produce any significant levels of mu suppres-sion, most likely, because in these conditions, participants did notperceive stimuli containing arrays of movements existing withintheir own motor repertoire (Buccino et al., 2001).

These findings also appear to be consistent with those ofMöttönen et al. (2004) and Sundara et al. (2001) who foundevidence of motor activation in response to visual but not audi-tory speech perception. As such, they support notions that motorrecruitment for speech is often mediated primarily by visual sys-tems (Skipper et al., 2007) and that when visual speech informationis presented, perception of the gesture is supported by an internalsimulation of the observed articulatory movements (Callan et al.,2003). This motoric supplementation which appears to be affordedby the presence of visual speech may contribute to previous find-ings showing that the addition of congruent visual cues to auditoryspeech is associated with increases in speech intelligibility (Sumby& Pollack, 1954), improved speech detection thresholds (Grant &Seitz, 2000), and accelerated neural processing (Van Wassenhove,Grant, & Poeppel, 2005).

Current findings may contrast with those that have found motorinvolvement during auditory-only speech perception tasks (e.g.,Fadiga et al., 2002; Watkins et al., 2003; Watkins & Paus, 2004;Wilson et al., 2004). Skipper et al. (2005) suggest that any briefmotor activation that occurs during auditory speech perceptionmay not be observed in methods that employ temporal averag-ing. Perhaps however, a closer examination of acoustic speechsignals may help foster a better understanding of their inconsis-tent motor activation patterns. Acoustic speech signals are richlyencoded and contain cues indicative of the gesture (e.g., formantstructures, glottal pulsing, onset spectra) from every level of thevocal tract. In fact, the redundancy of information pointing tothe intended gesture allows that the percept be recovered evenwhen considerable portions of the acoustic signal are removed. Anextreme case of this can be found when humans perceive linguis-tic information from sine wave analogs of speech that are devoidof traditional acoustic cues (Remez, Rubin, Pisoni, & Carrell, 1981).

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

One perspective derived from cognitive psychology that might con- 377

tribute to explaining this substantial human capacity is the view 378

of speech perception as an ‘embodied’ process (Aziz-Zadeh & Ivry, 379

2009; Galantucci, Fowler, & Turvey, 2006; Skipper et al., 2005; 380

Wilson, 2002). From this standpoint, the manner in which speech 381

D

INN

sycho

i382

C383

t384

p385

s386

p387

S388

s389

i390

s391

p392

393

v394

B395

c396

s397

&398

s399

a400

i401

n402

v403

i404

l405

O406

r407

t408

a409

r410

a411

i412

a413

v414

s415

f416

i417

t418

t419

420

n421

t422

i423

C424

m425

i426

m427

a428

b429

t430

i431

p432

t433

g434

f435

d436

f437

a438

w439

N440

r441

m442

s443

a444

b445

446

p447

448

449

450

451

452

453

454

455

456

457

458

459

460

461

462

463

464

465

466

467

468

469

470

471

472

473

Q1 474

475

476

477

478

479

480

481

482

483

484

485

486

487

488

489

490

491

492

493

494

495

496

497

498

499

500

501

502

503

504

505

506

507

508

509

510

511

512

513

NC

OR

RE

CTE

ARTICLEG ModelSY 3299 1–6

S. Crawcour et al. / Neurop

s perceived is experientially shaped through multiple interactions.onnections between the acoustics of speech and the motoric ges-ural configurations responsible for producing those acoustics (i.e.,arity) become strengthened over time. It seems logical that theseensorimotor connections are strengthened most during the earlyeriods of language acquisition and development (Kalinowski &altuklaroglu, 2003). Therefore, it may be plausible that once aufficient phonemic repertoire is acquired, motor involvement dur-ng basic speech perception becomes less essential; yet the motorystem may remain an available referent during more demandinghonemic or linguistic processing tasks.

The relationship between processing demands and motor acti-ation may be further considered when examining the role ofroca’s area in speech perception. Though Broca’s area is classicallyonsidered a key speech production area, it is also considered byome to mediate functioning of the mirror system (Hari, Levanen,Raij, 2000; Nishitani & Hari, 2002) and play a strong role in multi-

ensory processing. During passive speech perception tasks, Broca’srea has been found to be more active in conditions that requirencreased amounts of processing such as when an auditory sig-al is degraded or absent (Callan et al., 2003), when auditory andisual signals are not matched (Ojanen et al., 2005), when listen-ng to non-native phonemes (Wilson & Iacoboni, 2006) and whenistening to sentences paired with incongruent gestures (Willems,zyurek, & Hagoort, 2007). Hence if the suppression of the mu

hythm is influenced by activity in Broca’s area, it may be postulatedhat in the current study, the processing demands incurred by theuditory-only speech conditions were insufficient for necessitatingecruitment of the motor system and therefore, did not result inny significant suppression of the mu rhythm. That is, the mean-ngless acoustic speech signal was produced by an English speakernd was presented clearly and audibly. Its combination with eitherisual noise or the kaleidoscopic stimuli most likely did not tax thepeech processing demands. In addition, the participants were allemale native English speakers, free from speech, language or hear-ng disorders, and recruited from a university setting, suggestinghat they had little difficulty interpreting the phonemic informationhat was presented under these conditions.

Skipper et al. (2005) suggest that rather than being static inature, the mirror system may be a dynamic entity with variablehresholds for engagement across stimuli, both within and betweenndividuals and dependent on the nature and goals of a given task.onsidering this dynamic nature and the possibility that speechechanisms within individuals may be tied to their embodiment

n world experiences, it is not surprising that the original tenets ofotor theory have been recently re-examined (e.g., Galantucci et

l., 2006; Lotto, Hickok, & Holt, 2009). The current paradigm maye appropriate for shedding further light on questions surroundinghe role of motor involvement in speech perception. Future stud-es examining mu suppression to speech may consider increasingrocessing demands, perhaps by adding different types of noise tohe speech signal (e.g., Callan et al., 2003), using non-native lin-uistic forms (e.g., Wilson & Iacoboni, 2006) or employing speechrom disordered populations such as those who stutter or haveysarthric speech. Another application may be found in the dif-erent populations that may be tested, Much remains to be learnedbout the speech processing in disordered populations such thoseith autism, specific language disorders and aphasia. In addition,yström (2008) recently used high density EEG to find similar mir-

or responses in infants and adults in response to goal directedovements and as such, EEG may be appropriate for examining

U

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

peech perception mechanisms in children and comparing them todults, especially if theories about embodied processing prove toe valid.

To our knowledge, this is the first study to examine mu sup-ression during auditory and visual speech perception. Though we

PR

OO

F

PRESSlogia xxx (2009) xxx–xxx 5

believe the findings to be of importance, certain limitations shouldbe recognized in interpreting the findings. First, though the murhythm is thought to emanate from sensorimotor areas, the lim-ited spatial resolution of surface EEG recordings does not allowaccurate site mapping of the neural activity. As such, future inves-tigations continue to require that motor activity measured by musuppression is indexed as a function of the task and relative to otherconditions under investigation. Second, this study used an exclu-sively female cohort. Though females were chosen because of theirhigher sensitivity in recording mu suppression, these findings needto be replicated in males. Third, though the visual non-biologic ana-log was dynamic (i.e., the kaleidoscope), the non-biological analogto auditory speech (i.e., the pure tone) was not. In future stud-ies, to better understand the differential abilities of speech versusnon-speech acoustic signals to induce mirror activity, using moredynamic non-biological acoustic stimuli as controls may be advan-tageous.

These preliminary findings showed mu suppression, consideredevidence of cortical motor activation, to visually perceived streamsof meaningless syllables, irrespective of acoustic pairing. Thoughsuppression to acoustic only speech stimuli was not found herein,we are cautiously optimistic that it might be found in future stud-ies using this technique. As such, indexing mu suppression acrossaudio-visual tasks appears to be a non-invasive, cost-efficient tech-nique that may continue to provide clues to the many unansweredquestions about the nature of human speech perception.

Uncited reference

Lahav, Saltzman, and Schlaug (2007).

References

Altschuler, E. L., Vankov, A., Wang, V., Ramachandran, V. S., & Pineda, J. A. (1997).Person see, person do cells. Abstracts-Social Neuroscience, 23, 1848.

Aziz-Zadeh, L., & Ivry, R. B. (2009). The human mirror neuron system and embodiedrepresentations. Advances in Medicine and Biology, 629, 355–376.

Babiloni, C., Carducci, F., Cincotti, F., Rossini, P. M., Neuper, C., Pfurtscheller, et al.(1999). Human movement related potentials vs desynchronization of EEG alpharhythm: A high resolution EEG study. NeuroImage, 10, 658–665.

Buccino, G., Binkofski, F., Fink, G. R., Fadiga, L., Fogassi, L., Gallese, et al. (2001). Actionobservation activates premotor and parietal areas in a somatotopic manner: AnfMRI study. The European Journal of Neuroscience, 13, 400–404.

Buccino, G., Lui, F., Canessa, N., Patteri, I., Lagravinese, G., Benuzzi, F., et al. (2004).Neural circuits involved in the recognition of actions performed by noncon-specifics: An fMRI study. Journal of Cognitive Neuroscience, 16, 114–126.

Callan, D. E., Jones, J. A., Callan, A. M., & Akahane-Yamada, R. (2004). Phonetic per-ceptual identification by native- and second-language speakers differentiallyactivates brain regions involved with acoustic phonetic processing and thoseinvolved with articulatory-auditory/orosensory internal models. NeuroImage,22, 1182–1194.

Callan, D. E., Jones, J. A., Munhall, K. G., Callan, A. M., Kroos, C., & Vatikiotis-Bateson, E.(2003). Neural processes underlying perceptual enhancement by visual speechgestures. NeuroReport, 14, 2213–2218.

Campbell, R., MacSweeney, M., Surguladze, S., Calvert, G. A., McGuire, P. K., Brammer,M. J., et al. (2001). Cortical substrates for the perception of face actions: An fMRIstudy of the specificity of activation for seen speech and for meaningless lower-face acts (gurning). Cognitive Brain Research, 12, 233–243.

Calvert, G. A., & Campbell, R. (2003). Reading speech from still and moving faces: Theneural substrates of visible speech. Journal of Cognitive Neuroscience, 15, 57–70.

Cheng, Y., Lee, P.-L., Yang, Ch-Y., Lin, Ch-P., Hung, D., & Decety, J. (2008). Genderdifferences in the mu-rhythm of the human mirror-neuron system. PLoS ONE, 3,1–7.

Di Pellegrino, G., Fadiga, L., Fogassi, L., Gallese, V., & Rizzolatti, G. (1992). Understand-ing motor events: A neurophysiological study. Experimental Brain Research, 91,176–180.

Fadiga, L., Craighero, L., Buccino, G., & Rizzolatti, G. (2002). Speech listening specifi-cally modulates the excitability of tongue muscles: A TMS study. European Journalof Neuroscience, 15, 399–402.

Galantucci, B., Fowler, C. A., & Turvey, M. T. (2006). The motor theory of speech

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

perception reviewed. Psychonomic Bulletin & Review, 13, 361–377. 514

Grant, K. W., & Seitz, P. F. (2000). The use of visible speech cues for improving auditory 515

detection of spoken sentences. Journal of the Acoustical Society of America, 108, 516

1197–1208. 517

Hari, R., Levanen, S., & Raij, T. (2000). Timing of human cortical functions during 518

cognition: Role of MEG. Trends in Cognitive Sciences, 4, 455–462. 519

E

INN

6 sycho

J520

521

522

K523

524

525

K526

527

528

K529

530

K531

532

533

K534

535

K536

537

L538

539

540

L541

542

L543

544

L545

546

547

L548

L549

550

L551

552

M553

554

M555

556

M557

558

M559

560

M561

562

563

M564

565

M566

567

568

N569

570

N571

572

N573

574

O575

576

577

O578

579

580

581

O582

583

584

585

586

587

588

589

590

591

592

593

594

595

596

597

598

599

600

601

602

603

604

605

606

607

608

609

610

611

612

613

614

615

616

617

Q2 618

619

620

621

622

623

624

625

626

627

628

629

630

631

632

633

634

635

636

637

638

639

640

EC

T

ARTICLEG ModelSY 3299 1–6

S. Crawcour et al. / Neurop

asper, H. H. (1958). Report on the committee on methods of clinical examinationin electroencephalography. The ten-twenty system. Electroencephalography andClinical Neurophysiology, 10(Suppl.), 371–375.

alinowski, J., & Saltuklaroglu, T. (2003). Choral speech: The amelioration of stutter-ing via imitation and the mirror neuronal system. Neuroscience and BiobehavioralReviews, 27, 339–347.

eysers, C., Kohler, E., Umilta, M. A., Nanetti, L., Fogassi, L., & Gallese, V. (2003).Audiovisual mirror neurons and action recognition. Experimental Brain Research,153, 628–636.

latt, D. H. (1979). Speech perception: A model of acoustic-phonetic analysis andlexical access. Journal of Phonetics, 7, 279–285.

limesch, W., Doppelmayr, M., Russegger, H., Pachinger, T., & Schwaiger, J. (1998).Induced alpha band power changes in the human EEG and attention. Neuro-science Letters, 244, 73–76.

ohler, E., Keysers, Ch., Umilta, M. A., Fogassi, L., Gallese, V., & Rizzolatti, G. (2002).Hearing sounds, understanding action. Science, 27, 846–848.

uhl, P. K., & Miller, J. D. (1975). Speech perception by the chinchilla: Voiced-voicelessdistinction in alveolar plosive consonants. Science, 190, 69–72.

ahav, A., Saltzman, E., & Schlaug, G. (2007). Action representation of sound:Audiomotor recognition network while listening to newly acquired actions. TheJournal of Neuroscience, 27, 308–314.

iberman, A. M. (1957). Some results of research on speech perception. Journal of theAcoustical Society of America, 29, 117–123.

iberman, A., Cooper, F., Shankweiler, D., & Studdert-Kennedy, M. (1967). Perceptionof speech code. Psychological Reviews, 74, 431–461.

iberman, A. M., Delattre, P., & Cooper, F. S. (1952). The role of selected stimulus-variables in the perception of the unvoiced stop consonants. American Journal ofPsychology, 65, 497–516.

iberman, A., & Mattingly, I. (1985). The motor theory revised. Cognition, 21, 1–36.iberman, A. M., & Whalen, D. H. (2000). On the relation of speech to language. Trends

in Cognitive Sciences, 4, 187–196.otto, A. J., Hickok, G. S., & Holt, L. L. (2009). Reflections on mirror neurons and speech

perception. Trends in Cognitive Sciences, 13, 110–114.ann, V. A., & Liberman, A. M. (1983). Some differences between phonetic and

auditory modes of perception. Cognition, 14, 211–235.assaro, D. W., & Cohen, M. M. (1990). Perception of synthesized audible and visible

speech. Psychological Science, 1, 55–63.cGurk, H., & MacDonald, J. (1976). Hearing lips and seeing voices. Nature, 264,

746–748.öttönen, R., Järveläinen, J., Sams, M., & Hari, R. (2004). Viewing speech modulates

activity in the left SI mouth cortex. NeuroImage, 24, 731–737.uthukumaraswamy, S. D., Johnson, B. W., Gaetz, W. C., & Cheyne, D. O. (2004).

Modulation of neuromagnetic oscillatory activity during the observation of oro-facial movements. Neurology and Clinical Neurophysiology, 2, 1–4.

uthukumaraswamy, S. D., & Johnson, B. W. (2004). Changes in rolandic mu rhythmduring observation of a precision grip. Psychophysiology, 41, 152–156.

uthukumaraswamy, S. D., Johnson, B. W., & McNair, N. A. (2004). Mu rhythm mod-ulation during observation of an object-directed grasp. Cognitive Brain Research,19, 195–201.

ishitani, N., & Hari, R. (2000). Temporal dynamics of cortical representation foraction. PNAS, 97, 913–918.

ishitani, N., & Hari, R. (2002). Viewing lip forms: Cortical dynamics. Neuron, 36,1211–1220.

yström, P. (2008). The infant mirror neuron system studied with high density EEG.Social Neuroscience, 3, 334–347.

berman, L. M., Hubbard, E. M., McCleery, J. P., Altschuler, E. L., Ramachandran, V.S., & Pineda, J. A. (2005). EEG evidence for mirror neuron dysfunction in autismspectrum disorders. Cognitive Brain Research, 24, 190–198.

UN

CO

RR

Please cite this article in press as: Crawcour, S., et al. Mu wave suppressof motor recruitment. Neuropsychologia (2009), doi:10.1016/j.neurops

berman, L. M., Ramachandran, V. S., & Pineda, J. A. (2008). Modulation of musuppression in children with autism spectrum disorders in response to famil-iar or unfamiliar stimuli: The mirror neuron hypothesis. Neuropsychologia, 46,1558–1565.

hala, J. (1996). Speech perception is hearing sounds, not tongues. Journal of theAcoustical Society of America, 99, 1718–1725.

D P

RO

OF

PRESSlogia xxx (2009) xxx–xxx

Ojanen, V., Möttönen, R., Pekkola, J., Jääskeläinen, I. P., Joensuu, R., Autti, T., etal. (2005). Processing of audiovisual speech in Broca’s area. NeuroImage, 25,333–338.

Pineda, J. A. (2005). The functional significance of mu rhythms: Translating “seeing”and “hearing” into “doing”. Brain Research Reviews, 50, 57–68.

Pineda, J. A., Allison, B. Z., & Vankov, A. (2000). The effects of self-movement, obser-vation, and imagination on mu-rhythms and readiness potentials: Toward abrain-computer interface. IEEE Transactions on Neural Systems and RehabilitationEngineering, 8, 219–222.

Pineda, J. A., & Hecht, E. (2009). Mirroring and mu rhythm involvement in socialcognition: Are there dissociable subcomponents of theory of mind? BiologicalPsychology, 80, 306–314.

Pulvermüller, F., Huss, M., Kherif, F., del Prado Martin, F. M., Hauk, O., & Shtyrov, Y.(2006). Motor cortex maps articulatory features of speech sounds. PNAS, 103,7865–7870.

Remez, R. E., Rubin, P. E., Pisoni, D. B., & Carrell, T. D. (1981). Speech perceptionwithout traditional speech cues. Science, 212, 947–950.

Rizzolatti, G., & Arbib, M. A. (1998). Language within our grasp. Trends in Neuroscience,21, 188–194.

Rizzolatti, G., & Craighero, L. (2004). The mirror neuron system. Annual Review ofNeuroscience, 27, 169–192.

Rizzolatti, G., Fadiga, L., Gallese, V., & Fogassi, L. (1996). Premotor cortex and therecognition of motor action. Cognitive Brain Research, 3, 131–141.

Saygin, A. P., Wilson, S. M., Hagler, D. J., Jr., Bates, E., & Sereno, M. (2004). Point-lightbiological motion perception activates human premotor cortex. The Journal ofNeuroscience, 24, 6181–6188.

Skipper, J. I., Nusbaum, H. C., & Small, S. L. (2005). Listening to talking faces: Motorcortical activation during speech perception. NeuroImage, 25, 76–89.

Skipper, J. I., von Wassenhove, V., Nusbaum, H. C., & Small, S. L. (2007). Hearing lipsand seeing voices: How cortical areas supporting speech production mediateaudiovisual speech perception. Cerebral Cortex, 17, 2387–2399.

Stevens, K. N. (1981). Constraints imposed by the auditory system on the propertiesused to classify speech sounds: Evidence from phonology, acoustics, and psy-choacoustics. In T. Myers, J. Laver, & J. Anderson (Eds.), The cognitive representationof speech (pp. 61–74). Amsterdam, New York: Elsevier North-Holland.

Sumby, W. H., & Pollack, L. (1954). Visual contribution of speech intelligibility innoise. Journal of the Acoustical Society of America, 26, 125–212.

Sundara, M., Namasivayam, A. K., & Chen, R. (2001). Observation-execution matchingsystem for speech: A magnetic stimulation study. Neuroreport, 12, 1341–1344.

Sussman, H. (1989). Neural coding of relational invariance in speech: Human lan-guage analogs to the barn owl. Psychological Review, 96, 631–642.

Ulloa, E. R., & Pineda, J. A. (2007). Recognition of point-light biological motion: Murhythms and mirror neuron activity. Behavioral Brain Research, 183, 188–194.

Van Wassenhove, V., Grant, K. W., & Poeppel, D. (2005). Visual speech speeds up theneural processing of auditory speech. PNAS, 102, 1181–1186.

Virji-Babul, N., Moiseeva, A., Cheunga, T., Weeks, D., Cheynec, D., & Ribary, U. (2008).Changes in mu rhythm during action observation and execution in adults withDown syndrome: Implications for action representation. Neuroscience Letters,436, 177–180.

Watkins, K. E., & Paus, T. (2004). Modulation of motor excitability duringspeech perception: The of Broca’s area. Journal of Cognitive Neuroscience, 16,978–987.

Watkins, K. E., Strafella, A. P., & Paus, T. (2003). Seeing and hearing speech excites themotor system involved in speech production. Neuropsychologia, 41, 989–994.

Willems, R. M., Ozyurek, A., & Hagoort, P. (2007). When language meets action: Theneural integration of gesture and speech. Cerebral Cortex, 17, 2322–2333.

Wilson, M. (2002). Six views of embodied cognition. Psychonomic Bulletin & Review,9, 625–636.

ion during the perception of meaningless syllables: EEG evidenceychologia.2009.05.001

Wilson, S. M., & Iacoboni, M. (2006). Neural responses to non-native phonemes vary- 641

ing in producibility: Evidence for the sensorimotor nature of speech perception. 642

NeuroImage, 33, 316–325. 643

Wilson, S. M., Saygin, A. P., Sereno, M. I., & Iacoboni, M. (2004). Listening to speech 644

activates motor areas involved in speech production. Nature Neuroscience, 7, 645

701–702. 646