Visual Feedback of Tongue Movement for Novel Speech Sound Learning

William F Katz, Sonya Mehta, William F Katz, Sonya Mehta

Abstract

Pronunciation training studies have yielded important information concerning the processing of audiovisual (AV) information. Second language (L2) learners show increased reliance on bottom-up, multimodal input for speech perception (compared to monolingual individuals). However, little is known about the role of viewing one's own speech articulation processes during speech training. The current study investigated whether real-time, visual feedback for tongue movement can improve a speaker's learning of non-native speech sounds. An interactive 3D tongue visualization system based on electromagnetic articulography (EMA) was used in a speech training experiment. Native speakers of American English produced a novel speech sound (/ɖ/; a voiced, coronal, palatal stop) before, during, and after trials in which they viewed their own speech movements using the 3D model. Talkers' productions were evaluated using kinematic (tongue-tip spatial positioning) and acoustic (burst spectra) measures. The results indicated a rapid gain in accuracy associated with visual feedback training. The findings are discussed with respect to neural models for multimodal speech processing.

Keywords: articulation therapy; audiovisual integration; electromagnetic articulography; second language learning; speech production; visual feedback.

Figures

Figure 1
Figure 1
Illustration of the Opti-Speech system, with subject wearing sensors and head-orientation glasses (lower right insert). A sample target sphere, placed in this example at the subject's alveolar ridge, is shown in red. A blue marker indicates the tongue tip/blade (TT) sensor.
Figure 2
Figure 2
Close-up of tongue avatar during a “hit” for the production of the voiced, retroflex, palatal stop consonant. The target sphere lights up green, providing visual feedback for the correct place of articulation.
Figure 3
Figure 3
Accuracy for five talkers producing a coronal palatal stop. Shaded regions indicate visual feedback conditions. Baseline (pre-training) and post-training phases are also indicated.
Figure 4
Figure 4
Overlapping plots of short-term spectra for bursts of voiced, coronal, palatal stops produced before and after EMA training. Correct place of articulation (hits) are marked in blue, and errors (misses) in red. Computed averages of incorrect pre-training (red) and correct post-training (blue) spectra are shown at right, for comparison.
Figure 5
Figure 5
Simplified version of ACT model (Kröger and Kannampuzha, 2008), showing input pathways for external audiovisual stimuli (oval at bottom right) and optional feedback circuits to the vocal tract (shaded box at bottom). Visual feedback (dotted line) is provided by either external (mirroring) or internal (instrumental augmented) routes.

References

    1. Arbib M. A. (2005). From monkey-like action recognition to human language: an evolutionary framework for neurolinguistics. Behav. Brain Sci. 28, 105–124. 10.1017/S0140525X05000038
    1. Arnold P., Hill F. (2001). Bisensory augmentation: a speechreading advantage when speech is clearly audible and intact. Br. J. Psychol. 92(Pt 2), 339–355. 10.1348/000712601162220
    1. Badin P., Elisei F., Bailly G., Tarabalka Y. (2008). An audiovisual talking head for augmented speech generation: models and animations based on a real speaker's articulatory data, in Vth Conference on Articulated Motion and Deformable Objects (AMDO 2008, LNCS 5098), eds Perales F. J., Fisher R. B. (Berlin; Heidelberg: Springer Verlag; ), 132–143. 10.1007/978-3-540-70517-8_14
    1. Badin P., Tarabalka Y., Elisei F., Bailly G. (2010). Can you ‘read’ tongue movements? Evaluation of the contribution of tongue display to speech understanding. Speech Commun. 52, 493–503. 10.1016/j.specom.2010.03.002
    1. Ballard K. J., Smith H. D., Paramatmuni D., McCabe P., Theodoros D. G., Murdoch B. E. (2012). Amount of kinematic feedback affects learning of speech motor skills. Motor Contr. 16, 106–119.
    1. Berlucchi G., Aglioti S. (1997). The body in the brain: neural bases of corporeal awareness. Trends Neurosci. 20, 560–564. 10.1016/S0166-2236(97)01136-3
    1. Bernhardt B., Gick B., Bacsfalvi P., Adler-Bock M. (2005). Ultrasound in speech therapy with adolescents and adults. Clin. Linguist. Phon. 19, 605–617. 10.1080/02699200500114028
    1. Bernstein L. E., Liebenthal E. (2014). Neural pathways for visual speech perception. Front. Neurosci. 8:386. 10.3389/fnins.2014.00386
    1. Berry J. J. (2011). Accuracy of the NDI wave speech research system. J. Speech Lang. Hear. Res. 54, 1295–1301. 10.1044/1092-4388(2011/10-0226)
    1. Bislick L. P., Weir P. C., Spencer K., Kendall D., Yorkston K. M. (2012). Do principles of motor learning enhance retention and transfer of speech skills? A systematic review. Aphasiology 26, 709–728. 10.1080/02687038.2012.676888
    1. Boersma P., Weenink D. (2001). Praat, a system for doing phonetics by computer. Glot International 5, 341–345.
    1. Civier O., Tasko S. M., Guenther F. H. (2010). Overreliance on auditory feedback may lead to sound/syllable repetitions: simulations of stuttering and fluency-inducing conditions with a neural model of speech production. J. Fluency Disord. 35, 246–279. 10.1016/j.jfludis.2010.05.002
    1. Curio G., Neuloh G., Numminen J., Jousmäki V., Hari R. (2000). Speaking modifies voice−evoked activity in the human auditory cortex. Hum. Brain Mapp. 9, 183–191. 10.1002/(SICI)1097-0193(200004)9:4<183::AID-HBM1>;2-Z
    1. D'Ausilio A., Bartoli E., Maffongelli L., Berry J. J., Fadiga L. (2014). Vision of tongue movements bias auditory speech perception. Neuropsychologia 63, 85–91. 10.1016/j.neuropsychologia.2014.08.018
    1. Dagenais P. A. (1995). Electropalatography in the treatment of articulation/phonological disorders. J. Commun. Disord. 28, 303–329. 10.1016/0021-9924(95)00059-1
    1. Daprati E., Sirigu A., Nico D. (2010). Body and movement: consciousness in the parietal lobes. Neuropsychologia 48, 756–762. 10.1016/j.neuropsychologia.2009.10.008
    1. Dart S. N. (1991). Articulatory and Acoustic Properties of Apical and Laminal Articulations, Vol. 79 Los Angeles, CA: UCLA Phonetics Laboratory.
    1. Dayan E., Hamann J. M., Averbeck B. B., Cohen L. G. (2014). Brain structural substrates of reward dependence during behavioral performance. J. Neurosci. 34, 16433–16441. 10.1523/JNEUROSCI.3141-14.2014
    1. Engelen L., Prinz J. F., Bosman F. (2002). The influence of density and material on oral perception of ball size with and without palatal coverage. Arch. Oral Biol. 47, 197–201. 10.1016/S0003-9969(01)00106-6
    1. Engwall O. (2008). Can audio-visual instructions help learners improve their articulation?-an ultrasound study of short term changes, in Interspeech (Brisbane: ), 2631–2634.
    1. Engwall O., Bälter O. (2007). Pronunciation feedback from real and virtual language teachers. Comput. Assist. Lang. Learn. 20, 235–262. 10.1080/09588220701489507
    1. Engwall O., Bälter O., Öster A.-M., Kjellström H. (2006). Feedback management in the pronunciation training system ARTUR, in CHI'06 Extended Abstracts on Human Factors in Computing Systems (Montreal: ACM; ), 231–234.
    1. Engwall O., Wik P. (2009). Can you tell if tongue movements are real or synthesized? in Proceedings of Auditory-Visual Speech Processing (Norwich: University of East Anglia; ), 96–101.
    1. Erber N. P. (1975). Auditory-visual perception of speech. J. Speech Hear. Disord. 40, 481–492. 10.1044/jshd.4004.481
    1. Fagel S., Madany K. (2008). A 3-D virtual head as a tool for speech therapy for children, in Proceedings of Interspeech 2008 (Brisbane, QLD; ), 2643–2646.
    1. Fant G. (1973). Speech Sounds and Features. Cambridge, MA: The MIT Press.
    1. Farrer C., Franck N., Georgieff N., Frith C. D., Decety J., Jeannerod M. (2003). Modulating the experience of agency: a positron emission tomography study. Neuroimage 18, 324–333. 10.1016/S1053-8119(02)00041-1
    1. Felps D., Bortfeld H., Gutierrez-Osuna R. (2009). Foreign accent conversion in computer assisted pronunciation training. Speech Commun. 51, 920–932. 10.1016/j.specom.2008.11.004
    1. Fridriksson J., Hubbard H. I., Hudspeth S. G., Holland A. L., Bonilha L., Fromm D., et al. . (2012). Speech entrainment enables patients with Broca's aphasia to produce fluent speech. Brain 135, 3815–3829. 10.1093/brain/aws301
    1. Gentilucci M., Corballis M. C. (2006). From manual gesture to speech: a gradual transition. Neurosci. Biobehav. Rev. 30, 949–960. 10.1016/j.neubiorev.2006.02.004
    1. Goozee J. V., Murdoch B. E., Theodoros D. G. (1999). Electropalatographic assessment of articulatory timing characteristics in dysarthria following traumatic brain injury. J. Med. Speech Lang. Pathol. 7, 209–222.
    1. Guenther F. H. (2006). Cortical interactions underlying the production of speech sounds. J. Commun. Disord. 39, 350–365. 10.1016/j.jcomdis.2006.06.013
    1. Guenther F. H., Ghosh S. S., Tourville J. A. (2006). Neural modeling and imaging of the cortical interactions underlying syllable production. Brain Lang. 96, 280–301. 10.1016/j.bandl.2005.06.001
    1. Guenther F. H., Perkell J. S. (2004). A neural model of speech production and its application to studies of the role of auditory feedback in speech, in Speech Motor Control in Normal and Disordered Speech, eds Maassen B., Kent R., Peters H. F. M., Van Lieshout P., Hulstijn W. (Oxford University Press; ), 29–50.
    1. Guenther F. H., Vladusich T. (2012). A neural theory of speech acquisition and production. J. Neurolinguistics 25, 408–422. 10.1016/j.jneuroling.2009.08.006
    1. Gunji A., Hoshiyama M., Kakigi R. (2001). Auditory response following vocalization: a magnetoencephalographic study. Clin. Neurophysiol. 112, 514–520. 10.1016/S1388-2457(01)00462-X
    1. Haggard P., de Boer L. (2014). Oral somatosensory awareness. Neurosci. Biobehav. Rev. 47, 469–484. 10.1016/j.neubiorev.2014.09.015
    1. Hamann S. (2003). The Phonetics and Phonology of Retroflexes. Ph.D. dissertation, Netherlands Graduate School of Linguistics, University of Utrecht, LOT, Utrecht.
    1. Hardcastle W. J., Gibbon F. E., Jones W. (1991). Visual display of tongue-palate contact: electropalatography in the assessment and remediation of speech disorders. Int. J. Lang. Commun. Disord. 26, 41–74. 10.3109/13682829109011992
    1. Hartelius L., Theodoros D., Murdoch B. (2005). Use of electropalatography in the treatment of disordered articulation following traumatic brain injury: a case study. J. Med. Speech Lang. Pathol. 13, 189–204.
    1. Heinks-Maldonado T. H., Nagarajan S. S., Houde J. F. (2006). Magnetoencephalographic evidence for a precise forward model in speech production. Neuroreport 17, 1375. 10.1097/01.wnr.0000233102.43526.e9
    1. Hickok G. (2009). Eight problems for the mirror neuron theory of action understanding in monkeys and humans. J. Cogn. Neurosci. 21, 1229–1243. 10.1162/jocn.2009.21189
    1. Hickok G. (2010). The role of mirror neurons in speech perception and action word semantics. Lang. Cogn. Processes 25, 749–776. 10.1080/01690961003595572
    1. Hodges N. J., Franks I. M. (2001). Learning a coordination skill: interactive effects of instruction and feedback. Res. Q. Exerc. Sport 72, 132–142. 10.1080/02701367.2001.10608943
    1. Houde J. F., Nagarajan S. S., Sekihara K., Merzenich M. M. (2002). Modulation of the auditory cortex during speech: an MEG study. J. Cogn. Neurosci. 14, 1125–1138. 10.1162/089892902760807140
    1. Hueber T., Ben-Youssef A., Badin P., Bailly G., Elisei F. (2012). Vizart3D: retour articulatoire visuel pour l'aide à la pronunciation, in 29e Journées d'Études sur la Parole (JEP-TALN-RECITAL'2012), Vol. 5, 17–18.
    1. Jacks A. (2008). Bite block vowel production in apraxia of speech. J. Speech Lang. Hear. Res. 51, 898–913. 10.1044/1092-4388(2008/066)
    1. Jakobson R., Fant G., Halle M., Jakobson R. J., Fant R. G., Halle M. (1952). Preliminaries to Speech Analysis: the Distinctive Features and their Correlates. Technical Report, Acoustics Laboratory. No. 13. MIT.
    1. Katz W. F., Campbell T. F., Wang J., Farrar E., Eubanks J. C., Balasubramanian A., et al. (2014). Opti-speech: A real-time, 3D visual feedback system for speech training, in Procceedings of Interspeech. Available online at:
    1. Katz W. F., McNeil M. R. (2010). Studies of articulatory feedback treatment for apraxia of speech based on electromagnetic articulography. SIG 2 Perspect. Neurophysiol. Neurogenic Speech Lang. Disord. 20, 73–79. 10.1044/nnsld20.3.73
    1. Keating P., Lahiri A. (1993). Fronted velars, palatalized velars, and palatals. Phonetica 50, 73–101. 10.1159/000261928
    1. Kohler E., Keysers C., Umiltá M. A., Fogassi L., Gallese V., Rizzolatti G. (2002). Hearing sounds, understanding actions: action representation in mirror neurons. Science 297, 846–848. 10.1126/science.1070311
    1. Kröger B. J., Birkholz P., Hoffmann R., Meng H. (2010). Audiovisual tools for phonetic and articulatory visualization in computer-aided pronunciation training, in Development of Multimodal Interfaces: Active Listening and Synchrony: Second COST 2102 International Training School, Vol. 5967, eds Esposito A., Campbell N., Vogel C., Hussain A., Nijholt A. (Dublin: Springer; ), 337–345.
    1. Kröger B. J., Kannampuzha J. (2008). A neurofunctional model of speech production including aspects of auditory and audio-visual speech perception, in International Conference onf Auditory-Visual Speech Processing (Queensland: ), 83–88.
    1. Kröger B. J., Kannampuzha J., Neuschaefer-Rube C. (2009). Towards a neurocomputational model of speech production and perception. Speech Commun. 51, 793–809. 10.1016/j.specom.2008.08.002
    1. Kroos C. (2012). Evaluation of the measurement precision in three-dimensional electromagnetic articulography (Carstens AG500). J. Phonet. 40, 453–465. 10.1016/j.wocn.2012.03.002
    1. Ladefoged P., Maddieson I. (1996). The Sounds of the World's Languages. Oxford: Blackwell.
    1. Levitt J. S., Katz W. F. (2008). Augmented visual feedback in second language learning: training Japanese post-alveolar flaps to American English speakers, in Proceedings of Meetings on Acoustics, Vol. 2 (New Orleans, LA: ), 060002.
    1. Levitt J. S., Katz W. F. (2010). The effects of EMA-based augmented visual feedback on the English speakers' acquisition of the Japanese flap: a perceptual study, in Procceedings of Interspeech (Chiba: Makuhari: ), 1862–1865.
    1. Liu X., Hairston J., Schrier M., Fan J. (2011). Common and distinct networks underlying reward valence and processing stages: a meta-analysis of functional neuroimaging studies. Neurosci. Biobehav. Rev. 35, 1219–1236. 10.1016/j.neubiorev.2010.12.012
    1. Liu Y., Massaro D. W., Chen T. H., Chan D., Perfetti C. (2007). Using visual speech for training Chinese pronunciation: an in-vivo experiment, in SLaTE. 29–32.
    1. Maas E., Mailend M.-L., Guenther F. H. (2015). Feedforward and feedback control in apraxia of speech (AOS): effects of noise masking on vowel production. J. Speech Lang. Hear. Res. 58, 185–200. 10.1044/2014_JSLHR-S-13-0300
    1. Maas E., Robin D. A., Austermann Hula S. N., Freedman S. E., Wulf G., Ballard K. J., et al. . (2008). Principles of motor learning in treatment of motor speech disorders. Am. J. Speech Lang. Pathol. 17, 277–298. 10.1044/1058-0360(2008/025)
    1. Marian V. (2009). Audio-visual integration during bilingual language processing., in The Bilingual Mental Lexicon: Interdisciplinary Approaches, ed Pavlenko A. (Bristol, UK: Multilingual Matters; ), 52–78.
    1. Massaro D. W. (1984). Children's perception of visual and auditory speech. Child Dev. 55, 1777–1788. 10.2307/1129925
    1. Massaro D. W. (2003). A computer-animated tutor for spoken and written language learning, in Proceedings of the 5th International Conference on Multimodal Interfaces (New York, NY: ACM; ), 172–175. 10.1145/958432.958466
    1. Massaro D. W., Bigler S., Chen T. H., Perlman M., Ouni S. (2008). Pronunciation training: the role of eye and ear, in Proceedings of Interspeech, (Brisbane, QLD: ), 2623–2626.
    1. Massaro D. W., Cohen M. M. (1998). Visible speech and its potential value for speech training for hearing-impaired perceivers. in STiLL-Speech Technology in Language Learning (Marholmen: ), 171–174.
    1. Massaro D. W., Light J. (2003). Read my tongue movements: bimodal learning to perceive and produce non-native speech /r/ and /l/, in Proceedings of Eurospeech (Interspeech) (Geneva: 8th European Conference on Speech Communication and Technology).
    1. Massaro D. W., Liu Y., Chen T. H., Perfetti C. (2006). A multilingual embodied conversational agent for tutoring speech and language learning, in Proceedings of Interspeech (Pittsburgh, PA: ).
    1. Max L., Guenther F. H., Gracco V. L., Ghosh S. S., Wallace M. E. (2004). Unstable or insufficiently activated internal models and feedback-biased motor control as sources of dysfluency: a theoretical model of stuttering. Contemp. Issues Commun. Sci. Disord. 31, 105–122.
    1. McGurk H., MacDonald J. (1976). Hearing lips and seeing voices. Nature 264, 746–748. 10.1038/264746a0
    1. Mehta S., Katz W. F. (2015). Articulatory and acoustic correlates of English front vowel productions by native Japanese speakers. J. Acoust. Soc. Am. 137, 2380–2380. 10.1121/1.4920648
    1. Mochida T., Kimura T., Hiroya S., Kitagawa N., Gomi H., Kondo T. (2013). Speech misperception: speaking and seeing interfere differently with hearing. PLoS ONE 8:e68619. 10.1371/journal.pone.0068619
    1. Möttönen R., Schürmann M., Sams M. (2004). Time course of multisensory interactions during audiovisual speech perception in humans: a magnetoencephalographic study. Neurosci. Lett. 363, 112–115. 10.1016/j.neulet.2004.03.076
    1. Navarra J., Soto-Faraco S. (2007). Hearing lips in a second language: visual articulatory information enables the perception of second language sounds. Psychol. Res. 71, 4–12. 10.1007/s00426-005-0031-5
    1. Nordberg A., Göran C., Lohmander A. (2011). Electropalatography in the description and treatment of speech disorders in five children with cerebral palsy. Clin. Linguist. Phon. 25, 831–852. 10.3109/02699206.2011.573122
    1. Numbers M. E., Hudgins C. V. (1948). Speech perception in present day education for deaf children. Volta Rev. 50, 449–456.
    1. O'Neill J. J. (1954). Contributions of the visual components of oral symbols to speech comprehension. J. Speech Hear. Disord. 19, 429–439.
    1. Ojanen V., Möttönen R., Pekkola J., Jääskeläinen I. P., Joensuu R., Autti T., et al. . (2005). Processing of audiovisual speech in Broca's area. Neuroimage 25, 333–338. 10.1016/j.neuroimage.2004.12.001
    1. Ouni S. (2013). Tongue control and its implication in pronunciation training. Comp. Assist. Lang. Learn. 27, 439–453. 10.1080/09588221.2012.761637
    1. Pekkola J., Ojanen V., Autti T., Jääskeläinen I. P., Möttönen R., Sams M. (2006). Attention to visual speech gestures enhances hemodynamic activity in the left planum temporale. Hum. Brain Mapp. 27, 471–477. 10.1002/hbm.20190
    1. Pochon J. B., Levy R., Fossati P., Lehericy S., Poline J. B., Pillon B., et al. . (2002). The neural system that bridges reward and cognition in humans: an fMRI study. Proc. Natl. Acad. Sci. U.S.A. 99, 5669–5674. 10.1073/pnas.082111099
    1. Preston J. L., Leaman M. (2014). Ultrasound visual feedback for acquired apraxia of speech: a case report. Aphasiology 28, 278–295. 10.1080/02687038.2013.852901
    1. Preston J. L., McCabe P., Rivera-Campos A., Whittle J. L., Landry E., Maas E. (2014). Ultrasound visual feedback treatment and practice variability for residual speech sound errors. J. Speech Lang. Hear. Res. 57, 2102–2115. 10.1044/2014_JSLHR-S-14-0031
    1. Pulvermüller F., Fadiga L. (2010). Active perception: sensorimotor circuits as a cortical basis for language. Nat. Rev. Neurosci. 11, 351–360. 10.1038/nrn2811
    1. Pulvermüller F., Huss M., Kherif F., Moscoso del Prado Martin F., Hauk O., Shtyrov Y. (2006). Motor cortex maps articulatory features of speech sounds. Proc. Natl. Acad. Sci. U.S.A. 103, 7865–7870. 10.1073/pnas.0509989103
    1. Reetz H., Jongman A. (2009). Phonetics: Transcription, Production, Acoustics, and Perception. Chichester: Wiley-Blackwell.
    1. Reisberg D., McLean J., Goldfield A. (1987). Easy to hear but hard to understand: a lip-reading advantage with intact auditory stimuli, in Hearing by Eye: The Psychology of Lip-reading, eds Dodd B., Campbell R. (Hillsdale, NJ: Lawrence Erlbaum Associates; ), 97–114.
    1. Rizzolatti G., Arbib M. A. (1998). Language within our grasp. Trends Neurosci. 21, 188–194. 10.1016/S0166-2236(98)01260-0
    1. Rizzolatti G., Cattaneo L., Fabbri-Destro M., Rozzi S. (2014). Cortical mechanisms underlying the organization of goal-directed actions and mirror neuron-based action understanding. Physiol. Rev. 94, 655–706. 10.1152/physrev.00009.2013
    1. Rizzolatti G., Craighero L. (2004). The mirror-neuron system. Annu. Rev. Neurosci. 27, 169–192. 10.1146/annurev.neuro.27.070203.144230
    1. Sams M., Möttönen R., Sihvonen T. (2005). Seeing and hearing others and oneself talk. Cogn. Brain Res. 23, 429–435. 10.1016/j.cogbrainres.2004.11.006
    1. Sato M., Troille E., Ménard L., Cathiard M.-A., Gracco V. (2013). Silent articulation modulates auditory and audiovisual speech perception. Exp. Brain Res. 227, 275–288. 10.1007/s00221-013-3510-8
    1. Schmidt R., Lee T. (2013). Motor Learning and Performance: From Principles to Application, 5th Edn. Champaign, IL: Human Kinetics.
    1. Scruggs T. E., Mastropieri M. A., Casto G. (1987). The quantitative synthesis of single-subject research methodology and validation. Remedial Special Educ. 8, 24–33. 10.1177/074193258700800206
    1. Scruggs T. E., Mastropieri M. A., Cook S. B., Escobar C. (1986). Early intervention for children with conduct disorders: a quantitative synthesis of single-subject research. Behav. Disord. 11, 260–271.
    1. Shirahige C., Oki K., Morimoto Y., Oisaka N., Minagi S. (2012). Dynamics of posterior tongue during pronunciation and voluntary tongue lift movement in young adults. J. Oral. Rehabil. 39, 370–376. 10.1111/j.1365-2842.2011.02283.x
    1. Sigrist R., Rauter G., Riener R., Wolf P. (2013). Augmented visual, auditory, haptic, and multimodal feedback in motor learning: a review. Psychon. Bullet. Rev. 20, 21–53. 10.3758/s13423-012-0333-8
    1. Skipper J. I., Goldin-Meadow S., Nusbaum H. C., Small S. L. (2007a). Speech-associated gestures, Broca's area, and the human mirror system. Brain Lang. 101, 260–277.
    1. Skipper J. I., Nusbaum H. C., Small S. L. (2005). Listening to talking faces: motor cortical activation during speech perception. Neuroimage 25, 76–89. 10.1016/j.neuroimage.2004.11.006
    1. Skipper J. I., Nusbaum H. C., Small S. L. (2006). Lending a helping hand to hearing: another motor theory of speech perception., in Action to Language Via the Mirror Neuron System, ed Arbib M. A. (Cambridge: Cambridge University Press; ), 250–285.
    1. Skipper J. I., van Wassenhove V., Nusbaum H. C., Small S. L. (2007b). Hearing lips and seeing voices: how cortical areas supporting speech production mediate audiovisual speech perception. Cereb. Cortex 17, 2387–2399. 10.1093/cercor/bhl147
    1. Stella M., Stella A., Sigona F., Bernardini P., Grimaldi M., Gili Fivela B. (2013). Electromagnetic Articulography with AG500 and AG501, in 14th Annual Conference of the International Speech Communication Association (Lyon: ), 1316–1320.
    1. Stevens K. N. (2008). Acoustic Phonetics. Cambridge, MA: MIT Press.
    1. Stevens K. N., Blumstein S. E. (1975). Quantal aspects of consonant production and perception: a study of retroflex stop consonants. J. Phonet. 3, 215–233.
    1. Stevens K. N., Blumstein S. E. (1978). Invariant cues for place of articulation in stop consonants. J. Acoustic. Soc. Am. 64, 1358–1368. 10.1121/1.382102
    1. Suemitsu A., Ito T., Tiede M. (2013). An EMA-based articulatory feedback approach to facilitate L2 speech production learning. J. Acoustic. Soc. Am. 133, 3336 10.1121/1.4805613
    1. Sumby W. H., Pollack I. (1954). Visual contribution to speech intelligibility in noise. J. Acoustic. Soc. Am. 26, 212–215. 10.1121/1.1907309
    1. Summerfield Q., McGrath M. (1984). Detection and resolution of audio-visual incompatibility in the perception of vowels. Q. J. Exp. Psychol. Hum. Exp. Psychol. 36, 51–74. 10.1080/14640748408401503
    1. Swinnen S. P., Walter C. B., Lee T. D., Serrien D. J. (1993). Acquiring bimanual skills: contrasting forms of information feedback for interlimb decoupling. J. Exp. Psychol. Learn. Mem. Cogn. 19, 1328. 10.1037/0278-7393.19.6.1328
    1. Terband H., Maassen B. (2010). Speech motor development in Childhood Apraxia of Speech: generating testable hypotheses by neurocomputational modeling. Folia Phoniatr. Logop. 62, 134–142. 10.1159/000287212
    1. Terband H., Maassen B., Guenther F. H., Brumberg J. (2009). Computational neural modeling of speech motor control in Childhood Apraxia of Speech (CAS). J. Speech Lang. Hear. Res. 52, 1595–1609. 10.1044/1092-4388(2009/07-0283)
    1. Terband H., Maassen B., Guenther F. H., Brumberg J. (2014a). Auditory–motor interactions in pediatric motor speech disorders: neurocomputational modeling of disordered development. J. Communic. Disord. 47, 17–33. 10.1016/j.jcomdis.2014.01.001
    1. Terband H., van Brenk F., van Doornik-van der Zee A. (2014b). Auditory feedback perturbation in children with developmental speech sound disorders. J. Communic. Disord. 51, 64–77. 10.1016/j.jcomdis.2014.06.009
    1. Tian X., Poeppel D. (2010). Mental imagery of speech and movement implicates the dynamics of internal forward models. Front. Psychol. 1:166. 10.3389/fpsyg.2010.00166
    1. Uddin L. Q., Molnar-Szakacs I., Zaidel E., Iacoboni M. (2006). rTMS to the right inferior parietal lobule disrupts self–other discrimination. Soc. Cogn. Affect. Neurosci. 1, 65–71. 10.1093/scan/nsl003
    1. Wik P., Engwall O. (2008). Can visualization of internal articulators support speech perception?, in Proceedings of Interspeech (Brisbane: ), 2627–2630.
    1. Wilson S. M., Iacoboni M. (2006). Neural responses to non-native phonemes varying in producibility: evidence for the sensorimotor nature of speech perception. Neuroimage 33, 316–325. 10.1016/j.neuroimage.2006.05.032
    1. Wilson S., Saygin A. P., Sereno M. I., Iacoboni M. (2004). Listening to speech activates motor areas involved in speech production. Nat. Neurosci. 7, 701–702. 10.1038/nn1263
    1. Yano J., Shirahige C., Oki K., Oisaka N., Kumakura I., Tsubahara A., et al. . (2015). Effect of visual biofeedback of posterior tongue movement on articulation rehabilitation in dysarthria patients. J. Oral Rehabil. 42, 571–579. 10.1111/joor.12293
    1. Zaehle T., Geiser E., Alter K., Jancke L., Meyer M. (2008). Segmental processing in the human auditory dorsal stream. Brain Res. 1220, 179–190. 10.1016/j.brainres.2007.11.013

Source: PubMed

3
Iratkozz fel