245 16 3MB
English Pages 276 [274] Year 2014
Pronunciation is in the Brain, not in the Mouth
Pronunciation is in the Brain, not in the Mouth
A Cognitive Approach to Teaching it
Edward Y. Odisho
9
34 2014
Gorgias Press LLC, 954 River Road, Piscataway, NJ, 08854, USA www.gorgiaspress.com Copyright © 2014 by Gorgias Press LLC
All rights reserved under International and Pan-American Copyright Conventions. No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise without the prior written permission of Gorgias Press LLC. 2014
ܚ
9
ISBN 978-1-4632-0415-0
Library of Congress Cataloging-in-Publication Data Odisho, Edward Y. Pronunciation is in the brain, not in the mouth : a cognitive approach to teaching it / By Edward Odisho. p. cm. ISBN 978-1-4632-0415-0 1. English language--Pronunciation. 2. Cognitive grammar. 3. Psycholinguistics. I. Title. PE1137.O423 2014 421’.540071--dc23 2014032984 Printed in the United States of America
TABLE OF CONTENTS Table of Contents ...................................................................... v Foreword................................................................................ xiii Acknowledgments .................................................................. xix Lists of Symbols and Phonetic Labels ...................................... xxi Chapter 1: My Story with Languages, Pronunciation and Accent ............................................................................... 1 1.1. Prelude ....................................................................... 1 1.2. The Evolution of my Interest in Linguistics and Phonetics ................................................................... 2 1.2.1. Natural Language Internalization: Language Acquisition ............................................ 3 1.2.2. A Major in English Language in a nonEnglish Environment ............................................. 4 1.2.3. Full Immersion as an Adult in Two Languages ............................................................. 5 1.2.4. Phonetic and Linguistic Orientation in Graduate Education ............................................. 10 1.2.5. Educational and Professional Challenges in the U.S. ........................................................... 11 1.3. The Impact of my Linguistic/Professional Background on the Evolution of an Approach ........... 12 1.3.1. Impact of my Linguistic Background ................. 12 1.3.2. Impact of my Teaching Career .......................... 14 1.4. Concluding Remarks ................................................. 20 1.4.1. Childhood Trilingualism Triggered Interest in Languages........................................... 21 1.4.2. Learning Kurdish Triggered Interest in Linguistics ........................................................... 21 1.4.3. Graduate Study Immersed me in Phonetics and Linguistics .................................... 22 1.4.4. Professional Challenges in the U.S. .................... 23 v
vi
PRONUNCIATION IS IN THE BRAIN
Chapter 2: The Cognitive Base of Language ............................. 25 2.1. Language: A Species-Specific Code of Communication ........................................................ 25 2.2. Language: A Cognitive-Social System Superimposed on other Systems ...................................................... 27 2.2.1. Vocal Tract Modification ................................... 28 2.2.2. Vocal Folds (Cords) Modes ................................ 29 2.2.3. Tongue Functions and Maneuverability ............. 30 2.2.4. Lip Configurations............................................. 30 2.2.5. Cavities Resonance............................................ 31 2.3 Brain ‘Speaking’ via Respiratory and Digestive Systems .................................................................... 31 2.4. Economy in Language ............................................... 33 2.5. Conscious and Subconscious Brains ........................... 37 2.6. Concluding Remarks ................................................. 40
Chapter 3: Language in the Brain of a Child ............................ 41 3.1. Learning vs. Acquisition: Conceptual Differences ...... 41 3.2. The Brain of a Child and Language ........................... 42 3.2.1. Child Brain Formation and Maturation .............. 42 3.2.2. Formative Months and Years of Mother Tongue ................................................................ 44 3.3. Cognitive Transition in Sound Perception and Production ............................................................... 46 3.3.1. Transition from Phonetics to Phonology ............ 48 3.3.2. The Brain as the Commander-in-Chief of Language Acquisition: The Cognitive Roots of Linguistic Accent ............................................. 50 3.4. Fossilization or Psycholinguistic Insensitivity ............ 52 3.5. There is Room in the Human Brain for more than One Language .......................................................... 54 3.6. Narrowing Down the Broad Definition of Accent....... 55 3.7. Implications for Understanding the Cognitive Nature of Accent ...................................................... 56 3.8. Concluding Remarks ................................................. 57 Chapter 4: Linguistic Accent: Definition, Classification and Demonstration ................................................................. 59 4.1. Introductory Remarks ............................................... 59 4.2. Intralanguage and Interlanguage Accents .................. 60 4.3. Phonetic and Phonological Accents ........................... 62
TABLE OF CONTENTS
vii
4.4. Accent: A Normal Linguistic Phenomenon................. 64 4.5. What is Meant by Accent Acquisition, Accent Reduction and Accent Impersonation ....................... 65 4.5.1. Accent Acquisition ............................................ 66 4.5.2. Accent Reduction (Remediation) ....................... 67 4.5.3. Accent Impersonation or Faking ........................ 69 4.5.4. Intralanguage Accent Reduction and Impersonation ..................................................... 72 4.6. Cultural Accent ......................................................... 73 4.7. Transition of Accent into Orthography ...................... 74 4.8. Concluding Remarks ................................................. 76
Chapter 5: A Broad Base for Understanding the Pedagogy of Teaching Pronunciation ................................................... 79 5.1. Introductory Remarks ............................................... 79 5.1.1. Speech: A Cognitive Phenomenon ..................... 80 5.1.2. Pronunciation: Multisensory Access .................. 81 5.1.3. Pronunciation: Multicognitive Access ................ 82 5.1.4. Pronunciation: An Integrated and Holistic Process ................................................................ 83 5.1.5. Pronunciation: Top-Down & Bottom-Up Dynamics ............................................................ 84 5.1.6. Pronunciation: The Complementary Nature of Acquisition and Learning ..................... 85 5.1.7. Pronunciation: A Natural Gift for Children ........ 86 5.1.8. Pronunciation Should be Premised on a Triangular Base of Perception, Recognition and Production .................................................... 87 5.1.9. Pronunciation & Psycholinguistic Insensitivity......................................................... 89 5.1.10. Pronunciation: Understanding its Scientific Premises............................................... 90 5.1.11. Pronunciation: Its Feedback Mechanisms ........ 91 5.1.12. Pronunciation: In Light of Multiple Intelligences Theory ............................................ 91 5.1.13. Pronunciation: A Generative Skill.................... 92 5.1.14. Pronunciation: Interactive Involvement of Instructors and Learners .................................. 93 5.2. Concluding Remarks ................................................. 94
viii
PRONUNCIATION IS IN THE BRAIN
Chapter 6: Ten Commandments for Teaching Effective Pronunciation .................................................................. 95 6.1. Introductory Remarks ............................................... 95 6.1.1. Thou Shall Teach Pronunciation as a Cognitive Undertaking ........................................ 97 6.1.2. Thou Shall Teach Children and Adults Differently ........................................................... 97 6.1.3. Thou Shall be Qualified for Instruction in Pronunciation...................................................... 98 6.1.4. Thou Shall Familiarize Learners with Human Speech Production .................................. 99 6.1.5. Thou Shall Orient Learners Psychologically ................................................... 99 6.1.6. Thou Shall Use all Sensory Modalities to Prop up Instruction ........................................... 100 6.1.7. Thou Shall Use all Cognitive Modalities to Prop up Instruction ........................................... 101 6.1.8. Thou Shall Transform Learners from Listeners into Performers ................................... 101 6.1.9. Thou Shall Refrain from Insistence on a Learner.............................................................. 102 6.1.10. Thou Shall Make the Classroom a Place for Learning and Fun ......................................... 102 6.2. Concluding Remarks ............................................... 103 Chapter 7: Examples of Cross-Language Accent-Causing Consonants .................................................................... 105 7.1. Introductory Remarks ............................................. 105 7.2. Outline of the English Consonant System ................ 105 7.2.1. Interdental Pair /, / ..................................... 106 7.2.2. Approximant /r/ ............................................. 107 7.2.3. Voiceless and Voiced Alveolar Fricatives /s/ and /z/ ........................................................ 110 7.2.4. English Plosives: /p b, t d, k g/ ....................... 111 7.2.5. Labio-Dental Fricatives /f, v/ .......................... 113 7.2.6. The Affricates /ʧ ʤ/ ....................................... 114 7.3. Concluding Remarks ............................................... 115 Chapter 8: Examples of Cross-Language Accent-Causing Vowels ........................................................................... 117 8.1. Salient Features in General Vowel Description ........ 117
TABLE OF CONTENTS
ix
8.2. The Vowel System of English .................................. 121 8.2.1. Simple Vowels of General American English .............................................................. 122 8.3. Selections of Cross-Language Accent-Causing Vowels ................................................................... 124 8.3.1. Hispanic Learners of English Vowels ............... 125 8.3.2. Arab Learners of English Vowels ..................... 128 8.4. Concluding Remarks ............................................... 131
Chapter 9: Examples of Cross-Language Accent-Causing Suprasegmentals ............................................................ 133 9.1. A Description of the Most Salient Features of Suprasegmentals..................................................... 133 9.2. Stress and Rhythm .................................................. 136 9.3. Tone and Intonation................................................ 140 9.4. Basic Pitch Patterns................................................. 140 9.5. Consonant Clusters.................................................. 141 9.6. Concluding Remarks ............................................... 145
Chapter 10: The Role of Articulatory Settings in Pronunciation and Accent .............................................. 147 10.1. Introductory Remarks ........................................... 147 10.2. Salient Features of Articulatory Settings of Selected Languages................................................. 150 10.2.1. English Articulatory Settings ......................... 150 10.2.2. Spanish Articulatory Settings ........................ 155 10.2.3. Arabic Articulatory Settings .......................... 162 10.3. Concluding Remarks ............................................. 165 Chapter 11: Principles of a Multicognitive Approach to Teaching Pronunciation ................................................. 167 11.1. Introductory Remarks ........................................... 167 11.2. Multicognitive Principles for Teaching Pronunciation......................................................... 169 11.2.1. Think about L2 Speech Sounds ...................... 170 11.2.2. Transition from Hearing to Listening............. 171 11.2.3. Learn Something about Speech Production ........................................................ 171 11.2.4. Mechanical Repetition Hardly Works with Adults L2 Learning .................................... 172 11.2.5. Follow the ‘Perceive, Recognize and Produce’ Procedure ........................................... 173
x
PRONUNCIATION IS IN THE BRAIN 11.2.6. Instructor’s Academic and Professional Qualifications .................................................... 176 11.2.7. Plan Instructional Connection with Learners ............................................................ 177 11.2.8. Explain, Demonstrate and Demonstrate Multisensorily ................................................... 177 11.2.9. Deal with Pronunciation in a Holistic Fashion ............................................................. 178 11.2.10. Consider both Top-Down and BottomUp Perspectives ................................................. 179 11.2.11. Do not Confuse Memorization with Retention .......................................................... 179 11.2.12. Deal with Pronunciation as a Generative Skill ................................................. 181 11.3. Concluding Remarks ............................................. 181
Chapter 12: Principles of Multisensory Approach to Teaching Pronunciation ................................................................ 183 12.1. Introductory Remarks ........................................... 183 12.2. Multisensory Principles for Teaching Pronunciation......................................................... 184 12.2.1. Auditory Modality......................................... 184 12.2.2. Visual Modality ............................................. 186 12.2.3. Tactile, Kinesthetic, Proprioceptive Modalities ......................................................... 188 12.3. Developing Teaching and Learning Strategies ....... 189 12.3.1. Developing Teaching Strategies ..................... 189 12.3.2. Developing Learning Strategies ..................... 192 12.4. Concluding Remarks ............................................. 194 Chapter 13: Exemplary Applications of Accent Remediation Techniques .................................................................... 197 13.1. Introductory Remarks ........................................... 197 13.2. Techniques for Teaching Selected Consonants ....... 197 13.3. Techniques for Teaching Labial-Dental Sounds...... 198 13.4. Techniques for Teaching Interdental Fricatives / / ........................................................................... 202 13.5. Techniques for Teaching Tense (Long) vs. Lax (Short) Vowels ....................................................... 207 13.6. Techniques for Teaching Vowel Reduction ............ 212 13.7. Techniques for Teaching Accentuation (Stress) ..... 217
TABLE OF CONTENTS
xi
13.8. Concluding Remarks ............................................. 223
Chapter 14: Tips for Accent Reduction and Accent Detection. 225 14.1. Introductory Remarks ........................................... 225 14.2. Tips for Accent Reduction ..................................... 226 14.2.1. Tackle the most Salient Phonological Problems ........................................................... 226 14.2.2. Tackle the most Salient Phonetic Problems ........................................................... 229 14.2.3. Improve other Linguistic Skills ...................... 230 14.3. Accent Detection ................................................... 230 14.3.1. Accent Detection by Ordinary Individuals ........................................................ 231 14.3.2. Accent Detection by Professionals ................. 231 14.3.3. Telling the Linguistic Background of a Speaker through Accent .................................... 232 14.3.4 Hiding an Agent through Hiding an Accent ............................................................... 235 14.4. Concluding Remarks ............................................. 239 References ............................................................................. 241
FOREWORD My fascination with human language, in general, and pronunciation, in particular, seems to be intimately connected to my trilingual upbringing as a child in the city of Kirkuk/Iraq and my later exposure to three more languages. The immense diversity in the sound systems of those six languages afforded me extensive coverage of a wide range of sounds, sound patterns and sound systems. Gradually, throughout the five decades that I spent in teaching English language, linguistics and pronunciation in different countries and to speakers of a wide variety of languages, my passion for the diversity in their pronunciation patterns deepened. Thus, teaching pronunciation did not just become my favorite subject, but it also became the focus of my academic research. Throughout the last decade, I became fascinated with the concept of accent and whether it is a sensory deficiency with some people at a certain age or the side effect of cognitive and linguistic perfection in the internalization of the native language (L1). After lengthy observations and investigations, I came to the conclusion that accent is a cognitive phenomenon and seems to be the outcome of a deficit between what is known in psycholinguistics as language acquisition and language learning. The focus in this book is, above all, on accent and its cognitive roots. Such understanding will determine the framework of an approach to teaching pronunciation. In my book: Techniques of teaching pronunciation in ESL, bilingual and foreign language classes (2003),1 I introduced several innovative principles which outlined my cognitive approach to teaching pronunciation. One such principle emphatically stated 1
München: Lincom-Europa.
xiii
xiv
PRONUNCIATION IS IN THE BRAIN
that “pronunciation is in the brain prior to being in the mouth” which constituted the foundational premise for my cognitive approach to teaching pronunciation in general. Naturally, therefore, several of 2003 principles are reintroduced in certain chapters with further refinement, elaboration and application to accent in human speech. As I am still pursuing the teaching of pronunciation in more details in this book with the linguistic phenomenon of accent being a major constituent, it certainly further highlights the significance of the cognitive perspective especially in the case of adult learners’ rendition of an L2 pronunciation. Since the cognitive potential of human beings is fed and nurtured via the five senses, the approach to teaching pronunciation has been identified as ‘multisensory and multicognitive’. The former emphasizes the significance of the joint involvement of as many sensory modalities as possible especially those that feed the cognitive processing of human language and speech, namely, auditory, visual and tactile/kinesthetic, whereas the latter highlights the mobilization of all cognitive processes that human beings practice such as thinking, comparing, contrasting, analyzing, synthesizing, associating and memorizing, among others. It is worth underscoring that accent is not a pathological defect and its remediation is not exclusively the responsibility of a speech pathologist as claimed by some. In an accent modification instruction course announcement, it is stated that “a speech-language pathologist then teaches the client appropriate speech modification techniques”.2 Accent remediation falls within the professional domain of any person qualified in the science/art of speech production, foremost of whom is a phonetician. Nevertheless, each one of those who are professionally qualified to deal with the teaching of speech and pronunciation should combine his/her academic expertise with an effective instructional methodology that incorporates the latest linguistic and educational findings. 2
http://www.speechenrichmentcenter.com/index.php?option=com_Cont
ent&view=article&id=10
FOREWORD
xv
Hence, if accent is not a ‘pathology’, then what is it? From the perspective of this book, it is one of the normal symptomatic side effects of the cognitive and physical maturation of human beings and the gradual transformation in delegating the majority of conscious biological and social survival functions, foremost of which being language, to the subconscious brain. Stated differently, accent is consequential of cognitive transformation of humans from all-conscious operants to predominantly subconscious ones. The recent literature on human language internalization is replete with substantial evidence that the perfection in the mastery of the native language, especially its pronunciation, falls primarily within the years of childhood ranging from birth—in fact, even before birth—until puberty and adolescence. Once a person is fully immersed in the native language (L1) throughout those years, he/she will grow up accentless, but he/she is highly likely, beyond that period, to manifest a certain degree of accent in L2 ranging from light to heavy. It is the inherent cognitive bias to L1 that interferes with the accurate perception and recognition, hence production, of L2 sound system. The years of childhood are so significant for language internalization that a child can grow up as balanced bilingual or even balanced trilingual if there is ample exposure to and immersion in more than one language. My childhood is a typical example. This is so because the brain of any normal child is neurolinguistically wired for such a mission and is powerful enough to handle more than one language. Accent, therefore, tends to be a symptomatic trait of adulthood resulting from the cognitive maturation with regard to L1 rather than being a physical sensory deficiency in the perception and production of speech sounds by L2 learners. It is this difference that has convinced linguists to label the natural internalization of L1 by children as acquisition, while adults’ attempt at internalizing L2 as learning. Doubtless, the two processes are not treated here as mutually exclusive as some inexperienced persons may make it look like; rather they are complementary in nature depending on the age of the person among several other factors to be dealt with in due course. Nevertheless, the younger the person is the greater the role of acquisition in comparison to learning and the reversal of the roles with adults.
xvi
PRONUNCIATION IS IN THE BRAIN
The emergence of accent in the speech of adult L2 learners should never imply the closing of the window on the likelihood of improvement of pronunciation and the remediation of accent. The claim that adults suffer from the so-called fossilization in their learning of L2 pronunciation (Selinker, 1972) is rejected in this study partly due to the lack of a thorough understanding at the time of the cognitive nature of L1 acquisition (internalization) and its subsequent impact on L2 learning and partly due to the inefficient approach and methodologies used to teach pronunciation to adults and their accent remediation. A multisensory multicognitive approach has proven very effective in learning pronunciation and teaching it to adults of different linguistic backgrounds and the remediation of their accent to different degrees depending on several other specific factors such as aptitude, enthusiasm, focus, time allocation etc… All those factors have been relevant in my personal case as I, at the age of thirtythree (33), improved my pronunciation of English and immensely broadened my inventory of different sounds in perception, recognition and production. Certainly, remediation may not be perfect in most cases, but it is definitely evident. As hinted earlier on, the cornerstone of implementing this approach is primarily the ‘multisensory multicognitive’ principles; however, there are several other functional considerations for implementation foremost of which is the distinction between phonological accent and phonetic accent. The former represents sound substitutions that directly result in semantic confusion in words as well as in sentences, whereas phonetic accent may not result in semantic confusion directly, but it may generate noise or uncertainty that interferes with proper conveyance of meaning. Although I had actively implemented this functional distinction in my classes in mid 1990s, it appeared formally in print in 2003. Field and classroom observations substantiate the paramount functional significance of this dichotomy. In teaching pronunciation, the remediation of phonological accent should take precedence over phonetic accent. To demonstrate, it is far more significant to teach a Hispanic learner of English not to substitute a [j] sound as in = [ju] with [ʤ] and change the word to = [ʤu] than to substitute a tap or trill for the English approximant as the former ex-
FOREWORD
xvii
ample represents a phonological accent, whereas the latter represents a phonetic one. In light of what preceded, the effective and efficient teaching of pronunciation of an L2 to adults requires highly qualified professionals implementing an approach that seriously takes into consideration the cognitive nature of pronunciation and the implementational approach that is compatible with it. This book is all about such an approach. The book is in fourteen (14) chapters with the first chapter being a reflection on the author’s trilingualism as a child and later addition of three more languages as well as the experiences that led to his majoring in linguistics with focus on phonetic science and pronunciation. If the reader is eager to go directly to the core materials related to human speech and pronunciation, he/she can skip the first chapter although the chapter is rich with personal experiences. Chapters two through five are essential to understanding the cognitive nature of human language, pronunciation and accent. The reader is expected to encounter difficulty in understanding certain contents of the next chapters without thoroughly absorbing the contents of those four chapters. Chapter six represents a sketch of the most significant principles to be considered for the application of the multisensory multicognitive approach to teaching pronunciation. As for chapters seven through nine, they cover the most common segmental (consonants and vowels), suprasegmental (stress, rhythm, tone, intonation) that cause pronunciation problems and accent in cross-language learning. Intimately related to those chapters is chapter ten which deals with the very important but less known aspect of pronunciation the so-called ‘articulatory settings’. Chapters eleven and twelve elaborate on the main applications of the multicognitive modalities and multisensory modalities, respectively. The book ends with chapters thirteen and fourteen which focus more narrowly on accent and its acquisition, reduction and remediation. Finally, a couples stylistic notes are unavoidable so that the reader would not misinterpret my intention as the writer. First, parts of the book are written in the first person simply because they reflect my personal life and experience; it would, therefore, be awkward to state them in third person pronoun. Second, my use of ‘he’ pronoun for the learner does not exclude the ‘she’. I
xviii
PRONUNCIATION IS IN THE BRAIN
have tried the combination he/she’, but after a while it becomes stylistically too repetitive. I am a liberal thinker and my acknowledgements recognize three women that have deeply impacted my life Edward Y Odisho March 3, 2014
ACKNOWLEDGMENTS I recently celebrated my 75th birthday and with it came my decision not to write more books after this, my 11th one, but rather focus on producing more research papers which reflect some of the thoughts and themes that have been with me for a long time. Certainly, I have not had the opportunity to develop them and bring them all to fruition. However, I seize this opportunity to express my deep gratitude to all institutions and publishers which have assisted me in transforming my writings into books and promoting them locally and internationally. My first book was published by the Iraqi Ministry of Culture in 1971 followed by the second one published by Al-Mustansiriya University/Baghdad. After my escape from Iraq during Saddam’s regime in 1980, my third book was published by Otto Harrassowitz Verlag/Wiebaden, 1988. Between 1988 and 2002, my academic efforts were devoted to publishing research papers which put me on fast track for tenure and academic promotion to professorship. In 2003, I decided to produce a series of books to document my experience in classroom teaching. Towards the end of 2003 my book Techniques of teaching pronunciation in ESL, bilingual and foreign language classes was published by Lincom Europa/München followed by A linguistic and cognitive approach to the teaching of the English alphabet, Edwin Mellen Press/New York in 2004. Between 2007 and 2011, I published the following books all by Gorgias Press of New Jersey. Techniques of teaching comparative pronunciation: English-Arabic, 2005. Linguistic tips for Latino learners and teachers of English, 2007. Linguistic and cultural studies in Aramaic and Arabic, 2009. Modern Assyrian (Aramaic) language between speech and writing: linguistic examination, 2011. Although this book still deals with pronunciation, it has a somewhat narrower thematic focus as it attempts to deal with accent in pronunciation in greater academic and empirical depth. It certainly incorporates the views of many distinguished xix
xx
PRONUNCIATION IS IN THE BRAIN
scholars and experts in the fields that deal with human sound production intricacies, at large, and pronunciation, in particular. To those views, I added the personal reflections on my childhood as a balanced trilingual alongside my experiences accumulated during five decades of classroom teaching and real-life observations. In publishing this book, I must express my gratitude to tens of thousands of students that I have taught and who, in return, taught me through their difficulties as well as their successes in overcoming those difficulties. Finally, there are a few people to whom I am indebted. There are three women who have changed my life: my mother, Shakira, with whom I shared my interest in books, my doctoral supervisor, Celia Scully, who twisted my arm to be more science-oriented and my wife, Wardia Shamiran, for being my friend and partner in whatever I have accomplished during the last four decades of our marriage. Edward Odisho, Morton Grove, Illinois, March 3, 2014
LISTS OF SYMBOLS AND PHONETIC LABELS The conventions and symbols of the International Phonetic Association (IPA) and their acceptable substitutes are listed below and they have been used throughout this book. The Arabic alphabet and its diacritics are also listed because they have been employed where necessary. The following is a list of symbols and conventions used: Vowels
Phonetic Description Close front with spread lips
Close front to close-mid (somewhat central-
ized) with spread lips
Close-mid front with unrounded lips
Open-mid front with unrounded lips
Open-mid central with unrounded lips
Near-open front with unrounded lips Open front with unrounded lips Open back with unrounded lips
Open-mid back with rounded lips
Close-mid back with rounded lips
u
Close back with rounded lips
ɚ
ɝ
Near-close near-back with rounded lips Open-mid back with unrounded lips
Mid central (neutral) vowel (schwa)
R-colored (rhotacized) mid central (schwar) R-colored (rhotacized) open-mid central xxi
xxii
PRONUNCIATION IS IN THE BRAIN
Consonants Phonetic Description b Voiced bilabial plosive p p
h
Voiceless unaspirated bilabial plosive Voiceless aspirated bilabial plosive
d
Voiced alveolar plosive
t
Voiceless aspirated alveolar plosive
t
h
g
k k
h
c c
Voiceless unaspirated alveolar plosive Voiced palatal plosive Voiced velar plosive
Voiceless unaspirated velar plosive Voiceless aspirated velar plosive
Voiceless unaspirated palatal plosive h
q ʤ ʧ
Voiceless aspirated palatal plosive
Voiceless (unaspirated) uvular plosive Glottal stop
Voiced postalveolar affricate
Voiceless postalveolar affricate
v
Voiced labialdental fricative
Voiced interdental fricative
f
z s
h
Voiceless labialdental fricative Voiceless interdental fricative Voiced alveolar fricative
Voiceless alveolar fricative Voiced postalveolar fricative
Voiceless postalveolar fricative Voiced uvular fricative Voiceless uvular fricative
Voiced pharyngeal fricative
Voiceless pharyngeal fricative Voiceless glottal fricative
LIST OF SYMBOLS AND PHONETIC LABELS Voiced labialdental approximant
Voiced labialpalatal approximant
Voiced alveolar approximant
Voiced retroflex approximant
l
Voiced alveolar lateral approximant
w
Voiced labialvelar approximant
j
Voiced palatal approximant
m
Voiced bilabial nasal (approximant)
n
Voiced alveolar nasal (approximant)
Voiced velar nasal (approximant)
Voiced dental/alveolar tap Voiced retroflex tap
r
Voiced dental/alveolar trill
Diphthongs in RP English au as in
ai
as in
ou
as in
oi ei i
e
u
as in as in as in as in as in
Conventions / / Phonemic transcription [] _ _
Phonetic transcription Vowel full length
Vowel half-length
Superscript indicating aspiration
Superscript indicating strong stress
xxiii
xxiv –
C V
PRONUNCIATION IS IN THE BRAIN Subscript dot under /d t s/ = [ ] indi-
cates /ض, ط, ظ, ص/ the emphatic sounds of Arabic
In syllable structure patterns, ‘C’ stands for a ‘Con-
sonant’
stands for a ‘Vowel’
Arabic Symbols Consonants IPA Phonetic Description ء [] glottal stop
ب ت
[] voiced bilabial plosive
ث ﭖ ج ﭺ ح خ د
[] voiceless interdental fricative
ذ ر ز ﮊ س ش ص ض ط
[] voiced interdental fricative
[h] voiceless aspirated alveolar plosive [p] voiceless bilabial plosive (Farsi) [ʤ] voiced postalveolar affricate
[ʧ] voiceless postalveolar affricate (Farsi) [] voiceless pharyngeal fricative [] voiceless uvular fricative [d] voiced alveolar plosive [] voiced alveolar trill
[] voiced alveolar fricative
[] voiced postalveolar fricative (Farsi) [] voiceless alveolar fricative
[] voiceless postalveolar fricative [s
voiceless emphatic alveolar fricative
[t
voiceless (unaspirated) emphatic alveolar
[d
voiced emphatic alveolar plosive
plosive
ظ
[
voiced emphatic interdental fricative
LIST OF SYMBOLS AND PHONETIC LABELS
ع
[ʕ] voiced pharyngeal fricative
غ
[] voiced uvular fricative
ف ﭪ ق ك ﮒ ل م ن ه و
[] voiceless labialdental fricative
ي ــّــ
[]
[v] voiced labialdental fricative (Farsi)
[] voiceless unaspirated uvular plosive [] voiceless velar plosive
[g] voiced velar plosive (Farsi) []
voiced alveolar lateral
[] bilabial nasal
[] alveolar nasal
[] voiceless glottal fricative [] labialvelar approximant palatal approximant
Superscript on consonant indicating geminated (double) consonant.
Vowels (Letters)
ا ي و
[a] long counterpart of [a] [i] long counterpart of [i]
[u] long counterpart of [u]
Vowels (Diacritics) ــَــ ــِــ
Superscript over consonant indicating short [a]
vowel.
Subscript over consonant indicating short [i] vow-
el.
ــُــ
Superscript on consonant indicating short [u]
vowel.
ـــْـ
Superscript on consonant indicating absence of
vowel.
xxv
CHAPTER 1: MY STORY WITH LANGUAGES, PRONUNCIATION AND ACCENT 1.1. PRELUDE
Very simply, this book focuses on the linguistic nature of human language, in general, and pronunciation and accent, in particular. Two important principles govern the overall approach, namely, cognitive and the pedagogical principles. With regard to the first, language as a structure and system originates in society but its blueprint is in the brain. It is, therefore, a social product, but a cognitive entity. Whenever social survival needs the services of language, it signals to the brain which, in turn, activates its neurons and synapses to generate communication. As for native language (L1) pronunciation, it, ceteris paribus, is a process of natural acquisition with perfection; however, learning effective pronunciation of a second language (L2) by adults requires conscious effort by the learner assisted with the linguistic and educational knowhow of the instructor. This latter statement highlights the pedagogical principles of the approach taken in this book. The instructor should have a high level of professional competence and experience in the sound systems of the languages involved. He/she should also follow an educational philosophy that premises the success of a teaching approach on the extent of interactive connection with the learners to ascertain that there is an effective mode of two-way interaction. It is imperative that the instructor diversify his/her cognitive and sensory strategies and techniques of teaching as well as discover the individual learning styles of the learners and encourage them to get actively involved in the process.
1
2
PRONUNCIATION IS IN THE BRAIN
1.2. THE EVOLUTION OF MY INTEREST IN LINGUISTICS AND PHONETICS
There are five major linguistic experiences in my life that have contributed to the evolution of my interest in linguistics and phonetics. First, the natural acquisition of three languages in my childhood, namely, Assyrian (Modern Aramaic) Turkmeni and Arabic due to the multilingual environment in which I grew up in the city of Kirkuk/Iraq. My natural trilingualism was the result of full immersion in a context-embedded and situationembedded linguistic environment. In the framework of this study, the use of these two terms is somewhat different from their circulation in previous literature, especially by Cummins (Cummins, 1979, 1984). Context-embedded implies the use of language in discourse format rather than in isolated words and sentences, while situation-embedded implies the use of a discourse that matches the situation in which it actually takes place. To clarify the latter statement, a discussion of the action of airplanes taking-off and landing at an airport will be more authentic than a discussion about the same theme in a café or a classroom situation. This is simply because in the former scenario the actions are multisensorily and realistically perceived, hence readily comprehended. Second, for my first degree, I majored in English language at Baghdad University which implies partial immersion in the target language (L2) due to the absence of fully contextembedded and situation-embedded environments. During my four-year study for my degree, English was dominant only during the class sessions and occasional conversations with faculty members (the majority of whom were native English speakers). Outside those two environments, the day-to-day language of conversation was in Arabic or in any other native languages of Iraq such as Assyrian, Turkmeni, etc. Third, in my early adult life, I experienced full immersion in the Kurdish language. This type of full immersion was repeated later in my adult life with English in both England and the United States. Fourth, in my graduate studies of four years in England, I specialized in linguistics, in general, and phonetic science, in particular. This duration thoroughly exposed me to a wide varie-
CHAPTER 1
3
ty of sound materials from different languages. Such an experience afforded me a scientific insight into human language both as an acquisition process and a learning one. Fifth, the five-decade long professional experience in teaching I have had at different levels of education, in different countries and in diversified educational and professional situations added to the depth of my linguistic experience and honed it. Below is an elaboration on the above five experiences. 1.2.1. Natural Language Internalization: Language Acquisition
Any study of human language internalization should consider the manner in which normal children master their native language or any language they are immersed in as opposed to adults embarking on learning a second language. The two processes are known in psycholinguistic literature as acquisition vs. learning. Acquisition tends to be a subconscious, automatic and effortless process of internalizing a language, whereas learning tends to be more conscious, mechanical and effortful. Inasmuch as pronunciation acquisition is concerned, all that children need to accomplish it is ample exposure to speech in real-life contexts and situations. My hometown Kirkuk was at the time1 the most multilingual city in Iraq with the five languages of Turkmeni, Arabic, Kurdish, Assyrian and Armenian spoken in its different neighborhoods. The overwhelming majority of its population was, at minimum, bilingual with many being trilinguals. Which of the five languages one mastered, depended partly on the size of the population that spoke a given language and partly on the neighborhood in which one resided. In my case, Assyrian was my home language while Turkmeni and Arabic were the two languages to which I was most exposed. It should be clarified, however, that Arabic was then confined to a couple of neighborhoods because of the very small population of Arabs in Kirkuk at the time. Fortunately, I lived in one such neighborhood. What is important to highlight with regard to Arabic is the fact 1
Demographics have significantly changed since the 1940s.
4
PRONUNCIATION IS IN THE BRAIN
that it was the official language of the country especially in education and governmental business—everyone had to pick it up or, at least, be familiar with it. In light of the diversified linguistic environment in Kirkuk, I grew up fully orally competent in Assyrian, Turkmeni and Arabic. I always felt that functionally all three were my native tongues, the only difference between them being that Assyrian was my ethnic native tongue and my home language. The process of internalizing Turkmeni or Arabic to the degree of a native tongue can be attributed to the context-embedded and situation-embedded environment in which the two languages were encountered and used. No conscious effort was made to internalize them; they were simply naturally acquired as long as the exposure to them continued almost every day and throughout most of the day. In a nutshell, my native competency in those three languages as well as their cultures was a typical example of child language and culture acquisition. Turning to another aspect of this childhood linguistic experience and its impact on my later adult linguistic experience, several observations can be made especially with regard to my competence in the future performance of sounds and sound systems of other languages. First, this childhood experience afforded me a much broader inventory of sounds and richer competence in the production of some unfamiliar sounds of other languages. Second, on the flip side, in certain aspects of the sound systems, especially in the domain of accentuation (stress placement) my childhood experience was so dominant that it subconsciously permeated the English language that I learned in the early stages of my exposure to English. In many respects, my stress placement in English was seriously sullied by my childhood languages, especially Assyrian and Arabic. I lived with this misplacement of stress in English without knowing it until I settled in Britain for four years receiving systematic education in phonetic sciences and linguistics. There will be extended elaboration on this subject in due course. 1.2.2. A Major in English Language in a non-English Environment
Completing a major in English language at Baghdad University—obviously a non-native environment—implied partial im-
CHAPTER 1
5
mersion in the target language (L2) due to the absence of a fully context-embedded and situation-embedded environment. During my four-year study for my degree, English was dominant only during class sessions and occasional conversations with faculty members (the majority of whom were native English speakers). Outside those two environments, the day-to-day language of conversation was in Arabic and in any other native languages of Iraq such as Assyrian, Turkmeni, etc. In other words, my exposure to L2 was very limited in terms of time as well as in the variety of contexts and situations in which a language is normally used. Stated differently, we rarely, if ever, had real exposure to and use of English in, for instance, a market place situation or a casual family gathering. Throughout four years of education, my fellow students and I were capable of talking about the three witches in Shakespeare’s Macbeth more than conducting a conversation with the owner of a shop-keeper in a fruit and vegetable market. This limited exposure in terms of time, contexts and situations in which a language is naturally used is in striking contrast to the ample time and the multitude of contexts and situations in which my three childhood languages were acquired. Majoring in English in Baghdad was, to a large extent, a typical example of a context-reduced and situation-reduced experience; it was a classic example of an adult language learning model as opposed to a child language acquisition model. Consequently, my learning of English in my early adulthood in context-reduced and situationreduced environment had many deficiencies. 1.2.3. Full Immersion as an Adult in Two Languages
The following deals with my full immersion in the Kurdish and English languages as an adult and the impact of such experiences on my linguistic orientation and the later evolution of a passion for pronunciation and the pedagogy of teaching it. 1.2.3.1. Full Immersion as an Adult in all-Kurdish Environment
After gaining my degree in 1960, I was appointed as a teacher of English language in the Kurdish city of Sulaimaniya where the daily language of communication was predominantly Kurdish. Thus, I was immersed in another language as an adult. Learning
6
PRONUNCIATION IS IN THE BRAIN
Kurdish was a necessity in a city where almost all daily interactions were in this language. Additionally, for reasons unknown to me at the time, except for the fact that I was an open-minded and progressive young man with an interest in other languages and cultures, I embarked on learning the Kurdish language and culture exceptionally seriously, and with a passion. The manner in which I learned—or ‘acquired’—Kurdish turned out later to be innovative. At the time I was not a linguist to implement the approaches or techniques promoted by applied linguistics to master a second language as an adult, neither was I familiar with the modern concept of immersion in an L2 situation to acquire or learn it. I simply handled the Kurdish language in the following manner, which, decades later, I discovered to be full immersion in the target language: “The attempt at learning Kurdish went through four stages. The first stage was to focus entirely on the oral communication with minimum recourse to the written form except when a certain word or phrase was extremely necessary. For a couple months or so, I did not attempt to speak; it was simply a period of listening to other people’s communication and carefully watching their facial and body gestures that accompanied the conversation. Occasionally in classroom, I used to ask my students to translate some short statements in an English language dialogue into Kurdish and I would carefully listen to the translated segments. The second stage began around the third month when I decided to go one step further beyond the listening period. I ventured to speak by getting involved in very short conversations with some intimate Kurdish friends who appreciated my intention to learn their language. They would correct me when necessary and they would also repeat certain statements so that I would be able to internalize them. I was never intimidated by the mistakes or hesitations that would occur in my conversation. In order to carry my listening and speaking skills one step further and associate them with reallife contexts/situations, I took upon myself almost daily to go to the marketplace to do my shopping and carefully listen to live interactions between shop-keepers and customers. This experience was the most effective and efficient in help-
CHAPTER 1
7
ing me predict the meaning of many words and expressions that I did not know previously. The actual context/situation of the conversations aided me in the prediction of meaning. Perhaps, more significant than just predicting the meaning was the higher possibility of retaining the meaning also because of the clues from the context/situation. At the third stage, which began by the end of the first school year—usually nine months—I was able to sustain simple social conversations. After my marketplace ventures, I made more Kurdish friends with whom I spent my evenings in their homes or in cafés and social clubs for teachers and other civil service employees. The beginning of the second year was the final stage when my colleagues at school, my students and my friends in the community began to address me in Kurdish most of the time. I avoided using Arabic as much as possible; however, when I found myself groping for the right word in Kurdish I did not hesitate to double-dip in both Arabic and Kurdish. In other words, I resorted to some familiar linguistic devices that bilinguals use such as ‘code-switching’ or ‘code-mixing’ of two languages. I should not forget to reveal another strategy I used to teach myself Kurdish. This was a two-prong strategy of listening to songs and retaining their lyrics, as much as possible, as well as the retention of some popular proverbs and sayings. In both cases, the retention was aided by the music in the first instance and by the uniqueness of meaning and other linguistic niceties that this genre of human language usually has. By the end of the fifth year, which was the last year of my service in Sulaimaniya, my communication with people was predominantly in Kurdish. I was good in overall fluency, but excellent in pronunciation.”
After this brief outline of my experience with Kurdish, it is absolutely essential to point out that the manner in which I learned the language so successfully is still an experience which I cannot fully explain. Did I succeed with Kurdish because I was originally a multilingual or was it because I accidentally ran into an oral approach that turned out to be a context/situation-based linguistic pedagogy of which I knew nothing then except for the fact
8
PRONUNCIATION IS IN THE BRAIN
that it seemed a more natural way for human language acquisition or learning? Equally puzzling was the motive that pushed me towards a future in the study of linguistics, in general, and phonetics, in particular, as a profession while still living in a country (Iraq) where linguistics, let alone phonetics, was at the time much less known. After successfully finishing my graduate studies, the question of how I had become so obsessed with pronunciation and phonetic sciences was still nagging me and I craved for a satisfactory explanation. As I was accidentally sifting through the pages of some of my early college psychology and methodology textbooks I noticed that I had persistently underlined statements and paragraphs that were related to language learning, in general, and child language acquisition, in particular. In my view, my childhood multilingualism contributed significantly to my pursuit of linguistic studies as an adult. 1.2.3.2. Full Immersion as an Adult in an all-English Environment
My four-year stay in England placed me in a linguistic environment with maximum immersion in English coupled with extensive exposure to diversity in authentic contexts and situations in which the language was used. Those four years were, in many respects, virtually the reversal of my experience with English during the four years in Baghdad majoring in English. The longer I stayed in England and became acquainted with the linguistic and phonetic principles of language learning and teaching, the more I discovered the weaknesses and flaws in my English at all levels—grammar, style, lexicon and, above all, pronunciation which is the focus of this book. Fortunately, in the consonantal domain of English, there were no serious phonological problems simply because of my broad phonetic and phonological base attributed to my childhood trilingualism. The most noticeable consonantal mispronunciation was my rendition of the approximant English < r> as a tap or a trill one, which, fortunately, constituted a mere ‘phonet-
CHAPTER 1
9
ic accent’2 that did not interfere with meaning. Obviously, all consonant clusters involving an such as /pr, tr, cr/, etc., did cause some deviation in pronunciation. Nevertheless, this was a minor phonetic aberration compared to, for instance, stress placement in words and longer stretches of speech. For instance, in almost all patterns of verbs such as: , , , among many other word patterns, I, like the overwhelming majority of adult Assyrians and Arabs, placed stress on the final syllable which is hardly there in English. Also, in noun compounds such , , etc., where the accent is usually on the first word, I would place it on the second not knowing that it makes a huge difference in reference. A is a certain type of bird that happens to be ‘black’, whereas a is any bird that is ‘black’. These mispronunciations were all transfers from my Assyrian and Arabic languages. For example, in Assyrian the compound word for an (literally, a white-bearded man) is ܚܘܪܕܩܢܐwith accent on the second part. I did not know what stress was and how it functioned within a language, and neither did I know what intonation was. Throughout my four years of English language education in Baghdad, which was more or less literature-oriented, I had never heard an instructor draw attention to such important language dynamics.3 In a nutshell, the English we learned in Baghdad was subconsciously colored, to different extents, with the overall pronunciation of Arabic and/or the other native languages of the learners. In light of such facts, and with my schooling in the Department of Phonetics in the techniques of listening to a wide variety of sounds and producing them, I gradually began to concentrate my attention on the pronunciation of the native speakers of English with emphasis on both the segmental and supra2
It will be argued later that a ‘phonetic accent’ is a mispronunciation that
does not alter meaning as opposed to ‘phonological accent’ which does. 3
Bear in mind, I am referring to the years 1956–1960. In the years begin-
ning with the 1970s there were several faculty members in different universities in Iraq who specialized in linguistics and they would emphasize such prosodic features of both English and Arabic.
10
PRONUNCIATION IS IN THE BRAIN
segmental aspects of their pronunciation. I remember vividly how I rectified my pronunciation of the verbs and which I used to pronounce with stress on the first syllable instead of the second one. In fact, in the case of I had a transaction at the bank and I pronounced the word as and the banker said: “You mean ”. With a very mellow touch of linguistic embarrassment, I said: “Yes.” I also soon discovered that a certain category of words such as: could be rendered verbs or nouns depending on the position of stress. In this latter case, I do remember that one of our British instructors in Baghdad did mention in passing such a rule, but did not dwell enough on it neither did he demonstrate the difference for us in order to internalize what he was saying. Pedagogically, this example is typical of the fact when the instructor fails to establish a bridge of interaction with his students assuming that once a rule is mentioned everyone picks it up. The absence of such a methodological strategy of connection between instructors and learners has become one of the cornerstones in my approach to teaching: simply, make sure you and your students are on the same page and they connect with you. 1.2.4. Phonetic and Linguistic Orientation in Graduate Education
The first year of my graduate education was a two-pronged intensive orientation in theoretical and applied principles of phonology and phonetics. The theoretical component covered a wide variety of schools and theories of phonology. The applied side of it involved thorough exposure to human speech from the acoustic, physiological and aerodynamic perspectives which were reinforced with experimental laboratory work as well as practicals in the perception, recognition and production of a wide variety of human sound specimens. This type of educational orientation continued throughout the four years although it was most intensive during the first year. Such specialized intensive education in human speech had far-reaching impact on my perception, recognition and production of a broad array of sounds from different languages. For instance, prior to this education I was able to recognize and produce two bilabial sounds [b p]; however, after the orienta-
CHAPTER 1
11
tion the inventory doubled or even tripled into such sounds as [b b p p ]. This perceptual and productive enhancement in discriminative and articulatory skills afforded me far-reaching insight into other languages; perhaps, more importantly, it afforded me better insight into my own language repertoire. In the latter case, for instance, I was able to phonetically identify different bilabial plosives such as [b, p, p, p, p]4 some of which turned out to be phonologically significant, i.e., they triggered semantic differences between words. Also, equally important to broadening my range of sound perception, recognition and production, I nurtured an acute kinesthetic and proprioceptive sense for articulatory maneuvers of sound production. Stated differently, I began to feel the movements, positions and shapes of my tongue in the oral cavity, feel the tightness or relaxation of muscles and proprioceptively detect airflows, frictions and vibrations. Such kinesthetic and proprioceptive skills are extremely helpful for learning and teaching sounds especially in L2 situations. If the instructor does not bring such skills to the attention of the learners, they will pass unnoticed and learning fails. 1.2.5. Educational and Professional Challenges in the U.S.
I entered the U.S. as a refugee at the age of forty-three (43) with a family. I stayed unemployed for a year, desperately looking for employment to help my family survive and help myself pursue my academic profession. There was no room for me then to be choosy for employment. The English sayings “Beggars cannot be choosers” and “Don’t look a gift horse in the mouth” applied to me. Every job that I took in U.S. was only broadly within the realm of my academic and professional background, but was not in the core of my expertise so that I would immediately begin to be creative. As one will see later on, I went through three major professional and academic assignments each of which was a daunting challenge that occasionally pushed me to the brink of frustration. Nevertheless, I was a refugee and I was determined 4
The subscript dot indicates emphatic ( خمة ُّ ) ُُمفversions.
12
PRONUNCIATION IS IN THE BRAIN
to face head on any challenge. I succeeded in all three challenges and ended up gaining massive professional and academic experience from each one of them. Very humbly, I felt that the challenges I encountered during the first ten years in the U.S. granted me the value of several more doctorates. I felt that I was academically ‘reborn’ and professionally ‘baptized’ as an applied linguist that added depth to my skill as a teacher and enhanced my research fervor in both theoretical and applied linguistics and phonetics with pronunciation being at the core. 1.3. THE IMPACT OF MY LINGUISTIC/PROFESSIONAL BACKGROUND ON THE EVOLUTION OF AN APPROACH
No doubt, my linguistic upbringing coupled with my linguistic education provided the hidden dynamics that helped cultivate my passion for teaching pronunciation; in turn, it helped mold my approach to teaching pronunciation that is cognitive in essence, but requires further well-defined pedagogical principles and techniques for its implementation. 1.3.1. Impact of my Linguistic Background
The scenario of my linguistic background detailed in the above sections represents diversity in the types of languages to which I was exposed both as a child and adult. Briefly, I had different degrees of linguistic exposure to three major language families, namely Semitic (Assyrian and Arabic), Indo-European (English, Kurdish and German5) and Turkic (Turkmeni). A different perspective for my linguistic exposure is reflected in the nature of the particular experience I have had with each language. Was the experience in the form of acquisition very much akin to what children go through in their native language as subconscious internalization and full-time immersion in context/situation-embedded environment? Was it more in the form of an adult conscious learning without full-time immersion 5
I had some exposure to German through courses in my first degree cou-
pled with other courses at a German Institute in Baghdad. Most of all, my focus was on pronunciation.
CHAPTER 1
13
and with only minimum context/situation-embedded environment? Or was it a partial combination of the above two experiences in the sense that it was a full-time immersion in context/situation-embedded environment, but as an adult? The exposition of my cumulative linguistic experience in the previous sections indicates that I have been through all three types of linguistic environments. My early childhood experience with trilingualism in Kirkuk was of the first type: full-immersion as a child in context/situation-embedded environment which led to typical acquisition of the three languages involved. My exposure to English beginning with 5th grade throughout my high school on the basis of one hour per day could, at best, be described as minimal exposure to a language in almost context/situation-reduced environment using the most traditional approach to teaching known as ‘grammar-translation’. This approach uses translation into the native language (L1) as a medium to teach L2. It is a crude and utterly mechanical way of learning L2. My four-year college experience with English in Baghdad was relatively much better and more effective than the seven years in elementary and high school. There was more exposure to English, better contact with native speakers and more coordination between oral skills (listening and speaking) and literacy skills (reading and writing). However, the whole experience of school and college was a learning experience compared to my childhood trilingual acquisition. My experience with Kurdish and the four years of my graduate study in England had the characteristics of both acquisition and learning. Acquisition was facilitated by the linguistic environments that afforded me ample context/situation-embedded experiences in both languages. I also had plenty of opportunity in terms of time to intentionally learn what I wanted to. I vividly remember that when I was first introduced to ‘tones and intonation’ in human language I was totally lost because I had no idea, whatsoever, about these two aspects of human language. To familiarize myself with this aspect of human language, I imposed on myself a very strict schedule of attendance at the phonetic lab throughout one week and listened to a diversity of tones and intonation patterns. I also focused on stress and stress patterns in English as well as other languages. For English, I did two things to improve my ability in stress identification and placement. First, I listened
14
PRONUNCIATION IS IN THE BRAIN
very carefully to native speakers of English and carefully watched their body gestures. In one instance, I watched the economic correspondent of BBC television in 1970s (Dominick Harrod) whose straight black hair moved visibly down and up his forehead often synchronized with stressed syllables, especially those with primary stress. I rarely missed his appearances. For other languages, I used to listen to the English pronunciation of scores of foreign students attending Leeds University. Second, once I discovered my misplacement of stress, I focused all my attention on the correct pronunciation and repeated that forcefully as many times as needed and at times loudly so that I could hear my performance. I stopped when I felt that I had auditorily, kinesthetically and cognitively developed an acoustic image in my mind of the location of stress and its rendition in the targeted word and pattern of words. 1.3.2. Impact of my Teaching Career
Altogether, forty-nine (49) years of my life were spent in education. The first eleven (11) of which I spent in different Iraqi high schools teaching English. Four years after that were devoted to my graduate studies in England where I also taught Arabic on a part-time basis. After graduation, I returned to Iraq and taught linguistics and phonetics at both undergraduate and graduate levels for five years. I had to flee Iraq for political reasons and settle in the United States at the beginning of March 1981. In the United States, I first joined Loyola University as an adjunct professor teaching two courses in linguistics. In 1984, I was offered a full-time position of lecturer in English as a Second Language (ESL) with the City Colleges of Chicago (CCC) with which I stayed until 1990. During the years 1987–1990 I assumed the position of instructional advisor for ESL and Bilingual teachers within the CCC system. In August 1990, I resigned my position to join Northeastern Illinois University (NEIU) as associate professor in the Department of Teacher Education where I stayed until my retirement (effective January 2009). In the following sub-sections I will highlight the types of educational and professional challenges that I encountered and how each challenge contributed to building and shaping my cognitive and psychological approach to teaching pronunciation.
CHAPTER 1
15
1.3.2.1. Teaching Linguistics and Phonetics
The five years of university teaching of linguistics and phonetics in classroom situations in Baghdad helped put my theoretical knowledge into practice. While I was teaching my students, I was gaining real-life experience which helped me transform that into research papers that were accredited internationally. Nevertheless, I still did not have a vision of an approach to which I could claim ownership. The experience of twenty (20) years of teaching linguistics at Loyola University was, more or less, a continuation of what I did in Iraq except for the fact that one of the courses I was assigned focused on the ethnic and linguistic communities of Chicago. This course was quite challenging for me in nature since I was still a newcomer to Chicago. I devoted long hours to educating myself about the ethnic and linguistic composition of Chicago, which, in reality, represented almost all large urban areas of the United States. It was this course that introduced me to the ethnic and linguistic nature and composition of the overall society in my adopted country. 1.3.2.2. Teaching and Training ESL Teachers
The first three (3) of the six (6) years I spent with CCC were simply classroom teaching of English to L2 learners; however, the last three years were far more challenging both professionally and pedagogically. I was given the responsibility of training ESL teachers to assume the task of teaching English to a wide variety of speakers of other languages. At this juncture, I was professionally only a linguist who knew something about applied linguistics of which ESL is a discipline. The reader should be reminded that towards the end of 1980s ESL had become a primary discipline of applied linguistics with its own pedagogy in the form of theories and methodologies. Sensing that I was not prepared for the challenge, I decided to devote my whole time to review as much published literature in ESL as possible. I was so overwhelmed with sifting through the relevant materials, especially those most relevant to the daily teaching of ESL in classroom situations with all the needed teaching and learning strategies, I felt as if I was preparing for another doctorate. After approximately six months, I had built up some confidence in conducting lectures and running training workshops. I conduct-
16
PRONUNCIATION IS IN THE BRAIN
ed many of them in all of the seven colleges of CCC; in fact, I did more than what was expected of me. I began to receive kudos from the participants in those workshops, the deans of the ESL programs as well as the administrators in the central office of CCC. I was even offered an administrative position in the central office which I declined to have because I preferred to be in the classroom or field rather than in the office. After the end of those three years, I sensed that I had transformed myself into an ESL specialist. I began to present workshops and papers at local and state conferences as well as publish articles and papers. Personally, I felt that I had enhanced my academic reputation not just as a phonetician and theoretical linguist,6 but also as an applied one. Apparently, because of my performance and professional reputation I was approached by the College of Education at NEIU to apply for a position newly opened in the Bilingual/Bicultural Program (BLBC) which was part of the Department of Curriculum and Instruction later known as Department of Teacher Education. After the formalities of interviews, I was offered the position at the rank of associate professor to teach four courses—two bilinguals and the other two language arts (English)—in the regular program of Teacher Education. I believe I was offered the position not really because of my expertise in teaching bilingual education and language arts, but rather because of my academic credentials, my quality publications and my multilingual background in several languages. My academic assignment at NEIU turned out to be far more challenging than the ESL one with CCC simply because I had limited professional experience in both areas. Nevertheless, I welcomed the challenge simply because I still felt I was a refugee ready to confront any challenge in order to survive professionally and socially as a person with a family.
6
In 1988 I published my first book outside Iraq titled: The Sound System of
Modern Assyrian (Neo-Aramaic), Harrassowitz Verlag, Germany.
CHAPTER 1 1.3.2.3. Teaching Language Arts
17
In preparing teachers for English language in elementary and high schools in the United States, usually the course that covers its methods of teaching is named as ‘language arts’. Once I was assigned two language arts courses the first semester at NEIU, I had a quick look at the text and I was shocked at how remote the contents were from the modern linguistic perspective of teaching a language in an age when linguistics was permeating every study of every aspect of human language be it as L1 or L2. Unlike the ESL and bilingual language instruction, the teaching of English as a native language (L1), usually under the rubric of ‘language arts’, is so untouched by linguistics that, at times, the approach to its analysis is utterly surface-structured and orthography-based that many inaccurate practices and misconceptions riddle the approach. For instance, English vowels are taught as if they are five in number and occasionally six when is added. Such a statement, which is characteristic of phonics, is often a letter-based approach as opposed to a sound-based one. Linguistically such a statement is baseless because in all varieties of English there is a minimum range of 15–20 vowel phonemes (units) in the form of both simple vowels and diphthongs. Moreover, in the phonics-based approach consonant clusters (so called ‘blends’) may still be determined on the basis of letters rather than sounds. The teaching of spelling is so generically and obscurely conducted that there is hardly any distinction made between graphic spelling (based on written letters) and oral spelling (based on letter-names) two totally different language processes that require different methodologies for implementing. I was so frustrated with this assignment that I was almost about to ask the chair of my department to replace them with other courses. After further contemplation I thought it would be unwise on my part to ask for replacement since I was a ‘rookie’ professor who had to build up his professional and academic reputation in order to secure employment, promotion and tenure. In face of such a professional dilemma, I recalled fondly my successful experience with my ESL professional struggle. Once again I decided to confront the professional and academic challenge. I embarked on a massive reading and researching campaign to familiarize myself with all aspects of teaching language arts including the so-called theories, approaches, techniques and
18
PRONUNCIATION IS IN THE BRAIN
styles. Everywhere I thought I could infuse knowledge from linguistics into teaching language arts I did not hesitate to do so. I was relieved when the first semester came to an end and I managed to survive satisfactorily. With the next semester, I felt more prepared and at ease. As the years passed, I fell in love with the teaching of language arts because I succeeded in gearing it in the direction of modern linguistics in combination with up-to-date cognitive theories in language teaching pedagogy. At this juncture, I began to feel that I had accumulated enough knowledge and experience to formulate what could be called ‘my personal approach’. Theoretically, the pedagogy of this approach was premised on the works of four intellectual pillars of cognitive pedagogy to human language acquisition and learning. They were Noam Chomsky with his transformation-generative (TG) approach to the nature of human language and child language acquisition premised on the concept of language acquisition device (LAD); Howard Gardner with his multiple intelligence theory (MIT); Lev Vygotsky with his zone of proximal development (ZPD); and Jean Piaget with his child cognitive development (CCD). With the knowledge I acquired from these giants, I infused my own academic and professional experience of decades of teaching and researching to formulate what I had called at an early stage: ‘Multisensory Multicognitive Approach’ to teaching pronunciation (Odisho, 2003; 2007/a). Since then I have benefitted from further theoretical insight into child vs. adult acquisition and/or learning of sounds and sound systems. I have equally benefitted from the feedback that I received during the last ten years of the application of the approach in classroom situations with a large number of adult learners from a broad range of linguistic and cultural backgrounds. Gradually, I began to feel that there were so many weaknesses in the traditional approach to teaching ‘language arts’ that I had to unveil them publicly for other people in the field. My focus was primarily on areas related to my academic orientation and research interests. I began presenting and publishing research works with focus on the shaky and linguistically indefensible premises of ‘phonics’ as a tool to teach letter-sound correspondence and the ensuing vowel and consonant systems as well as the overall pronunciation and spelling. Phonics, for instance, confuses letters with sounds in both domains of conso-
CHAPTER 1
19
nants and vowels but more so in the latter. The alleged dichotomy of long vs. short vowels is the most striking example of confusion in the quality and quantity7 of the English vowel system. The manner in which phonics approaches the study of the sound system of English is utterly vulnerable from the perspective of modern linguistics. After several years of teaching language arts, I had so many inaccuracies and misconceptions to reveal that I had to write a book titled: A Linguistic approach to the application and teaching of the English alphabet (2004). 1.3.2.4. Teaching Bilingual Education
I have to emphasize the fact that I was hired by NEIU primarily to fill a position in the BLBC program. In other words, teaching bilingual/bicultural courses was my primary responsibility. Certainly, my multilingual background and my professional linguistic orientation coupled with my quality research work were of great help in handling some of the courses in the program. Nevertheless, there were areas in bilingual theories and bilingual education with which I was not familiar. Since the late 1960s bilingual/bicultural education and bilingualism, at large, have taken gigantic strides forward in theory and pedagogy. Very much like ESL, bilingual education has during the last few decades constituted a major domain in applied linguistics. In order to be ready to handle all aspects of bilingual/bicultural education I had to seriously acquaint myself with the latest innovations and applications in the field. The works of Jim Cummins, Stephen Krashen, Colin Baker, among many others, were thoroughly studied. As if I had not done enough self-education in the field of bilingual education, the Dean of the College of Education called for a special meeting of the BLBC faculty during which he assigned me the responsibility of developing two master level programs in bilingual education. This implied the need to develop 7
Because the words quality and quantity will keep recurring and acting
jointly in vowels, I have opted to blend them together in the form of ‘qualtity’ to be used where necessary.
20
PRONUNCIATION IS IN THE BRAIN
from scratch a minimum of six (6) graduate courses supported with the needed bibliography. Fortunately I was relieved of 50% of my teaching load for one semester. Throughout the semester, besides teaching my other two courses, I had exclusively devoted my time to this new professional assignment. The first two months of the semester were dedicated to surveying the literature on bilingualism and biculturalism in combination with pertinent insights from applied linguistics especially ESL. The focus during the remaining two months was on designing the framework of each course, identify its contents and supplement a core bibliography. The courses were approved by the College of Education and certified by the educational authorities in Springfield/Illinois. Very humbly stated, this achievement in bilingual education enhanced my professional status among my colleagues; besides, personally I felt it significantly added to my professional stature. I also began to feel that I stepped beyond my theoretical linguistic or phonetic skills into the realm of applied linguistics and the pedagogy of teaching language in a wide variety of classroom and real-life situations. Concurrently with my triumphant academic wrestling with ESL, language arts and bilingual/bicultural teaching I began a massive campaign of national and international academic presentations and publications. I published scores of papers in refereed journals and special volumes in honor of renowned scholars throughout the world. In 2002, I decided to put my long experience in teaching and research in several books prior to my retirement. Thus, I published: Techniques of teaching pronunciation in ESL bilingual and foreign language classes in 2003; A Linguistic and cognitive approach to the teaching of the English alphabet in 2004; Techniques of teaching comparative pronunciation: English-Arabic, in 2005; and Linguistic tips for Latino learners and teachers of English, in 2007. 1.4. CONCLUDING REMARKS
My early multilingualism seems to have been the motive behind my interest in languages, while learning Kurdish seems to have been the turning-point that guided me in the direction of linguistics. As for my early pre-doctorate writings, they narrowed down my general interest in linguistics focusing it on phonetics and pronunciation. Some light should be shed on this statement.
CHAPTER 1
21
1.4.1. Childhood Trilingualism Triggered Interest in Languages
There seems to be ample evidence that my trilingual/tricultural upbringing in Kirkuk was the flicker that lit the road for me in the direction of interest in languages and their teaching. Since the formal teaching of my native Assyrian language was very limited, I gradually developed interest in the last two years of high school in both Arabic and English beyond their formal classes. I finally ended up majoring in English for my first degree and graduated with honors. 1.4.2. Learning Kurdish Triggered Interest in Linguistics
My first degree in English in Baghdad does not seem to have triggered my future pursuit of self-education in linguistics in mid 1960s; instead, I give most credit to the five years of my stay in Sulaimaniya where I passionately pursued the learning of Kurdish. The first inkling of an infatuation with what I later found to be linguistics was born there. During the six years of teaching after that in Basra and then in Baghdad most of my readings were in the realm of human language with focus on the history of the English language, the etymology of words and some pronunciation issues. My pursuit of language studies in the direction of linguistics was utterly self-motivated. I published more than twenty (20) articles all in Arabic8 in different Iraqi newspapers and magazines followed by my translation of a book from English into Arabic titled Sounds and signs which was published by the Ministry of Culture in 1971. By this time, in Iraq, I was known as a ‘linguist’. Remember this was all prior to my formal graduate study in phonetics and linguistics in England.
8
Except for one article in English, published in the ‘Baghdad Observer’
dealing with Arabic loanwords in English via Spanish which retained their definite article < [ = >أل]. 25 years later I revisited the theme extensively and in depth with a publication in Zeitschrift für arabische Linguistik, Vol. 33, 1997.
22
PRONUNCIATION IS IN THE BRAIN
1.4.3. Graduate Study Immersed me in Phonetics and Linguistics
In 1971, after eleven years of applying for some sort of scholarship, I was granted a study leave, which was financially a third class sort of scholarship. The reason why it took this long to have even a third class privilege was simply due to political, ethnic and religious discrimination. Nevertheless, after gaining this scholarship I joined Leeds University determined to do my utmost. I joined Leeds because it was the only university in Britain that granted degrees in phonetics as a separate subject from linguistics. In my wildest imagination, I never ever thought that to major in phonetics would be so challenging, demanding and frustrating on top of the cultural shock of moving from Iraq and settling in Britain. The shock was so powerful in the first weeks that it pushed me to the verge of frustration; however, I put all my thoughts together and remembered all those eleven years of my struggle to continue my higher education as an ethnic Assyrian. I, then, decided to face the challenge and do everything possible to succeed. My work was the second doctoral project since the founding of the Department of Phonetics in 1948, the first being granted in 1970. I had two supervisors: a linguist and a phonetician. The latter was a young, very active and academically ambitious, faculty member in charge of the phonetics lab and the courses in acoustics, anatomy and physiology. She wanted me to be in line with her academic interests and ambitions. She pushed me almost beyond my scientific tolerance; fortunately, I managed to respond and began gradually to be more scienceoriented in my research. The bulk of my research was lab-based. I had to conduct a wide variety of experiments using different gadgets. The most challenging, and at times scary, experiments were the ones that involved inserting catheters or polyethylene tubes through my nostrils for two purposes. One catheter, with a photo-transducer at the end of it, was to be positioned just above the vocal folds to detect the opening and closing of the glottis and whether there was any vibration in the vocal folds. The other catheter had two small openings at its end to be inserted through the nasal passage and positioned between the larynx and pharynx.
CHAPTER 1
23
The purpose of this experiment was to monitor the intraoral pressure changes in the vocal tract. Such experiments were initially very intimidating and I did not have the support of anyone who had conducted them before. Nevertheless, the dream of a doctorate in phonetics by an Assyrian who suffered from discrimination and struggled for eleven years to have a go at it, made the trial less daunting. Inserting the catheters was a truly scary and disgusting ordeal. I suffered from severe coughing and nausea and my nose was almost dripping with mucous. During one year, I conducted fifteen successful attempts and accumulated more data than what I needed for my doctoral thesis. Being a fast writer, I finished my work, had my thesis of 420 pages typed and bound 6 months before the first day of my eligibility to submit. My graduate studies radically changed me as an educated person. It gave me academic depth, enhanced my tolerance for research and innovation and infused more scientific orientation in my approach to problem finding and problem solving. I became a passionate researcher with a broader horizontal perspective and deeper vertical vision. 1.4.4. Professional Challenges in the U.S.
The professional challenges in the U.S. in the form of teaching ESL, bilingual education and traditional language arts were not in nature what a phonetician was trained to do. Nevertheless, for a refugee professor facing those professional challenges was a fact of survival; indeed, I did not just survive, but I also gained a treasure trove of professional expertise in those three fields of teaching. As a researcher and professional, I developed a better understanding of my strengths and weaknesses, promoted more intimate connection with my students and cultivated richer teaching and learning styles. Perhaps more importantly, my knowledge base in linguistics and phonetics became more applied especially because of a better understanding of human language and speech as a multilingual person from childhood to adulthood. My greatest discovery was that any study of language, especially human speech, should begin with the evolutionary potential of the brain and its application through interactions in real-life situations. The knowledge and experience I accumulated throughout almost five decades of teaching led me
24
PRONUNCIATION IS IN THE BRAIN
to the pedagogy of teaching pronunciation that is multisensory and multicognitive in theory and in application.
CHAPTER 2: THE COGNITIVE BASE OF LANGUAGE 2.1. LANGUAGE: A SPECIES-SPECIFIC CODE OF COMMUNICATION
Language as a species-specific entity implies that only human beings are genetically born with a potential for language acquisition in its generative sense that Chomsky has intensively promoted. In this study, the generative nature of human language entails the potential to produce and comprehend instinct-free and stimulus-free infinite chains of meaningful structures using finite number of rules. In its underlying structure, the generative characteristic of human language also implies the potential for producing infinite number of meaningful structures from a very finite number of meaningless sound units (minimal structures) that are traditionally known as phonemes. This potential is part of the genetic makeup of human beings and it is ingrained in their brain. The code of language with human beings is radically different from that of other beings including apes and birds simply because with human beings the code is open-ended (infinite in the generation of meaning) whereas with non-humans it is closeended (finite in the generation of meaning). Thus linguistically, the word ‘language’ is exclusively used for human beings, while its use for other creatures is, at best, figurative; theirs is simply a finite code of communication to serve finite functions during their life span. This faculty of language is the result of millions of years of natural evolution of Homo sapiens away from chimpanzees, especially with regard to encephalization or “amount of brain mass to body mass”.1 In more strictly scientific terms, the brain 1
http://en. wikipedia.org/wiki/Encephalization.
25
26
PRONUNCIATION IS IN THE BRAIN
of a human being has the highest number of nerve cells (neurons) compared to all other animals; these neurons number in their hundreds of billions. A certain percentage of those billions constitutes the blueprint of the most sophisticated code of meaning-generation and meaning-decipherment: language. Although certain locations or areas in the brain such as “Broca’s expressive and Wernicke’s receptive areas” (Joseph, 2011) are more commonly associated with human language, language as it is understood nowadays is the function of the brain as a whole. Many scholars whose scientific works are in line with Darwin’s theory of evolution lump the gradual human brain encephalization with the gradual emergence of the faculty of language. This reciprocal relationship of evolution between the brain and language has been identified as a co-evolution phenomenon in nature (Pinker, 1994; Deacon, 1997; Christensen, 2001). However, the evolution of modern man goes beyond just the coevolution of brain and language. Developing an upright gait and freeing the hands (so-called bipedality, see Ackerman, 2006) seems to have preceded this particular co-evolution and, indeed prepared for it. Bipedality afforded man three extremely significant advantages. First, it freed the front limbs and transformed them into what became hands to be manipulated for more creative survival strategies. Second, it granted him a more efficient panoramic vision. Thus, vision wise, he did not have to stand on his hind legs, as many animals still do, to see beyond the level of his eyes when resting on four legs. Third, the upright gait gradually helped prepare the vocal tract for both physical survival and powerful potential for speech sounds generation. These three socalled sub-evolutions, jointly with others, enhanced his power to control his environment and interact with it more intimately and creatively. It is such combinations of sub-evolutions that have triggered the reciprocal evolution of human brain in the form of massive multiplication of its neurons. Figuratively speaking, the evolution of human brain capacity is similar to the gigantic increase in the capacity of the hard drive of modern computers from megabytes to Gigabytes to terabytes, etc., but in a more creative manner. Obviously, natural survival forced man to be as creative as possible. The most important strategy for survival was to become a social ‘animal’ and form a cooperative community with
CHAPTER 2
27
others. In turn, living in a community necessitated an efficient means of communication to plan and consolidate cooperation. Consequently, this need for communication led to the gradual emergence of a system to facilitate it. It is highly likely that in the beginning, the system was based on primitive hand and body gestures combined with different types of murmurs, grunts and noises which were more suprasegmental (long) in nature. No doubt, this was a very primitive and crude system. With time, and in order to grant the system more efficiency, it eventually evolved into well-defined segmental sounds to be coalesced together in conjunction with melody and rhythm to generate longer meaningful structures. In a sense, it was an evolution and progression from more generic to more specific, similar to the evolution of writing from longer and more general units to shorter and more specific (i.e. pictographic, ideographic, syllabic and alphabetic). In order for such meaningful messages to be comprehended by all members of the community, they had to be governed by rules: rules for sound combinations and others for word combination or syntax. It was these minimal sound units that were joined together with rules into larger meaningful units that paved the way to what we identify now as language. This gradual evolution of rule-governed systems and structures of meaning generation and comprehension could not have been possible without the qualitative and quantitative enhancement of the human brain. In turn, this enhancement through the power of selective evolution was genetically internalized. Such a genetic transformation granted newly-born babies the potential for activating, in appropriate social environments, a code of infinite production/comprehension of semantic messages that justified identifying language as species-specific. It is this potential that convinced Lenneberg to name his book Biological foundations of human language (1967). 2.2. LANGUAGE: A COGNITIVE-SOCIAL SYSTEM SUPERIMPOSED ON OTHER SYSTEMS
The gradual evolution of language required the generation of certain types of neurons and neural connections to assume responsibility for the internalization and activation of language. The more language distanced itself from grunts, murmurs and
28
PRONUNCIATION IS IN THE BRAIN
gestures in the direction of more readily generated and easily identified sounds (noises and voices), the greater the need became for manipulating many of the basic organs that help the other systems to assume additional functions. Along this line of thinking, this newly-evolving sociocognitive system had no specific organs assigned to it to initiate the needed acoustic and aerodynamic conditions for speech generation. Hence, in order for speech to be generated, it had to manipulate the existing organs in the human body, especially organs from the digestive and respiratory systems. From the evolutionary perspective this phenomenon of assigning double functions to certain organs has been recently named ‘exaptation’.2 A standard definition of the term is “an evolutionary process in which a given adaptation is first naturally selected for, and subsequently used by the organism for something other than its original, intended purpose” (Croom, 2003). Stated differently, the process stands for assigning an extra function to a system or organ in addition to its original function or purpose. The most commonly cited example for exaptation as an evolutionary process is the case of feathers in birds. Biologically, it has been stated that the original purpose of feathers in birds had been for the control of body temperature; it was only later that the evolutionary process adapted them for flying. In the case of speech, a typical example of exaptation is the radical modification in the shape of the vocal tract and the additional functions assigned to different digestive and respiratory organs. Below are some of the typical exaptation examples that facilitated the proper production of speech. 2.2.1. Vocal Tract Modification
The most significant outcome of exaptation accompanying language evolution is the modification of the tract between the lips and the vocal folds (cords). Presently, in human beings, instead of the tract having an approximately 45º curve, as it still is with many mammals as well as with newborn babies, in a mature 2
Etymologically, the word was coined from the prefix +
with the deletion of (ad).
CHAPTER 2
29
adult that tract is now almost at 90º curve and it is known as the vocal tract. A newly-born baby “has a mammalian larynx that can rise, enabling concurrent breathing and eating, and not until the age of three months are its speech organs ready for producing vowels” (Pinker 1994:354). It is because of the early 45º shape of the vocal tract and their ability to separate the breathing and eating tracts that babies can drink liquids while lying on their back whereas adults have serious difficulties doing that with a 90º tract. This 90º curve has resulted from the considerable lowering of the larynx, thus pulling the epiglottis away from the velum, and making contact between the epiglottis and velum no longer possible. It is because of this separation that swallowing is not possible while breathing for adults and vice versa. In modern man, the 90º vocal tract has two fairly distinct dimensions: a horizontal dimension beginning with the lips in front and the end of the oral cavity in the back and a vertical dimension beginning with the pharynx down to the vocal folds enclosed in the larynx. This evolutionary modification of the vocal tract has remarkably enhanced the articulatory, aerodynamic and acoustic suitability of the vocal tract for immensely diversified sound generation (in the larynx) and noise generation almost along the entire vocal tract. 2.2.2. Vocal Folds (Cords) Modes
The vocal folds (commonly known as cords) are small lip-like muscular tissues that are jointly and horizontally anchored to the thyroid cartilage (Adam’s Apple) in front and separately connected at the back to the arytenoid cartilages. Their natural function in the human body is to seal against the accidental entry of any object into the lungs while swallowing and to open them while breathing. With the evolution of speech, the vocal folds assumed an extremely important function, namely, voice generation through their different modes of vibration. In fact, vocal folds constitute the major source of human sound inventory enrichment. They are involved in the generation of both segmental (consonants and vowels) and suprasegmental (stress, tone and intonation) sounds. It is worth mentioning that the presence or absence of vocal folds vibration constitutes the richest and most economic principle for the generation of contrastive pairs of sounds. The voiced vs. voiceless
30
PRONUNCIATION IS IN THE BRAIN
dichotomy, which is a universal feature in human speech (Aitchison, 1996: 183), is the most economic and convenient distinctive feature to double the basic sounds in any language. Aesthetically, when voice was embellished with harmony and melody, humans crafted the most popular form of entertainment in human history—singing. 2.2.3. Tongue Functions and Maneuverability
The role of the tongue in all mammals is of prime importance for survival. “It has vital functions in feeding: It plays a major role in ingestion, as in licking, lapping, and browsing; and it moves food distally through the oral cavity from the incisors to the post-canines for chewing, and then to the pharynx for bolus formation and swallowing” (Hiiemae and Palmer, 2000). Nevertheless, the role of the tongue in speech is equally important. It is so important that in many cultures, such as English, Turkish, Assyrian, Hebrew, etc., the word ‘tongue’ is synonymous with ‘language’. In actual articulation, its configurations, distances from and approximations to the other passive or active articulators inside or outside the vocal tract (such as the lips) give birth to almost all common vowels in human languages. In the production of consonants, its role is no less significant. It is the primary articulator that determines the classification of the majority of unmarked (common) speech sounds in terms of place and manner of articulation. The versatile muscular structure of the tongue grants it so much plasticity that it can even endure antagonistic movements such as having the tip of the tongue placed at the incisors and alveolar ridge with a simultaneous gesture of a drastic push of its back/root posteriorly into the pharynx as is the case with the emphatics طُُص ُُض ُُظin Arabic. In short, evolution has given the tongue, the forefront organ of the digestive system, equally important functions in human speech production system, especially through its greater maneuverability within the vocal tract. 2.2.4. Lip Configurations
The essential biological function of the lips is to help food begin its journey along the digestive system. In speech production, they are the organs that generate most of what could be called
CHAPTER 2
31
the visible sounds or sound features for both vowels and consonants (Odisho, 2003). In the case of vowels, for instance, lipposition (spread, neutral, rounded), which is the only distinctly visible feature in vowel production, constitutes one of the three primary parameters in vowel formation and description. As for consonants, the lips, jointly or severally, are active in the formation of all bilabial e.g., [b, p, , ], labialdental [v, , f] and labialvelar [w] sounds. The labial/bilabial feature is not only felt by the speaker, but is also seen by the listener. 2.2.5. Cavities Resonance
The current 90º shaped vocal tract has augmented the sources of speech resonance, thus contributing considerably to the diversification of speech sounds and allowing different languages throughout the world to select their own sound inventories based on places and manners of articulation. Among such resonance cavities are the laryngeal, buccal (mouth), nasal and pharyngeal cavities. No doubt, all those cavities were biologically exclusively designed to serve primarily the respiratory and digestive systems. The laryngeal and buccal cavities have responded to evolutionary pressure leading to clear-cut exaptation; less so, for instance, with the nasal cavity. The structure and the anatomy of the nasal passage with the absence of a movable part (articulator), with many side chambers and heavy coating of mucous membrane render it a highly efficient airconditioning tract for respiration, but a very inefficient cavity for sound production and resonance. This explains why in human speech production mechanism, it is the oral passage that is predominant. Only a few nasals and nasalized sounds are attested in human languages, in general. Proprioceptively, it is quite difficult to feel nasal airflow, but one certainly can prove its presence when a nasal sound, such as [m] or [n], is sustained and then suddenly the nostrils are shut off with the fingers. 2.3 BRAIN ‘SPEAKING’ VIA RESPIRATORY AND DIGESTIVE SYSTEMS
It is clear from the above brief descriptions that the biological functions of certain parts and organs of the digestive and respiratory systems and the additional functions assigned to them throughout millions of years of evolution to generate language
32
PRONUNCIATION IS IN THE BRAIN
have all led to radical changes in the brain of human beings and other biological systems. The evolutionary growth in brain capacity and the concomitant evolutionary modifications in some organs have granted human beings much better qualifications for physical, cognitive and social survival and creativeness. In the forefront of such qualifications is the emergence of language as a unique sociocognitive privilege that sets human beings apart from other primates. When the brain fires its instructions for a message to be delivered, it is primarily the respiratory system and the upper end of the digestive system that facilitate the transformation of the cognitive message into an audible one through aerodynamics and acoustics. The lungs pump the air (the dynamic power) and send it through the necessary channels for vibration generation (voice) at the vocal folds level, if needed, combined with the appropriate degree of turbulence noise at different junctions along the vocal tract. It is this concomitant combination of systematic and rule-governed acoustic signals that impact the ear of the listener whose brain then decodes those signals according to pre-conceived code of a given language. Without such a preconceived code, the on-coming acoustic signals would be meaningless. It is just like listening to a language that one does not know. Usually in a face-to-face communication, speech is more readily transmitted and decoded between speakers and listeners because it is naturally accompanied by facial, hand and body gestures. This is yet another aspect of human speech where it out-performs the communication code of other primates. According to more up-to-date research, “it seems that gestures have a tight and perhaps special coupling with speech in present-day communication. In this way, gestures are not merely add-ons to language—they may actually be a fundamental part of it” (Kelly, et al, 2009). In other words, the authors conclude: “If you really want to make your point clear and readily understood, let your words and hands do the talking.”
CHAPTER 2 2.4. ECONOMY IN LANGUAGE
33
Some of the plain definitions of the term ‘economy’ read like the following: “Careful, thrifty management of resources; an orderly, functional arrangement of parts; an organized system”.3 According to such definitions human language turns out to be one of the most economic systems ever developed in nature. It is a system that uses finite minimal meaningless units to generate infinite meaningful multi-length units. It is this genius in human language that Martinet (1964) calls ‘double articulation’. The principle of double articulation is mathematically generative in nature. It first creates minimal units without meaning because if they were with meaning then any single unit or any combination of them would result in meaningful units. With such a system it would be too confusing for the brain to retain so many units with meaning. In the face of such vulnerability to confusion, the brain imposes rules that generate redundancy which, in turn, ‘provides the sufficient stimulus needed to acquire the system of language. Redundancy, therefore, provides the stimulus needed to acquire a complex grammar system’.4 One such powerful rule of redundancy is to deprive the minimal units of the bottom-most layer of language—the so-called phonemes—of meaning and start assigning meaning to units at a higher level generally known as morphemes. Other sets of rules are added to regulate the syntactical relationships of longer stretches of speech. It is this multi-layered reversed-pyramidal structure of language that becomes infinitely generative. It uses a finite set of rules to create finite systems which jointly are capable of generating infinite meaningful structures. These sets of rules serve two primary purposes: first, enable the speaker/listener to guess meaningful units from meaningless ones within a given language; second, enable the brain to internalize the finite rules subconsciously and interfere consciously when needed. Such a transformational production/reception of meaning is made possible by the very nature of the human brain which 3 4
http://www.thefreedictionary. com/economy.
http://en.wikipedia.org/wiki/ Redundancy_linguistics.
34
PRONUNCIATION IS IN THE BRAIN
has the capacity to function as the perfect encoder (generator of the code) in the case of the speaker and perfect decoder (decipherer of the code) in the case of the listener. In order for the speaker-listener flow of interaction to be triggered and continued, the embedded cognitive code in the brain of the speaker has to be able to reciprocate with the cognitive code in the brain of the listener and vice versa. Stated differently and plainly, they have to be speakers of the same language. Meaningful linguistic communication will be impeded if the underlying cognitive code is not the same or identical. Evolution through natural selection has empowered certain centers in the human brain to function as the decoding or encoding centers. Although the human brain is holistically responsible for human language, more specifically the former is identified as the reception center, commonly known as Broca’s area and the latter as the production center commonly known as the Wernicke’s area. Nature is simultaneously a creation and a creator. It rests on three pillars, namely biology, physics and mathematics. Biology creates, physics balances and mathematics computes. It is the change in the biology of the brain and other survival systems (respiratory and digestive) brought about by evolution that made language possible. Then, the rules of physics kicked in to balance the ensuing changes through computations that mathematics made available. This is nature at its best. Under the influence of evolution, creations in nature are susceptible to change. At times, with changes problems emerge. Nevertheless, nature is conscientious enough not to disturb its own balance by only initiating problems; its problems are followed by solutions. For example, when nature enabled the brain to grow larger and more powerful and capable of housing the species-specific language, it (nature) forced the respiratory and the digestive systems to adapt and facilitate the physical, aerodynamic and acoustic prerequisites of language generation. For instance, the 45º mammalian digestive/respiratory tract evolved into a 90º tract to accommodate the speech apparatus. The vocal folds had to acquire far more maneuverability than merely adducting the passage to the lungs when swallowing. They increased in elasticity to develop different and more sophisticated patterns of abduction/adduction and tension/relaxation. The relevant muscular apparatus and innervations had to be able to enhance the
CHAPTER 2
35
raising/lowering capability of the larynx to adjust the aerodynamics of voice and noise generation not just for speech production but also for different forms of singing. Another of the most characteristic features of the human brain is its hemispheric functional asymmetry which is completed through the so-called lateralization process according to which the left brain controls the right side functions of the body and vice versa. This lateralization process is usually completed by the age of puberty. If one seeks an explanation for this hemispheric specialization the answer seems to combine the principle of functional economy with increased specialization.5 The ability of human beings to develop their cognitive, physical, aerodynamic and acoustic potentials to accommodate language is the best gift uniquely endowed by nature upon them. The evolution of such a potential to express limitless meaning with minimum physical and mental effort is a superb example of the principle of economy in action when economy is defined as ‘the minimum amount of effort to achieve the maximum result’ (Vicentini, 2003). It is because of this dominating principle of economy that human beings “are able to produce about three words per second or one sound every tenth of a second on average and make only about one sound error per million sounds and one word error per million words” (Caplan, 1995). Within this grand system of economy in language, there are different sub-systems that contribute to building the unique system of speech generation. Without such sub-systems the burden of language on the brain would simply be too much to endure. To illustrate, the human speech apparatus can hypothetically generate infinite number of sounds; however, two questions immediately arise. First, does a human language need that many sounds to generate speech? Second, does the brain, which has millions of other biological functions to handle, like to burden itself with stocking up thousands of speech sounds? The answer to both questions is ‘no’. In the first instance, the generative de5
http://pandora.cii.wu.edu/vajda/ling201/test4materials/language_and
the_brain.html.
36
PRONUNCIATION IS IN THE BRAIN
sign of language requires only tens of sound units (phonemes) to be recycled again and again in a recurrent and generative manner. This highly economic design is a universal feature of human language without which it will lose its limitless creativeness. In the second instance, there is a salient tendency in language to manipulate sounds whose articulatory maneuvers are predominantly easy to produce and easy to perceive. The above two dynamics minimize ambiguity and enhance clarity. It is, therefore, not accidental that the majority of speech sounds are produced in the anterior half of the vocal tract in the form of labial, bilabial, inter-dental, dental, alveolar, etc., sounds; moreover, although in actual speech each sound unit may have a large number of different phonetic variations (allophones) for the same unit,6 the speaker does not recognize and store all those phonetic variations; rather, it only cognitively internalizes one abstraction for all the variations of a given sound unit to be known as a ‘phoneme’. It is only in the context of live speech that the phonemes mold themselves to the context in which they occur, thus yielding suitable contextual variants (allophones). Let us consider yet another example to demonstrate the principle of economy in language. When speaking, the brain fires instructions to the phonemes (abstractions) to construct a word, but the dynamic nature of speech, especially the mutual interactions of sounds in the flow of speech, creates different shades (allophonic variations) for each fired phoneme. It is only the listener who hears in phonemes because they go directly to his brain which stores only a finite inventory of phonemes. To state the same fact in more accurate linguistic terms, the speaker speaks phonetically while the listener recognizes phonemically (phonologically).
6
For instance, the /p/ phoneme may have several different realizations in
different contexts; it can occur as aspirated as in or unaspirated as in , with lip-rounding as in or with lip spreading as in .
CHAPTER 2 2.5. CONSCIOUS AND SUBCONSCIOUS BRAINS
37
In the preceding section, different pieces of evidence were knit together to highlight the economic premise of language as a reflection of the robust tendency towards economy in brain energy. It is the massive data storage capacity of the brain that renders language an open-ended system capable of producing and recognizing endless meaningful stretches of speech. Nevertheless, one of the most salient attributes of the brain that relates to human language and empowers it to be such a rich, creative and sublime medium of communication will be discussed in this section. This attribute is the twin-nature of human brain as conscious and subconscious (or unconscious). Although some nonhuman creatures may manifest a few hints of such a division of labor for the brain, in reality it is only a nominal division compared to that of human beings. It is, therefore, logical and substantiable to say that the human dichotomy of conscioussubconscious brains has been one of the main evolutionary developments that gradually emerged to manage, administer and execute millions of biological, social and cultural functions that humans have to successfully execute in order to survive healthily and rationally. One such fundamental function of the brain is language; indeed, without a highly sophisticated brain there would be no language. Furthermore, without the dichotomy of conscious and subconscious brains, language would be too much of a mental burden on the conscious brain to be able to handle it so smoothly and effortlessly. Here again there is a distinct division of labor between the two brains for the sake of economy in effort through securing maximum coordination and harmony in the execution of the myriad functions of the brain. At best, the conscious brain is responsible for approximately 10% of total brain functions, the rest being the responsibility of the subconscious brain. The conscious brain is responsible for any action that it decides to initiate. Once it decides on a certain action, most of the requirements for completion are automatically delegated to the subconscious brain. For instance, to participate in a marathon running competition is the responsibility of the conscious brain, but the physical preparation of the body (neuromuscular coordination and respiration) and the implementation of the actual running are executed by the subconscious brain. To cite another exam-
38
PRONUNCIATION IS IN THE BRAIN
ple, when engaged in an informal conversation with family members and friends, a sizeable percentage of the conversation is managed by the subconscious brain simply because one does not seriously engage in planning the contents of the conversation such as the selection of the needed vocabulary and monitoring the morphological and syntactical rules because the storage of the lexicon and the rules of grammar are in the subconscious. However, unlike this informal conversation, if a person engages in delivering a formal speech orally, the conscious brain assumes a greater role to cater for targeted contents, careful selection of the needed lexicon and greater adherence to formal grammatical rules of morphology and syntax. Because of the conscious role of the brain, the speaker might have more pauses, hesitations and repetitions than in a casual social conversation due to a covert conflict between the two brains. For all biological functions, the subconscious brain never sleeps because it has to monitor the heartbeat, respiration, blood circulation, secretion of the necessary glands, digestion and scores of other survival tasks. Most important of all, the subconscious brain is the sentinel of our normal continued existence especially inasmuch as language is concerned; it is the seat of the long-term memory unlike the conscious brain which usually handles sensory memory (for just a split second) and at best shortterm memory (for a few seconds). With every impression, experience and event that a person intends to maintain for the long term, it is the duty of the conscious brain to serve as a medium to transfer them to the subconscious brain followed by some reinforcement strategies.7 It is not unrealistic to say that when in the morning we prepare ourselves to go to work most of what we do is to obey our autopilot (subconscious brain) which will help us dress ourselves, have breakfast, start the car, leave the garage, close the garage door and drive. If all those habitual actions were to be handled by our conscious brain, by the time we were at work we would already be stressed out.
7
The retention in the long-term memory is the result of anatomical or bio-
chemical changes that occur in the brain (Tortora and Grabowski, 1996).
CHAPTER 2
39
In an earlier study of the strategies for teaching pronunciation, it was pointed out that language acquisition is a process of mental (cognitive) habit formation (Odisho, 2003). I used the term ‘cognitive’ to mean storing the linguistic habits in the brain and retrieving them instantaneously when needed. It is common knowledge to say that “whenever anything has been repeated a sufficient number of times to have become habitual, it becomes second nature, or rather a subconscious action” (Larson, 1912). We all as babies, toddlers or young children have slowly and gradually learned how to grasp things with our fingers, balance our bodies, walk, run, ride a bicycle and perform countless number of functions which through constant and systematic repetition have been transformed into automatic, effortless and subconscious survival functions. Such transformations are the greatest relief that nature has ever bestowed upon human beings. It is this transfer of the mental load from the conscious brain to the subconscious that helps the former avoid collapse under the pressure of too many requests and commands for action. Without mental habit formations, life would be too stressful and burdensome on the conscious brain. In our day-to-day life, when we say that life is becoming stressful we simply mean that we are engaging the conscious brain in responding to several problems simultaneously, knowing that the brain prefers to handle one problem at a time. Thus, normal survival of human beings without a powerful subconscious brain is virtually impossible. It is estimated that the subconscious brain has 40,000,000 nerve impulses per second, while the self-conscious brain fires 40 nerve impulses per second.8 The ease with which human beings use language as their most efficient social and cultural tool has only been possible through continuous mental transferences from the conscious to the subconscious. All the required basic constituent units of sounds, the rules that combine them in different formations and assign meaning to them are gradually transferred from the conscious brain to the subconscious especially in childhood. “It is estimated that the window for language is from birth to 10 years 8
http://brainforsuccess.com/howyourbrainwork.html.
40
PRONUNCIATION IS IN THE BRAIN
old. Note how quickly children learn a new language compared to adults. Moreover, unless the children learn the new language at a very early age, they will most likely have an accent in that language for life” (Carpenter, 2004). 2.6. CONCLUDING REMARKS
Evolution is the magic with which Mother Nature works its miracles. If nature imposes a new function or creates a new problem for its beings, it must, later on, devise a road map to solve the problem. In this instance, nature forced man to innovate a culture of survival. The foremost tool for such a culture of survival was language. However, language as a sophisticated generative system needs a large brain to store its building blocks, its rules and organize them to generate the semantic strings that speech is made of. So, the human brain gradually grew larger and larger to be the largest relative to body size of all brains of all animals. In order for language to be spoken, it had to have the necessary organs to initiate physical energy, to create the aerodynamic conditions that, in turn, afford the needed acoustic prerequisites that are translated into speech in the brain of the listener. This is how nature instructed the brain to negotiate with the respiratory and digestive systems to collaborate without compromising their biological functions. Both systems responded by imposing some modifications on their basic biological structures and functions. Thus, evolution gave human beings a modified physical tract that became capable of (almost) perfectly operating three systems (respiration, digestion, speech) instead of two. The evolution of human language is a typical example of “compromise between nature and nurture” (Bates, 1999). If the human brain is the most powerful computer ever to have evolved (even compared to our modern computers) it is because the brain is the result of millions of years of evolution whereas the latter are only the product of the last few decades. Stephen Hawking, in The Universe in a nutshell, says, “Present computers remain outstripped in computational power by the brain of a humble earthworm” (cited in Carpenter, 2004).
CHAPTER 3: LANGUAGE IN THE BRAIN OF A CHILD 3.1. LEARNING VS. ACQUISITION: CONCEPTUAL DIFFERENCES
In dealing with human mastery of language, the two terms ‘learning’ and ‘acquisition’ are the most frequently used to describe the process. Obviously, ‘learning’ is more generic in connotation than ‘acquisition’. The latter carries more technical denotation inasmuch as the phenomenon of human language internalization is concerned. The technical denotation has emerged with the rise in the popularity of modern linguistic studies, especially applied linguistics with specific focus on child language mastery (i.e. acquisition), bilingualism and second language learning in adulthood. In its specific sense, the term ‘acquisition’ has been used to identify the natural process through which a normal child masters its native spoken language perfectly and almost effortlessly compared to second language learning by an adult. There are very significant theoretical and applied distinctions between the two processes that are unfortunately known only to people who have been exposed to neurolinguistics, cognitive psychology and linguistics, per se. The majority of people use the term ‘learning’; some use the two terms interchangeably. The applied side of the distinction will be dealt with in the upcoming chapters. Theoretically, however, there are three rationales for shedding light on the distinction this early. First, the distinction between the two processes is related to the age of the learner and the nature of brain involvement in each case. Second, the implicit association of the two processes with the dichotomy of nature vs. nurture is significant in the context of this book. Third, although learning and acquisition are occasionally used as mutually exclusive, in the application of the pedagogy promoted here the two processes are handled as complementary in nature. 41
42
PRONUNCIATION IS IN THE BRAIN
3.2. THE BRAIN OF A CHILD AND LANGUAGE
Children are the best learners of language on earth simply because nature has endowed them with this gift. They begin internalizing their mother tongue not just in the postnatal period, but also before birth. Language begins burgeoning with the early formation of the fetus. One of the key findings of an ongoing research project by Canadian and Chinese researchers who are studying infant development suggests that while still in the womb, our brains learn speech patterns, laying the groundwork for language acquisition.1 “Although we think of human infants as being behaviorally immature at birth, there are many aspects in which the human brain is unusually mature at birth” (Clancy and Finlay, 2001). 3.2.1. Child Brain Formation and Maturation
In terms of the power and speed for information encoding, generation, transmission, reception and decoding, there is no match, to be identified yet, in nature and in modern technology to the human brain. If a comparison is made between the human brain and the latest most powerful computer in the world, research shows that “each neuron is comparable to a single computer. With one hundred billion neurons in the brain, we have the neural capacity of approximately one hundred billion networked computers each of which gets modified and updated on an everyday basis”.2 Since the focus of this book is on language, in general, and pronunciation, in particular, many of the biological and physiological aspects of the human brain fall outside the realm of the book. All attention will be centered on the function of the neurons (brain cells) and their synapses,3 which con1 2
http://abcnews. go.com/Technology/ story?id=97635&page=1.
Wesson, Neuroscience: http: //www.sciencemaster.com/columns/wes
son/wesson_part_03.ph 3
A synapse is a junction between two nerve cells, consisting of a minute
gap across which impulses pass by diffusion of a neurotransmitter. When a nerve impulse reaches the synapse at the end of a neuron, it cannot pass directly to the
next one; instead, it triggers the neuron to release a chemical neurotransmitter.
CHAPTER 3
43
stitute the neural network of the brain in crafting and operating human language. The neuron transmits information through its axons and receives information through its dendrites (França, 2006). The intricate manner in which this network enables a child to internalize the code of a language into which it is born with so much ease is simply the prime miracle of Mother Nature. It is startling to realize how much of fundamental brain morphology and organization is already laid down by the end of the first trimester even before many mothers realize that they are pregnant (Clancy and Finlay, 2001). At birth, each neuron in the cerebral cortex has approximately 2,500 synapses.4 By the time an infant is two or three years old, the number of synapses is approximately 15,000 synapses per neuron (Gopnick, et al., 1999). At its peak, the cerebral cortex of a healthy toddler may create 2 million synapses per second.5 In general, young children may have more synapses than they will ever need. The development of synapses occurs at an astounding rate during children’s early years in response to childhood experiences. In fact, the sudden increase in synapses amounts to more than what an adult brain usually has. The significant question at this juncture is: why is there such a gigantic surge in synapse formation or synaptogenesis? The obvious answer lies in the immense environmental pressure on the brain brought about by the physical and mental growth of a child. In fact, the abundance in synapse formation is a safety and precautionary measure on the part of nature to successfully carry a child through the most significant phase in human life. Once the formative years carry the child along the path to The neurotransmitter drifts across the gap between the two neurons. On reaching the other side, it fits into a tailor-made receptor on the surface of the target neuron, like a key in a lock. This docking process converts the chemical signal
back into an electrical nerve impulse. (http://www.sciencemuseum.org.uk/Who AmI/FindOutMore/Yourbrain/Howdodrugsaffectyourbrain/Whatsasynapse.aspx). 4 5
http://faculty.washington.edu/chudler/plast. html.
https://www.childwelfare. Gov /pubs /issue_briefs/brain_development
/how.cfm, 2009
44
PRONUNCIATION IS IN THE BRAIN
maturity and enable it to successfully and creatively engage in normal physical and cognitive survival, a visible sudden drop in the number of synapses, called synapses pruning, takes place. Human and animal studies show that the mammalian brain undergoes massive synaptic pruning during childhood, removing about half of the synapses before puberty (Chechik, et al 1999). It is, therefore, “fair to say that the infant arrives in the world with a nervous system whose working components are in place and organized. All cells are generated, major incoming sensory pathways are in place and all have already gone through a period of refinement of their total number of cells, connections, and topographic organization” (Clancy and Finlay, 2001). There is ample scientific evidence that the human brain is primed right from its fetal phase throughout adulthood to be the unique and powerful organ to construct, operate and manage the perfect communication system ever known in history—human language. The relationship between the human brain and human language typifies a perfect balance between nature and nurture: the former creates and demands while the latter reacts and heals. As for building cognitive habits in the brain, it is worth mentioning that the more frequently the neuron connections are used, the more they retain information and the stronger they become.6 It is in this manner that children execute the process of natural acquisition of language. 3.2.2. Formative Months and Years of Mother Tongue
Research indicates that infants are able to respond to sound 10 weeks before birth through bone conduction (Shiver, 2001). Nevertheless, the actual acquisition7 of language begins with birth when the infant is not only exposed to its mother’s speech 6
http://www.brighthubeducation.com/infant-development-learning/
35203-sensory-stimulation-for-infant-brain-development/ 7
‘Acquisition’ is used here in its technical denotation as a natural, effort-
less and subconscious process of language internalization. It is collectively, but not absolutely, opposed to the process of ‘learning’ which tends to be effortful and conscious.
CHAPTER 3
45
but also to the speech of other members of the family and to all natural sounds in its environment. The period from birth throughout childhood and early adolescence is the prime time for language acquisition. The brain, through billions of neurons and trillions of synapses, is ready to assimilate any structures of language beginning with its minimal sounds through larger combinations of them in the form of words, phrases, clauses and sentences leading to discourse. This burgeoning of language acquisition neatly coincides with the mushrooming of the synapses in the brain that become the pathways to the neurons for a twoway communication of speech production and perception. Clancy and Finlay (2001) neatly summarize child language evolution during the first thirty (30) months as follows: “First, the period between 8-10 months is a behavioral watershed, characterized by marked changes and reorganizations in many different domains including speech perception and production, memory and categorization, imitation, joint reference and intentional communication, and of course word comprehension. It seemed plausible that this set of changes (which are correlated within individual children) might be related to patterns of connectivity and brain metabolism. Second, the period between 16 and 30 months encases a series of sharp non-linear increases in expressive language, including exponential increases in both vocabulary and grammar. A link seemed possible between this series of behavioral bursts and a marked increase in synaptic density and brain metabolism that was estimated to take place around the same time.”
Once the base of language is consolidated in the brain of a child, the massive synaptic activity slows down in volume and speed. It is at this stage the synaptic pruning phase kicks in. A huge number of neurons perish and even greater numbers of synapses are simply eliminated; it is truly a “use-it-or-lose-it” situation (Shiver, 2001). Other synapses are redirected to handle additional cognitive functions. For instance, not all millions or billions of synapses that were engaged in internalizing the sound system (phonology) of the native language are needed because the phonology and the articulatory maneuvers required for its actual production/perception are already perfectly internalized
46
PRONUNCIATION IS IN THE BRAIN
to the extent of being fully cognitively habitual or subconscious. It is also quite likely that a sizeable percentage of the neurons and synapses are assigned different linguistic functions such as focusing on enriching the lexicon (vocabulary) and upgrading the morphological and syntactical rules. For example, it is quite common for very young native speakers of English to demonstrate excellent mastery of their sound system, but fail to correctly conjugate past tense and past participle of irregular verbs such as . For instance, instead of conjugating as and , they apply the dominant rule of ending resulting in in both cases. In light of such grammatical flaws, it is likely that some of the phonologyoriented synapses in the brain of those young learners will be redirected to engage in the mastery and refinement of other grammatical rules. This neural and synaptic scenario in terms of peak activity, slow down and redundancy is very reminiscent of the working conditions in a factory. When there is high demand for goods the laborers work overtime, when the demand decreases a slowdown follows forcing the management to lay off some laborers and/or reassign others to different jobs. In this game of brain and language, the brain is the manager, the neurons and synapses are the laborers and the requirement of language acquisition is the demand. 3.3. COGNITIVE TRANSITION IN SOUND PERCEPTION AND PRODUCTION
Once the child is born into the world, it has to transition from the world of hearing to the world of listening as the former is a “sense while the latter is a skill… indeed, listening can be thought of as applying meaning to sound, allowing the brain to organize… listening is where hearing meets brain… listening to language is uniquely human” (Beck, 2011). In the process of language acquisition, the child has to hone its listening skills to be able to discriminate one sound from the other. Discriminative listening is a cognitive process through which sounds are internalized into the long term memory to be subconsciously and effortlessly identified, selected and produced when needed. Neurolinguistic literature is fraught with evidence substantiating the fact that very young infants discriminate not only the segmental contrasts of their native language, but many
CHAPTER 3
47
nonnative contrasts as well (Best, 1991). This implies that the phonetic inventory of infants and young children in their native language is richer than when they become adults. Best’s (2002) findings further suggest that infants begin life with languageuniversal abilities for discriminating phonetic sounds that do not exist in their language environment. However, these abilities gradually decline in cross-language speech perception between infancy and adulthood; in fact, adult speech perceptual ability is more limited, reflecting discrimination of only those contrasts which are phonemic in the listener’s native language (Werker and Tees, 2002). It might be that the difficulties adults have will reflect a permanent, absolute loss of sensory-neural sensitivity to the acoustic properties of those contrasts discriminating between the pair members of many nonnative contrasts, or their linguistic properties (Eimas, 1978). True, the loss in sensory-neural sensitivity is a serious one, but is neither absolute nor permanent; there is always room for neuroplasticity in early adulthood. Nevertheless, the gradual loss in sensory-neural sensitivity is a natural phenomenon and seems to correspond neatly with the synaptic pruning that the infant brain undergoes in the early years. The synaptic pruning is an attempt on the part of the brain to internally reorganize its neural functions and avoid any redundancies in the whole process of phonology acquisition. Thus, when the baby hears people speak a certain language, the brain strengthens connections for the sounds of that language; consequently, the connections for the sounds of other languages become weaker and may eventually wither.8 When the brain of the infant senses that it has an almost complete mastery over the sound system (phonology) and communication is sustained almost flawlessly with parents and other siblings, it reinforces the system and locks it in the subconscious brain and indirectly excludes other sounds that are not experienced and do not recur in his native language. Linguistically, this transition from discriminating a large number of sounds in childhood to recognizing only a limited core in adulthood is 8
http://www.fcs.uga.edu/outreach.
48
PRONUNCIATION IS IN THE BRAIN
known as the transition from phonetics to phonology or a transition from physical entities to cognitive ones. Stated differently, it is a transition from the world of concrete speech sounds in the environment to abstract sounds in the brain. Once this concreteto-abstract transformation takes place the child gradually becomes psycholinguistically insensitive or, at times, deaf9 to sounds that are absent in his language or occur only allophonically (as contextual variations of a phoneme). This transition is the foremost reason why adults learning a second language (L2) usually fail to accurately pronounce sounds that are alien to the native phonology of their language or dominant dialect; it is this failure that we identify as accent in the pronunciation of a given language or even a given dialect within a language. 3.3.1. Transition from Phonetics to Phonology
In linguistic studies, phonetics is the study of human capabilities for speech sounds production: their places and manners of production, their voicing and voicelessness, among others. Human speech organs are capable of generating thousands of sounds, some of which differ from each other in only minuscule and insignificant features (Catford, 1977). However, this profusion in physical sound generation is incompatible with the dominating tendency in the brain towards economy in both physical and mental effort. Stated differently, the rule of the brain is to store finite sound units in memory, but generate infinite stretches of meaning. In order to attain this goal, the brain applies a sound compression technique in the form of the process of abstraction. In other words, the brain, right from infancy, begins the process of compressing many physically and/or perceptually related sounds under one cognitive unit stored in the subconscious brain which is traditionally known as a phoneme or a sound unit. 9
When in England as a graduate student, I had an Arab friend from Syria
who used to pronounce = [ʤʤ] as []. When I brought that to his attention, he said: “I said [ I didn’t say [ ”. He absolutely did not realize
that he was still pronouncing [ʤʤ] as []. Indeed, he was deaf to this sound. It took me a long time to convince him that he had to pronounce with [ʤ] not [].
CHAPTER 3
49
As pointed out earlier, each phoneme is the abstraction of many physical sounds identified as allophones. To illustrate, the sounds of
in the following words: , , , and are different from each other. Generally speaking, they are simply phonetically described as follows: aspirated; unaspirated or deaspirated; aspirated with lip-rounding; and aspirated with lip-spreading. Each of those four represents an allophone of the phoneme /p/. In reality, the human brain is not simply satisfied with this process of abstraction; abstraction is coupled with a process of selection and elimination. The brain of a typical native speaker of a given language selects a number of sounds10—usually in tens—from a huge inventory of possible natural sounds to form the inventory of his L1 sound system. It is this limited inventory of units that is known as phonology. Once this transition is completed in the brain of a child, two major linguistic consequences are distinctly noticed. First, the natural phonetic gift of a child in being able to perceptually discriminate a wide variety of sounds including many that are not part of the inventory of his native language (L1) gradually begins to subside with age. Accordingly, his skill in perceiving, discriminating, let alone producing non-native sounds erodes progressively. Naturally, the erosion continues with older age. It is worthwhile emphasizing that the erosion is not the result of loss in the plasticity of the vocal organ as much as it is a consequence of the loss of sensitivity (neural plasticity) to sounds and sound phenomena outside the periphery of the native language phonology. Second, barring any speech deficiency, the mastery over the phonology of L1 becomes perfect, but the perfection will be at the expense of non-L1 sounds perception/production. Such biases facilitate rapid perception of native speech, but seriously impede perception of speech sounds in a language other than the native one (Port, 2007). It is in the above transitions in the skill of sound perception and production that the root cause of the linguistic phenomenon of accent is buried. In sum, it is the 10
The majority of languages throughout the world have an inventory
range between 20–50 units.
50
PRONUNCIATION IS IN THE BRAIN
phonology of the native language (L1) that imposes phonetic limitations on perceiving and producing L2 sounds. 3.3.2. The Brain as the Commander-in-Chief of Language Acquisition: The Cognitive Roots of Linguistic Accent
Detailed discussion of the applied implications of accent in language will be dealt with in a specific chapter. At this juncture, since the focus of this book is on the cognitive pedagogy of teaching pronunciation, only the cognitive roots of accent will be tackled in this section. In a nutshell, accent is a psycholinguistic problem; it is a cognitive problem whose solution should be primarily cognitive as well. It emerges as a result of the brain trying to immaculately internalize the phonology of the native language and transfer it from the conscious brain to the subconscious and render the process of pronunciation as effortless as possible with the purpose of alleviating the stress on the brain if all language interactions are conscious. It seems that the brain of an infant assigns a massive contingent of neurons and synapses to perfect the process of internalizing the phonology. As if during the formative years, the brain keeps communicating mutely with the infant and negotiating the conditions for accomplishing the mission. The mute dialogue between the infant and his brain might go on in the following manner: Brain: Do you love your mother tongue? Infant: Yes, I do; indeed, I adore it. Brain: Do you want to master its pronunciation perfectly? Infant: Yes, I love to. Brain: I will help you with it. Infant: Oh! Thank you! Thank you! Brain: Are you planning to learn a second language as an adult? Infant: May be. Brain: It may be difficult for me to help you master its phonology as perfectly as your L1. Infant: What do you mean?
CHAPTER 3
51
Brain: You may have an accent. Infant: Why? Brain: You have either to start much earlier or you have to have a gift as well as a creative instructor to help you reduce your accent. Infant: O.K. I have no better choice.
The gist of the above assumed and tacit dialogue between the brain and the self implies that once a person approves the conditions of the brain for securing perfection in the phonology of L1 at the expense of the phonology of L2, he/she implicitly approves the likelihood of emergence of accent in L2. The argument that accent is a cognitive problem and its solution should also be cognitive-oriented with the assistance of as many sensory modalities (channels) as possible is premised on three cognitive strategies on the part of the Commander-inChief, the brain. First, it is the economic principle of employing almost negligible cognitive effort to generate the maximum product by internalizing a finite inventory of sound units and storing them in the subconscious for spontaneous and effortless retrieval. Second, store the sound system in a safe zone as if the brain intends to grant it immunity against interferences from non-native sound systems and other outside interferences. Third, redirect the redundant neurons and synapses after successful internalization of the sound system to reinforce other linguistic skills including the lexical, morphological, syntactical and stylistic. It is the above argument that has inspired the title of this book: ‘Pronunciation is in your Brain, not in your Mouth’. If one intends to ‘eliminate’, or avoid an acute accent the only option to achieve that is to negotiate with the brain for ‘entry visas’ to accommodate additional ‘alien’ sounds. Children are always granted such ‘visas’ whenever they have ample linguistic exposure to an ‘alien’ language. Adults, unfortunately, have to try hard and even wait for longer time to secure a ‘visa’. Oftentimes, adults are granted the ‘visa’, but with constraints attached to it—you have to have a certain degree of accent. It is almost a universal tradition that a child born and raised in a certain country secures the citizenship of that country. By the same to-
52
PRONUNCIATION IS IN THE BRAIN
ken a child born and raised in a certain language becomes the citizen of that language. It is a natural privilege of children. 3.4. FOSSILIZATION OR PSYCHOLINGUISTIC INSENSITIVITY
There is no doubt, whatsoever, that children show more adeptness in language mastery than adults do, especially in the area of pronunciation. As has been discussed earlier, this distinction is attributed to the natural age-bound neuronal and synaptic proliferation and activity with regard to native language internalization in early childhood followed by a gradual slowdown. With age, adults begin to progressively lose their aptitude for the automatic and subconscious internalization of pronunciation. Consequently, the process of mastering the pronunciation of a second language becomes increasingly more conscious, mechanical and effortful. In the available literature, failure of adults to further improve their mastery of L2 pronunciation beyond a certain limit has been identified by some researchers as fossilization. (Selinker, 1972).11 According to the cognitive pedagogy adopted here, the use of the term ‘fossilization’ to identify the slowness and imperfection of adults’ mastery of L2 pronunciation is rejected for several reasons. First, the choice of the term ‘fossilization’ is unfortunate since it implies excessive rigidity to be associated with the enormous potential plasticity of the human brain. Second, the neuronal and synaptic activities which are behind the internalization of L1 sound system do not seem to be exhausted because many adults, young and middle-aged, can still secure a highly satisfactory proficiency in pronunciation. Some investigators (Johnson and Newport, 1989) conclude that there is no single moment when the window of linguistic opportunity slams shut.12 Third, the inefficient methods of teaching pronunciation 11
The definition of the word ‘fossilization’ is “the process of being turned
to stone”. 12
The author significantly improved his English pronunciation at the age
of thirty-three (33) by eliminating many phonetic and phonological residues left
over in his English from his Aramaic and Arabic languages. He also added a wide variety of new sounds to his phonetic inventory. It was a very demanding
CHAPTER 3
53
that rely exclusively on the auditory modality (i.e. repeat-afterme technique) instead of an approach that is multisensory and multicognitive in essence and secures significant results (Odisho, 2003). Fourth, the extent of exposure to L2 and the context of exposure13 are of considerable importance in reducing the acuteness of L2 accent. Finally, if self-motivation on the part of the learner is added to the above four factors the accent can be considerably reduced and, occasionally, minimized. The rejection is justified based on the fact that systematic multisensory and multicognitive orientation helps all learners, regardless of age and aptitude for pronunciation, to improve their skills, though to different degrees, in the acquisition/learning of L2 pronunciation. Using a combination of diversified multisensory and multicognitive techniques and exercises, the learning process can continue, albeit slowly, but it will hardly cease completely to the extent claimed by fossilization. In an earlier work (Odisho, 2003), this slowness or intransigence of adults in the acquisition/learning of L2 pronunciation was known as psycholinguistic deafness which, unlike fossilization, does not imply total cessation of learning; rather, it keeps the doors of acquisition/learning of L2 pronunciation open depending on the rationales mentioned above. Maybe the choice of ‘deafness’ was too loaded a term; nevertheless, what I implicitly meant was a transitional decline in efficiency or sensitivity to speech sound perception/production, but not a total cessation. Stated differently, with adulthood, speakers of any language gradually develop a certain degree of psycholinguistic insensitivity to the perception, hence production, of speech sounds outside the phonological realm (inventory) of their native language (L1). It is, however, noteworthy that the distinction between the adeptness of children and adults to language acquisition is priand time-consuming task, but it was doable for two reasons: first, it was full
immersion in an all-English speaking environment; second, it was an environment of academic specialization that imposed on the brain to react to the specific situation. 13
Is it classroom exposure only or is it combined with sufficient social in-
termingling in the community of the targeted language?
54
PRONUNCIATION IS IN THE BRAIN
marily confined to the linguistic domain of pronunciation and not necessarily to other domains, such as morphology, syntax and lexicon in which adults may be equally adept or even more adept than children in many cases. It is, therefore, logical to conclude that for a child there is no easy or difficult language or languages to acquire, especially their pronunciation; the difficulty in mastering the pronunciation of second language tends to be a trait of adults. 3.5. THERE IS ROOM IN THE HUMAN BRAIN FOR MORE THAN ONE LANGUAGE
Bilingualism is a normal natural phenomenon in human civilization. There are as many bilinguals around the world as there are monolinguals; indeed, there is hardly any country that does not exhibit a certain degree of bilingual communication. In the long history of human civilization, bilingualism has never been a marginal or accidental linguistic phenomenon that emerges sporadically and intermittently here and there. Conversely, bilingualism is a constant component of the overall structure of human civilization; it automatically emerges when two or more language communities or speakers come into contact. Bilingualism is an easily justifiable normal and natural sociolinguistic and psycholinguistic phenomenon. The sociolinguistic naturalness of bilingualism is substantiated by its pervasiveness throughout all linguistic communities; likewise, psycholinguistically, bilingualism is a normal and natural phenomenon because human beings, especially the young, internalize it readily, implying that the human brain is endowed with enough cognitive potential to absorb more than one language (Odisho, 2002). The brain of a child with billions of virgin neurons and synapses is a massive generator of cognition, imagination, creation and innovation. It is, therefore, quite natural for a child to automatically acquire two, or even three languages if it has ample access and exposure to them. If the exposure is balanced then naturally the competency in the two languages will be balanced as well. In fact, each language may be handled as a system in its own right and the competency will greatly resemble that of monolingual child in either language (de Houwer, 1990). Stated differently, if the child is exposed to two languages from a very early age, he will essentially grow as if there were two
CHAPTER 3
55
monolinguals housed in one brain.14 The linguistic competency with many bilingual children can be so high that enables them to transition instantaneously and subconsciously from one language to the other. Obviously, some unintended switching and mixing between the two languages is to be expected. The human brain is powerful enough for more than one language, but in the absence of a bilingual or multilingual environment, the child will naturally grow up as an immaculate monolingual. 3.6. NARROWING DOWN THE BROAD DEFINITION OF ACCENT
The broadest definition of linguistic accent is that it is a deviation from a given norm of pronunciation acceptable to a group of speakers. With this broad definition, it is irrational and impractical to design one approach to remedy all types of deviations from the given standard or acceptable model of pronunciation. Often, the term ‘accent’ carries a negative connotation since it is usually associated with comments such as: ‘He speaks with an accent’; ‘He has a strange accent’; ‘I can’t understand him because of his very heavy accent’. With today’s linguistic refinement in the description and assessment of human languages, the term is still often used generically without much specificity. The only distinction that has relatively been made somewhat clearer is between a ‘dialect’ and an ‘accent’. The former is usually used to refer to a combination of grammatical, lexical and pronunciation differences, whereas the latter is essentially confined to pronunciation differences. In light of the above information, two points should be taken into consideration. First, highlight all identifications and descriptions of the term ‘accent’ and select the one that will be the focus of this study. Second, identify the approach that will be used to teach pronunciation, in general, and accent remediation in L2, in particular. The response to the first point is covered by the following two dichotomies: intralanguage accent vs. interlanguage accent and phonetic accent vs. phonological accent which will be the focus of Chapter 4. The approach will be touched 14
Petitto, quoted in [email protected].
56
PRONUNCIATION IS IN THE BRAIN
upon lightly in this chapter, but details will be dwelt upon thoroughly in some of the remaining chapters. 3.7. IMPLICATIONS FOR UNDERSTANDING THE COGNITIVE NATURE OF ACCENT
There are both theoretical implications and practical applications for the cognitive identification of the underlying causes of accent especially with adult L2 learners in contrast to children internalizing their L1 or even L2.15 The practical applications will be elaborated on elsewhere in this book; thus, the focus here is on the theoretical implications that should outline the pedagogical roadmap for application. Foremost among those theoretical implications are the following. 1) Develop a cognitive approach that reflects the latest understanding of distinctions between the nature of human language acquisition and language learning. 2) Envisage the methodology or methodologies and identify the teaching and learning techniques that take the above distinctions into consideration and implement them. 3) Be aware of the phonetic vs. phonological differences in pronunciation and how the latter should receive the priority as they cause semantic (meaning) confusion. 4) Determine the proficiency level of pronunciation that is targeted; is it native, near-native or just acceptable? 5) Determine the main functional goals and objectives of teaching pronunciation.
The striking disparity between children and adults is that children internalize language through a natural process identified in applied linguistics as acquisition. Cognitively, they are primed to master their L1 or even beyond that. Their brain is ready to masterfully absorb any linguistic materials to which they are 15
In fact, being fully exposed to and immersed in three different languages
as a child, I grew up trilingual with native oral mastery of all three of them (Assyrian, Arabic and Turkmeni).
CHAPTER 3
57
amply exposed in authentic social contexts and situations. Collectively, all those three conditions of cognitive readiness, ample exposure and authentic context lead to the process of acquisition which is a teacherless process in the formal meaning of teaching. All that a normal child needs is the habitat of a family and the community. Children acquire the core of the native language, especially pronunciation, from mere exposure to the community around them which is also known as ‘incidental learning’. “Incidental learning through overhearing occurs when children listen to speech not directly addressed to them, yet they learn from it. Amazingly, very young children learn approximately 90% of the information they acquire incidentally” (Beck, 2011). 3.8. CONCLUDING REMARKS
In chapter 2, the focus was on the interaction between brain and nature and the power of the latter in the evolution of the former. This chapter had a triangular emphasis on nature, brain and social environment in that language is a product of all three and children are a most fortunate beneficiary of the combination of the three forces. Language acquisition for a child is identical with the acquisition of walking; they both evolve, ceteris paribus, naturally and are seamless in their function. This is why children do not have an accent when thoroughly exposed to a language besides their own. Accent is a linguistic phenomenon that is associated with adults. Naturally, no adult wants to speak an L2 with an accent; nevertheless, there will often be a ‘sliver’ of accent. How ‘thick’ or ‘thin’ the sliver would be depends on many conditions. In a social-linguistic context, a thinner sliver is better than a thick one. It is for this reason that the next chapter emphasizes the distinction between a phonological accent and a phonetic one, the former being the thick (implying a thick accent), and the latter being the thin. Functionally and professionally, priority should be given to thinning (i.e., reducing) the phonological accent first then handling the phonetic one.
CHAPTER 4: LINGUISTIC ACCENT: DEFINITION, CLASSIFICATION AND DEMONSTRATION 4.1. INTRODUCTORY REMARKS
As mentioned earlier on with regard to the difference between dialect and accent, the former is usually used to refer to a combination of grammatical, lexical and pronunciation differences, whereas the latter is essentially confined to pronunciation differences. In fact, in ESL classroom situations or real-life interactions, one is able to identify the accent of an L2 speaker not necessarily based exclusively on pronunciation inaccuracies and deviations from acceptable targeted standard pronunciation. In many instances, there are grammatical, morphological and lexical hints that trigger an accent in pronunciation. Take, for instance, the case of some English past tense and past participle formations of verbs which end with suffixes that are in the form of voiced or voiceless consonant clusters. Often such clusters are broken up or reduced in one way or another by Hispanic learners of English because they are alien to their native phonology. For example, the past tense of is and are pronounced as ]. Traditionally, Hispanics try to overcome the problem in one of two ways both of which are, unfortunately, wrong: 1) Reduce the cluster by dropping one of its elements, usually the latter. This means that instead of pronouncing [] and [] correctly, they would simply reduce them to [] and []. 2) Insert a vowel between the two elements of the cluster, break up the cluster and cause a reshuffle in the syllabic structure of the word. To demonstrate, instead of pronouncing the word or as [] and [], they are rendered roughly as [] and []. Undoubtedly, this is a grammatical problem but its roots are in pronunciation. In the totality of the performance of English, 59
60
PRONUNCIATION IS IN THE BRAIN
such instances could be treated as both morphological and grammatical errors besides being pronunciation errors. Errors such as those are different from replacing [v] with [b] or [z] with [s] as in the words and , respectively, which are sheer pronunciation inaccuracies. By the same token, an Arab student may make a statement as follows: ‘The house beautiful’ instead of ‘The house is beautiful.’ Surely, this is a grammatical error attributed to the absence in Arabic of what is equivalent to the verb ‘to be’ in English; nevertheless, it is more readily captured by the listener through a gap in the overall pronunciation. Certainly, errors in the overall pronunciation of a statement in L2, such as the ones above, tend to point in the direction that accent as part of speech production and perception process on the part of the speaker/listener can receive interference from other linguistic systems besides pronunciation. In light of such broadening of the concept of accent, one can justify considering it in terms of surface layer vs. deep layer with the former ascribed exclusively to mispronunciations, whereas the latter ascribed to other linguistic factors including grammatical and lexical. 4.2. INTRALANGUAGE AND INTERLANGUAGE ACCENTS
The intralanguage accent refers to sound differences that exist among different dialects or varieties of a given language such as the difference in the pronunciation of // as an alveolar trill or a uvular trill or fricative across different German dialects. Similarly, the pronunciation differences between New York and Chicago dialects or even the differences between Received Pronunciation (RP) of England and the General American English (GAE) of the Midwest constitute intralanguage accent. In the latter case, there are, for instance, some major pronunciation differences between the two varieties. The differences are more distinct in the vowel system than the consonant system. Within the vowel system, the difference is more in diphthongs than in pure (simple) vowels. In consonants, except for the phonetic nature of /r/ and the pronunciation of the intervocalic /t/ as in the word , there are hardly any significant differences. In Assyrian (Modern Aramaic) dialects, the grapheme ܟhas different renditions such as [k], [k], [c and [ʧ]. In Arabic, the standard realization of the graphemes جand ق, for ex-
CHAPTER 4
61
ample, are different throughout the Arabic speaking countries. The former has at least four different realizations, namely [ʤ , a voiced postalveolar affricate as in Iraq; [g], a voiced velar plosive as in Egypt; [ʒ , a voiced postalveolar fricative as in Syria and Lebanon and [Ɉ , a voiced palatal plosive as in Sudan. The قhas at least three different realizations. It is a voiceless unaspirated uvular plosive [q] in Standard Arabic and in several other dialects; in Iraq, it may be realized as voiced velar plosive [g]; and in Egypt as a glottal stop []. In sum, intralanguage accent represents pronunciation differences within L1. Conversely, the interlanguage accent stands for pronunciation differences that emerge when one moves from the native language (L1) to the target language (L2). For instance, when a native speaker of English embarks on learning French or Spanish, a major pronunciation difference one encounters is the radical shift in rhythm. The same is true for Frenchmen and Spaniards learning English. Some typical vowel differences that result in acute accent for native speakers of English learning German and French are the front rounded vowels such as [ ] and also the French nasal vowels such as [ ]. A foremost difficulty in learning the pronunciation of the Semitic languages, especially Arabic, is the dominance of the so-called guttural sounds [, , , , ] and the emphatic (also known as pharyngealized) sounds [ ]. For the specific teaching of the latter category see Odisho, 1981. In teaching pronunciation in cross-linguistic situations, it is important to be aware of the difficulties that arise when one moves from one language to another or even when one moves within the different varieties of one given language. There are other specific situations in which two individuals of the same L1 learning an L2 may have different pronunciation problems due to dialectal differences. To illustrate, if two German speakers, one with an alveolar trill-r dialect and the other with a uvular trill-r dialect embark on learning Spanish or Italian, the former is less vulnerable to r-pronunciation problem than the latter. There are scores of such cross-language pronunciation problems. With native speakers of Arabic learning English, the /ʧ/ as in is not a problem for those speakers of Iraqi Arabic as opposed to Egyptian, Lebanese or Syrians speakers because Iraqis have the sound in their local dialect. The focus in this study
62
PRONUNCIATION IS IN THE BRAIN
is essentially on interlanguage phonological and phonetic accent. 4.3. PHONETIC AND PHONOLOGICAL ACCENTS
From the functional perspective, the phonetic vs. phonological distinction in the nature of accent is extremely significant in teaching pronunciation. It was a major distinction that I developed in the mid-1990s and began implementing in my classes. The distinction was in print in 2003. Pedagogically, it made a substantial difference in helping learners focus on more important problems of pronunciation facing them rather than scratching on the surface of phonetic accent. One needs to understand the difference prior to any elaboration on the applied side of the dichotomy. Phonetic accent refers to a mispronunciation that does not result in a semantic (meaning) change, though it may negatively interfere with the proper comprehension of meaning due to partial detraction from the acceptable standard rendition of a given pronunciation. In other words, it is a mispronunciation that does not directly cause a miscomprehension, but it may hamper it or delay it and become a distraction. Let us now elaborate on the key words in the last statement. An example for the non-semantic nature of this accent is the massive replacement of the English approximant /r/ (English [] and American []) by a tap/flap [], or trill [r] or retroflex tap/flap [] by millions of learners of English. This replacement does not cause a change in meaning in English; it simply phonetically deviates from the normal standard and acceptable rendition of it. It is, therefore, a form of phonetic accent. Similarly, if a native speaker of English learning Spanish aspirates the unaspirated Spanish [p, t, k] or a native speaker of Spanish learning English deaspirates the normally aspirated English [p, t, k] no miscomprehension will result, only a phonetic deviation. Although a phonetic accent may not directly hamper comprehension by replacing one sound for another, it may indirectly interfere with or impede comprehension because a phonetic mispronunciation or a combination of phonetic mispronunciations may serve as an element of noise that can confuse the phonological filter of the listener and, hence, cause miscomprehension or, at least, a delay of comprehension. The latter two symp-
CHAPTER 4
63
toms portray themselves verbally through repeated questions or statements such “What?” “What did you say?”, “I beg your pardon!” or “I did not understand”, etc. Nonverbally, the listener may give you a facial impression of lack of understanding. As for the phonological accent, it is a mispronunciation that directly causes semantic confusion and impedes comprehension. Obviously, a phonological accent always implies a phonetic accent. If, for instance, the inaccurate phonetic rendition of aspiration and non-aspiration between the speakers of English and Spanish was a good example of phonetic accent, a similar inaccuracy in phonetic rendition between the speakers of English and Thai languages will certainly amount to a phonological accent on the part of the English learner of Thai since in the latter language aspiration vs. nonaspiration amounts to a phonological distinction. To illustrate, if a speaker of English fails to aspirate the name [tai] (a person from Thailand), he will be pronouncing it as if it were the word [tai], which means (kidney). A real conversation between an Indian student from Kerala/India and a Thai student went on like this. Indian student: “Are you Tai? She replied: “No.” He went on: “Aren’t you from Thailand?” She replied: “Yes”. He continued: “Why do you say you are not Tai?” She replied: “Tai, means ‘kidney’. I’m not kidney. I am Thai’”. There was laughter.
Similarly, the absence of [ʤ] sound, as in [ʤʤ], in German is easily felt in the pronunciation of German learners of English. Germans usually replace the [ʤ] with its voiceless counterpart [ʧ] as in the English words and which are rendered [ʧk and [ʧouk], respectively, thus causing radical semantic change, a typical outcome of phonological accent. Interestingly, when Arnold Schwarzenegger occasionally appeared on Jay Leno’s show one could hear him say “Jay, are you joking?” which sounded:
64
PRONUNCIATION IS IN THE BRAIN
“Chay, are you choking?” Therefore, the English pronunciation of the word and its derivations constitute a problem for Germans because it sounds as . However, luckily, they do not have this problem in the native language because they either pronounce those derivations with a [g] sound as in [g] similar to in the English word or use the Germanic counterpart of the word, . For Spanish learners of English, one of the main sources of phonological accent is the vowel system both for quantity (short/long or lax/tense) and quality (overall vowel impression). The absence of short/long or lax/tense contrasts in Spanish and their presence in English creates a very serious problem for Hispanic learners of English because virtually thousands of words are semantically confused—some of them resulting in embarrassing situations. Luckily for English learners of Spanish, hardly any phonological accent results from their handling of the five vowels of Spanish; however, English learners of Spanish may have noticeable phonetic accent, especially because of the imposition of their lax vowels, such as schwa [], and other vowel reductions and schwaizations1 that accompany unstressed syllables. In addition to the semantic difference that ensues from the distinction between phonetic and phonological accent, perceptually the phonological accent is much more readily identified by the native listener than the phonetic one because it either triggers confusion in meaning or propagates meaninglessness. 4.4. ACCENT: A NORMAL LINGUISTIC PHENOMENON
The overall tone of the preceding discussions of the phenomenon of accent should give the impression that accent is a normal linguistic fact. It emerges as a result of the cognitive attitude of the brain of a child acquiring L1 versus that of an adult learning L2. Every adult human being has to demonstrate a certain de1
Schwaization, after the neutral vowel schwa []; a vocalic change in the
direction of this vowel.
CHAPTER 4
65
gree of accent whether trying to learn a dialect within his language or a second language besides his. Thus, to speak a dialect or a language with an accent is not a stigma or a pathology and should not be treated as such; however, when the accent seriously interferes with meaning it then begins to obstruct the normal conveyance of the message between two individuals. Earlier on, this was identified as phonological accent versus phonetic accent, the latter of which does not seriously obstruct meaning directly except when it involves a certain degree of deviation from the targeted rendition of several segmental (consonants and vowels) and suprasegmental components (stress, tone, intonation) of the targeted language. Briefly, the general goal of teaching pronunciation should be an attempt at eliminating phonological accent and reducing the phonetic one. Some learners or speakers of an L2 claim that they do not care whether they manifest an accent in their speech or not. Such a claim is made under the pretext that the speaker wants his accent to reflect his ethnic and linguistic identity. In my view, this is a baseless claim and it is no more than a cover-up for the linguistic failure. Logically and aesthetically, any speaker of L2 should portray his ethnic and linguistic identity through perfect demonstration of his/her L1 rather than through the loading of an L2 with mispronunciations infused through L1. 4.5. WHAT IS MEANT BY ACCENT ACQUISITION, ACCENT REDUCTION AND ACCENT IMPERSONATION
To have or not to have an accent is a particularly controversial question. How much ‘accent’ one has is even more controversial simply because it can be a very subjective judgment especially by non-linguists. In real-life situations, there are different routes to accent minimization, namely through: accent acquisition, accent reduction (remediation) and accent impersonation or faking.2 In all three cases, the goal of the person is to try to diminish the differences in pronunciation between himself and a typical na2
Accent impersonation and faking will be used interchangeably according
to the context in which they occur.
66
PRONUNCIATION IS IN THE BRAIN
tive speaker of a target language (L2). In what follows, an attempt will be made to afford a descriptive account of each term coupled with some highlights. 4.5.1. Accent Acquisition
Accent acquisition is primarily the gift that nature bestows upon a child who grows up immersed in the authentic environment of a language—be that the language of the country of birth or the country of the resettlement where he grows up abiding by the rules and conventions of the pronunciation of that given language. The gift could be broadened in scope to include two more conditions that qualify a child or a person for accent acquisition. The former is the total and equal immersion of a child in two languages while the latter is a linguistically talented adolescent or even a young adult who is amply exposed to an L2 and has a passion for the pursuit of mastering that language. No one should exclude others from qualifying for accent acquisition, but the instances are very rare for adults. Nevertheless, very many adults can excel and even outsmart native speakers in their immaculate and creative competence in L2 lexicon, morphology syntax, stylistics and overall fluency. Unfortunately, they may not master the pronunciation simply because the latter seems to be bound cognitively with the limitations of age—the younger the more perfect, the older the less. The renowned novelist Joseph Conrad, a Pole in origin, is a typical example of a person who achieved the highest level at all aspects of linguistic performance in English except for his Polish accent in pronunciation which is a case commonly identified as ‘Joseph Conrad syndrome’. Another example is Henry Kissinger, the former Secretary of State whose fluency, syntax, lexicon are immaculate, but his accent is striking to the ear. The contrast in matters of pronunciation between the skill of a child and an adult reminds one of the difference between the terms ‘acquisition’ vs. ‘naturalization’ in becoming a citizen of the United States. Briefly, obtaining citizenship by acquisition is primarily a right by birth, whereas citizenship by naturalization, especially for adults, is a privilege to be earned by a set of procedures and requirements. Translating this difference in terms of ‘accent-free speech’ and ‘accented-speech’ does not amount to an exact analogy, but it hints at the fact that being
CHAPTER 4
67
born and raised in the same language is different from being born and raised in a language, but relocated to a second one later in life as an adult. Simply, accentless speech is a natural gift for a child born and/or raised in what is supposed to be practically an L1 environment. 4.5.2. Accent Reduction (Remediation)
Nevertheless, if accent acquisition is a difficult goal to achieve because it requires perfection or near-perfection in the execution of a language skill, it should not deter anyone from emulating the targeted pronunciation through accent-reduction as much as possible. In accent reduction, the objective is to suppress the most salient features of one’s L1 and attempt to replace them with those of L2 when speaking it in order to camouflage any readily detectable indications of an accent. Accent reduction is targeted for different purposes foremost of which are the following. 4.5.2.1. Linguistic and Aesthetic
Attain the highest proficiency in pronunciation as part of the overall competency in L2 to avoid accent as much as possible. This could simply be for linguistic and aesthetic purposes. Needless to say, in any L2 language classes, learners aim at a better level of proficiency in the target language, especially in its pronunciation through which they give the native listener a very positive impression especially because of higher intelligibility. 4.5.2.2. Spontaneous Interpretation
Spontaneous oral interpretation that is a daily practice in the United Nations’ General Assembly, Security Council or in any high level negotiation where room for misinterpretation and miscomprehension attributed to any linguistic component (syntactical, morphological, lexical and phonological) should be eliminated or reduced to the minimum. 4.5.2.3. Acting and Broadcasting
An actor or actress would certainly be far more impressive and convincing if he/she were to impersonate a foreign-speaking character as accurately as possible. For instance, Ben Kingsley’s performance in Gandhi’s film was extremely impressive in act-
68
PRONUNCIATION IS IN THE BRAIN
ing and demeanor, although his language would have needed a touch of retroflexion, especially with his ‘r’ sound to impact the native Hindi audience more authentically. Equally important is the role of newscasters in delivering their news bulletins loaded with names of foreign personalities. For instance, now that the name of the Syrian President بَ ّشار ُأالis in the news, I have yet to listen to an American newscaster or anchor pronouncing it correctly; in reality, its pronunciation is seriously distorted. Ironically, the name does not contain any sound that does not exist in English; rather, the distortion in its rendition lies in the misplacement of stress, reduction of Arabic vowels to schwas [] and the overall shift in syllabification. The Arabic pronunciation of the name sounds like this: [baar alasad]3 as opposed to an approximately typical mispronunciation by a native English-speaking newscaster: [br lsd]. The following are some of the major differences between the two renditions. a) Almost all Arabic vowels are Anglicized, especially in the direction of vowel neutralization, i.e. infusion of [] vowels in place of [a]; also the replacement of vowel [a] with [] b) The germination (doubling) of [] (= ) is reduced to a single one. c) The elimination of the germination results in reshuffling the syllabic structure of the first name; thus, instead of [ba] + [ar] it becomes [b] + [r]. d) Finally, the Arabic rolled = [r] is replaced with an approximant retroflex one [] or [] which is not a major change.
Cumulatively, the change in the overall rendition of the name goes beyond being an accent; rather, it is an overall distortion in pronunciation both phonologically and phonetically. A very similar example of mispronunciation in broadcasting that is retained in my memory from some decades ago is the 3
The bold syllables indicate the stressed syllable in each case.
CHAPTER 4
69
pronunciation of the name of the former French President Valéry Giscard d’Estaing. The proper pronunciation of the name is typically [valeʁi ʒiskaʁ dɛ ]; however, the Arabic-speaking Baghdadi television anchor pronounced the name with such a heavy Arabic accent that it was stripped off all of its French language characteristics. Some such typical phonetic/phonological features of pronunciation in French language include: strong tendency for word stress to fall on the final syllable; nasal vowels; voiced uvular fricative [] instead of an alveolar ; and a typical voiced postalveolar fricative [ʒ in place of a voiced postalveolar affricate [ʤ . None of those features were found in the rendition of the name by the Iraqi news anchor. 4.5.2.4. Spying and Espionage
Spying or espionage is a ‘profession’ in which the person may use language, especially pronunciation to camouflage his/her personality without raising suspicion. In other words, the person hides his identity behind an ‘adopted’ pronunciation. Because this ‘profession’ is an extremely risky assignment, it requires a high level of accent reduction and/or accent impersonation. This dimension of accent reduction will be revisited from a different perspective. 4.5.3. Accent Impersonation or Faking
In this section the term ‘accent faking’ is used with a specific connotation; hence, it undoubtedly needs some clarification. In a broad sense, ‘accent reduction’ and ‘accent faking’ overlap to some extent because they involve impersonation of a targeted speaker of a language except for the fact that the purpose can be different. Accent faking can project itself in different forms or strands. 4.5.3.1. Comedians Faking an Accent
A common example of someone impersonating or faking an accent is what comedians do especially when mimicking speakers of languages other than their own to generate laughter. It is interesting to note that not all comedians are equally skillful in impersonation—some are better than others. Let us just consider the case of Jay Leno impersonating Arnold Schwarzenegger. Le-
70
PRONUNCIATION IS IN THE BRAIN
no is a great comedian, but is not a good impersonator; nevertheless, he is very impressive when tackling Schwarzenegger. It is noteworthy that phonetically, his success is not attributed to his imitation of Schwarzenegger’s vowels and consonants, but rather to his overall rhythm, tempo and intonation. What comedians usually do is highlight the most salient features of the targeted speech and mimic them in a caricaturized way to capture attention. This means that a comedian may not necessarily be skillful in meticulous impersonation, but may be good in highlighting some salient pronunciation features similar to what caricaturists do in drawing—to highlight the most striking facial and bodily features and exaggerate them for comedic effect. 4.5.3.2. Building Intimacy
Accent faking may also take the form of socially impersonating the language or dialect of an interlocutor to sound friendly, intimate and trustful; however, in some instances, all those three attributes may be used either honestly or with a twist of dishonestly. In the first instance, the impersonator aims sincerely at bonding with the interlocutor for no ill intention. In the latter instance, he may be aiming at enticing the interlocutor to unveil personal information or even secrets. This may sound like spying, but in reality it tends to be more the action of a curious and nosy individual; however, there is always the likelihood that it can cross over to indirect interrogation or information gathering. As an instructor of phonetics and pronunciation, I have used accent faking of languages that I did not speak, but I knew the sound systems of, such as Russian, Greek and Hindi, among others. I did the faking for two purposes. First, to demonstrate the power of the knowledge of phonetics, especially articulatory phonetics, as a science. Second, to assess my own skill in faking as a strategy for teaching. In the former instance, faking an accent was used as a tool to attract the attention of the learners to my presentation and build up confidence in their ability to hone their skills through attention and practice. In the latter instance, I did it to test my own skill in faking a given accent. Let me cite some examples of my attempts at accent impersonation. In one instance, I had the following encounter with an adult American lady of Russian ethnic background:
CHAPTER 4
71
“One day on the campus of my university, I stopped by a colleague of the Department of Chemistry to arrange for a university event. While we were chatting, the lady in charge of the chemistry labs stopped by to ask my colleague a couple questions. Just as a courtesy gesture, my friend introduced her to me as ‘Ludmila’. Looking at my non-fair complexion, she apparently became more curious as to my ethnic background. I also became equally curious because she portrayed a distinct Russian accent in her English. Once I detected the Russian accent I wanted to ‘play’ my so-called accent impersonation ‘game’ as a means to collect data for this book. The following conversation went on between the two of us: She asked me: “Where are you from?” I said, infusing some sort of Russian phonetic features in my English: “I am from ex-Soviet Union.” She looked at me strangely as if in disbelief: “Are you Russian?” I said: “No, my parents were originally from Azerbaijan.” I chose Azerbaijan to justify my facial complexion as non-fair. She went on: “Do you speak Russian.” I replied: “Just a few words.” At this stage, I felt that she was extremely confused because she was suspicious of my story. I also felt guilty of the confusion I caused her. I immediately, apologized to her and told her I was simply trying to impersonate a Russian accent. One could readily notice that she was relaxed after my apology. She further looked at me and said: “But you don’t know how good you make it.” Linguistically, I was very happy after this encounter because I felt that my impersonation of the Russian accent seemed to have been good enough to make her believe in my fake story.”
With regard to the Greek language, I know only a few words and phrases including: , , , , ; however, when I use them in a Greek shop or with a Greek person, the immediate question is: “Are you Greek?” When my re-
72
PRONUNCIATION IS IN THE BRAIN
sponse is ‘No’, the next utterance is always: “I don’t believe you.” All that I do is apply my articulatory knowledge of the Greek sound system and use the right accentuation. This makes the difference between a phonetician and a ‘lay’ speaker of a language. Let me explain this difference with the help of the following real anecdote: One day, one of my friends asked me whether I knew Greek. I said: “No”. He went on to say: “Do you know the greetings?” I said: “Yes”. He then said: “What is ‘Good Morning’?” I said: “[kali mea ”.
When I asked him why of all Greek language he wanted this phrase, his response was that his mother shared a room in a nursing home with an old Greek lady and he simply wanted to greet her in her native language as a courtesy gesture. A few days later, I saw my friend and asked him about his visit to his mother and her Greek roommate. He said: “Yes, I greeted her, but she did not show any response at all.” I said: “Are you sure you used the right phrase?” He responded: “Yes, I said ‘[kl m].
If one notices carefully there are at least seven (7) phonetic differences between the authentic Greek pronunciation and his Americanized rendition of it. I then thought to myself that she did not respond to his greeting because she did not realize it was in Greek due to the very heavy English accent. 4.5.4. Intralanguage Accent Reduction and Impersonation
Accent reduction and impersonation can be at both dialect level (intralanguage) as well as language level (interlanguage). Although the emphasis is on the latter, some comments must be made about the former. All languages have many dialects that evolve for geographic, socio-economic or ethnic reasons. In
CHAPTER 4
73
many instances, a certain stigma may be attached to a given dialect causing many of its speakers to avoid using it either completely or in certain situations. On the contrary, one of the dialects, usually the standard,4 tends to be the most prestigious, hence a large number of speakers of other less prestigious or local dialects tend to adopt it through formal education or through special orientation such as radio and television anchors and announcers. Many such educated and professional people tend to reverse back to their own social, regional and local dialects in casual communication; hence, they are typically bidialectals. Usually, a good percentage of educated people tend to be bidialectal moving back and forth between the standard and their own dialects as the situation dictates. 4.6. CULTURAL ACCENT
As equally important as the linguistic accent, there are solid grounds for justifying the existence of a cultural accent. Thus, following the common pattern of L1 for native language and L2 for a target language, C1, henceforth, will stand for native culture and C2 for target culture. However, the question still remains as to those grounds that justify the recognition of a cultural accent. There are certainly some non-verbal gestures such as hand, eye and body movements as well as some interjectional filler ‘words’ that differ from culture to culture.5 Strictly speaking, the filler ‘words’ are not necessarily supposed to be words in the linguistic sense. Some of them can be simply interjectional utterances with no well-defined lexical denotations such as = [:] or [], and = [m]; [m] or [m]. The reason why such interjections are phonetically transcribed somewhat differently is because they do not have an exact pronunciation across individuals, dialects as well as languages. For instance, in English is more popular than ; besides, even if the latter is shared, for example, between native English and Russian 4
Linguistically, the so-called ‘standard language’ is also a dialect which
happens to be associated with the schooling system and formal education per se. 5
http://en.wikipedia.org/wiki/Filler_(linguistics).
74
PRONUNCIATION IS IN THE BRAIN
speakers, its pronunciation is somewhat different. Interestingly, the use of such language-specific filler ‘words’ is so deeply ingrained in one’s native language due to acquisition that they tend to be some of the last remaining accent traces of L1 and/or C1 in L2 and/or C2. If one carefully listens to some fairly competent Russian speakers of English, their speech tends to be punctuated with [m]s as fillers. An equally interesting observation about these filler words is that the vowel element in them tends to be consistent with the phonology of the given language. It is not unexpected to have the schwa [] as a dominant vowel in the coinage of filler words in English since this vowel has the highest frequency of occurrence in English. In the absence of a schwa in Arabic and the high frequency of [a] or [] ()فَتحه, it is this vowel that coins the filler word = [a] or [] in Arabic. With regard to hand and body gestures, as part of the cultural accent, there are many examples of them differing with different peoples and cultures. For example, one of the most typical examples of cultural accent for the Japanese while meeting and greeting natives of other languages and cultures is to bow with or while exchanging hand shaking. Equally noticeable is the tradition among Arabs, especially of rural areas or of rural background, who immediately and subconsciously touch their chest with the right hand after shaking hands. 4.7. TRANSITION OF ACCENT INTO ORTHOGRAPHY
Unquestionably, phonological accent is far more detectable than phonetic accent not just in speech, but also occasionally in orthography (written form). In fact, one of the indications of the seriousness of a phonological accent is when it is carried on by the L2 speaker into the orthographic renditions of the mispronunciations. Let us consider some misspellings of adult Hispanic students learning English. It is quite likely for a Hispanic to spell as or the vice versa. This is certainly attributed to mispronunciation. It is an established fact that the Spanish vowel system has a limited inventory of vowels based chiefly on quality differences with no quantitative (length) distinctions. This is a system that was first identified technically as a centrifugal one as opposed to a centripetal system of English in which
CHAPTER 4
75
vowels allow both qualitative and quantitative distinctions (Odisho, 1992). In Spanish, the vowel in (without) has the qualtity that is almost half way between the English vowels in and . Consequently, a Hispanic student fails to distinguish English words such as vs. , vs. or vs. . It is exactly because of the absence of such qualtity differences in their language, pairs of words such as the above are misspelled by adult Hispanic learners of English. In light of such examples, it is not uncommon for a Hispanic person to write a sentence such as: ‘This cars are expensive.’ For Arabic, the sound
is phonologically irrelevant, though phonetically the sound may occur in certain contexts such as when followed by an aspirated sound as in ابتداء [] (beginning); nevertheless, without training and practice, /p/ is the most difficult, and at times embarrassing, sound for Arabs because in the absence of such a sound, hundreds, perhaps even thousands, of words are confused. Words such can become . The impact of this mispronunciation can occasionally overflow into orthography in the form of replacing words of
spelling with . In fact, at times, the phenomenon known as over-compensation may develop, according to which the fear of mispronouncing a given sound leads the speaker to reverse the rendition of the relevant two sounds. Once this over-compensation kicks in, the situation worsens because the Arab learner of English will not only reverse the pronunciation of /b/ and /p/, but will also reverse their orthographic renditions. I have come across many Arab students who pronounce or write as and as . In one instance, in the Iraqi city of Basra, a traffic officer had ordered the sign to be engraved on a concrete slab; unfortunately, it ended up being spelled as . In Kurdish, which is an Indo-European language, the English interdental fricative pair [, ] is absent. It is consistently replaced in pronunciation with the alveolar fricative pair [s, z]. This sound substitution is so powerful that it occasionally stealthily sneaks into their orthography. The anecdote below is relevant to this phenomenon.
76
PRONUNCIATION IS IN THE BRAIN “During the years 1960-1965, I worked as a teacher of English in a high school in the Kurdish city of Sulaimaniya/Iraq. Once, I gave my students a written test. During the test, I moved around the classroom monitoring the performance. On the desk of one of the students, I noticed a small piece of paper with Arabic scribbling on it. At first glance, I did not pay attention to it because the written test was in English; however, when I moved away from student’s desk, I had a different vision of the Arabic scribbling. There were some strange orthographic features in the scribbling that made it look different from the overall visual impression of standard Arabic orthography. The writing had more than usual recurrence of the letters سand ز. I went back to the student’s desk and picked up the piece of paper and began reading it. To my utter surprise, it was a text in English transliterated in Arabic and had relevance to the exam. The reasons why it had an extraordinary number of سand زwas simply because all English
can be the source of phonetic and phonological accent for speakers of some languages. Below are further comments on some of the cross-language phonological inconsistencies.
105
106
PRONUNCIATION IS IN THE BRAIN
7.2.1. Interdental Pair /, /
In terms of markedness and unmarkedness of human language sounds, this pair [, ] can be considered marked because it is rarely attested in the majority of known languages. Due to its rarity, many learners of English encounter serious pronunciation problems resulting in a distinct phonological accent. More widely, the pair [, ] is often replaced with the alveolar fricative pair [s, z] or the alveolar plosive pair [t, d] depending on the phonology of the specific language. The less common replacement of [, ] is with the labialdental fricatives [f, v]. This last substitution is also known in literature as ‘th-fronting’ which is common in some dialects of British English as well as in AfricanAmerican English.1 It is also observed in New Zealand English.2 To demonstrate examples of the above-mentioned substitutions, a German or Kurd learner of English, for example, renders the words and as and , while a Pole, Filipino or Assyrian learner of English renders the same two words as and , respectively. In each case, the outcome is a very serious phonological accent as well as a phonetic one. It results in a phonological accent because it changes the meaning of thousands of words causing crucial semantic confusion. Even without the semantic change resulting from the failure to properly pronounce the pair [, ], the mere phonetic change can cause serious phonetic accent. If one listens to exPope Benedict XVI3 delivering speeches in English, there is a sibilant impression running throughout his English pronunciation. The impression of sibilance is generated by the fact that English is naturally a language rich in sibilant sounds, typically [s, z, , ʒ . Now, if a German, Frenchman or Kurd learner of English were to convert all the [, ] sounds into [s, z] that 1 2
http://en.wikipedia.org/wiki/Th-fronting.
http://www.victoria.ac.nz/lals/resources/publications/nzej-
backissues/2003-elizabeth-wood.pdf. 3
Listen
to
his
reading
com/watch?v=w3fi93umuc4)
a
text
in
English
(http://www.youtube.
CHAPTER 7
107
would seriously reinforce the dominance of sibilance. It should be pointed out, however, that the high frequency occurrence of [, ] sounds, especially [], is not because they occur in the structure of many English words; rather, because the words in which they occur are of high frequency of occurrence such . Thus, when all the sibilance imposed by an L2 learner of English is added to the existing sibilance that is already naturally in English, it generates a pervasive semantic confusion and disturbing sibilance which reverberates as noise. Because of this dominating pervasiveness of sibilance in the pronunciation of a German or Kurd learner of English, the influence at times may be reflected even in orthography (as was pointed out in 4.6, above). The other form of conversion of [, ] is into [t, d] which causes equally serious phonological accent as it semantically confuses scores of words such as which are rendered , respectively. Phonetically, it creates an overflow of [t, d] sounds which, in turn, infuses considerable noise into the flow of oral discourse in English. 7.2.2. Approximant /r/
The most unmarked (common) types of sounds in the majority of languages throughout the world are the tap = [] and the rolled = []. Although such sounds exist in several dialects of English, the one dominant in Standard English varieties such as the English RP and GAE tends to be an approximant or the so-called frictionless continuant despite some phonetic difference between the two. The English RP is a voiced postalveolar approximant [], whereas the American one is a voiced postalveolar retroflex approximant []. Perhaps, the most interesting and atypical fact about the phonetic behavior of this sound in English is its suprasegmental impact on the context of the words in which it occurs. The contextual rules of ‘r’ pronunciation in the two varieties are different. In the general classification of the English dialects and varieties, GAE is categorized as an ‘r-dialect’ (or rhotic dialect) meaning that the ‘r’ is pronounced in all linguistic contexts, whereas RP is, perhaps, the most typical of all ‘r-lessdialects’ (non-rhotic dialect) meaning that the ‘r’ is not pro-
108
PRONUNCIATION IS IN THE BRAIN
nounced except in certain linguistic contexts. The rule for the positional pronunciation of ‘r’ in RP is very simple—‘r’ is pronounced when it is followed by a vowel sound within a word and across the word boundary. For instance, in the following pre-consonant, pre-silent vowel and word-final positions, the ‘r’ is not pronounced:
Pre-consonant: , and ,
Pre-silent vowel: , and ,
Word-final: ,
,
,
,
,
and
whereas in the following pre-vowel within a word or across a word boundary it is pronounced:
Pre-vowel: , , and ,
Cross-word pre-vowel: , .
It should be reiterated once again that the difference in RP or GAE pronunciation by learners of English as L2 is essentially of a phonetic nature implying that it rarely amounts to phonological accent;4 nevertheless, the phonetic difference is extremely noticeable. The retroflexion in GAE colors the adjacent vowels, whereas the positional constraints on RP ‘r’ results in two phonetic variations in the overall pronunciation. First, dropping the creates a consonantal vacuum for an L2 learner. Second, dropping it in word final positions or pre-silent vowels results in more diphthongal renditions of the vocalic elements as in [] and [] versus their renditions in GAE as less diphthongized vowels [] and []. Additionally, the substitution of the approximant English with a tap, rolled or retroflex flap /r/ as with speakers of Span4
Except, of course, when the is confused with another sound, such
as in Japanese when it is replaced with .
CHAPTER 7
109
ish, Italian, Arabic, Russian, Turkish, Indian, among many other languages, tends to result in so much phonetic accent that it may snowball into some sort of acoustic noise that, in turn, interferes with comprehension on the part of the native English listener. A special note is in order with regard to the typical retroflex flap /r/ of the sub-continental Indian languages. In such languages, retroflexion is pervasive and phonetically covers almost all consonants and vowels. Actually, from the perspective of Firthian linguistics,5 phonetic retroflexion in such languages functions as a prosody that runs throughout complete words and even discourse. In languages such as Hindi and Urdu, retroflexion seems to run throughout their entire oral discourse. This explains why many Hindi or Urdu learners of English, or any other language, color their L2 with a heavy trace of retroflexion. Most strikingly, retroflexion constitutes the most salient feature in the articulatory settings of Indian learners of other languages, in general. It is not a thematic diversion in this section to bring in the /l/ and /r/ confusion by Korean, Japanese and Chinese learners of English. Let us consider the case of Japanese learners. Phonetically, in Japanese language the , which is, generally speaking, identified as a tap or flap, while the , which is identified as a lateral, are two phonetic variations (allophones) of the same phoneme. In terms of the neurolinguistic storage of sounds in the brain, the two sounds of Japanese are stored in one slot. Unlike Japanese, in English the and are two autonomous phonemes independently stored in two separate slots. Besides the primary difference between the phonological function of the s and s, their phonetic realizations are also different. It is, therefore, quite natural for Korean, Japanese and Chinese students to encounter serious phonological difficulty in mastering the independent contextual production of the two phonemes of English. In fact, these two liquid phonemes in English remain some of the last sounds of English that those oriental learners of English manage to successfully pronounce. 5
After the late Professor J. R. Firth of London University.
110
PRONUNCIATION IS IN THE BRAIN
7.2.3. Voiceless and Voiced Alveolar Fricatives /s/ and /z/
This pair in English represents the unmarked variety of fricatives. Their mispronunciation is not as pervasive as the interdental pair /, /. However, in cross-language studies, differences arise that lead to mispronunciations in the form of accent. For instance, the voiced alveolar fricative /z/ phoneme is missing in Spanish which results in its replacement with /s/ in Latin American Spanish and a // in Continental Spanish leading to both phonological and phonetic accent. In Latin American Spanish, words such as , , and are confused with , , and resulting in semantic confusion. Furthermore, what is linguistically significant about [z] is that even when it does not cause phonological confusion, it certainly causes a very noticeable phonetic accent because of the high frequency of occurrence of /z/ sounds in English. The failure to pronounce the [z] is extremely noticeable even among otherwise fluent Hispanic speakers of English. Just recently, CNN television broadcast a program on the techniques of interrogation of foreign terrorists in which a CIA high-ranking officer of Hispanic background pronounced most of his [z] sounds as [s] in spite of the fact that his English was fluent and native-like. I vividly recall that the word = [p] was rendered [ps]. English words such as: , , , etc. in which the is pronounced [z], are traditionally realized as [s] by Hispanics. It is also noticeable that for Greek learners of English, the alveolar sounds [s, z] are often replaced by the alveolo-palatal sounds [ ] which are auditorily and impressionistically confused with the [, ʒ pair for the native listener. The confusion is not unexpected as they actually sound like them because [, ] are located exactly half-way between [s, z] and [, ʒ in terms of their place of articulation. It is worth noting that because Greek lacks the postalveolar pair [, ʒ sibilants “their sounds like Sean Connery’s ”.6 Traditionally, classical Greek historians have confused the alveolar pair [s, z] or the postalveolar 6
http://greek.kanlis. com/ phonology.html.
CHAPTER 7
111
one [, ] of other languages with their alveolo-palatal sounds [, ]. A piece of historical evidence that supports this claim is embedded in the chronicles of ancient Greek historians documenting Alexander the Great’s invasion of the Middle East. For example, notice that the English word ‘Assyria’ is the anglicized rendition of the Greek name ‘Aσσυρία’ in which the geminated sigma originally represents the germinated = [] in the ancient Assyrian name = [aur]. This indicates that the Greek historians identified the ancient Assyrian [] as sigma which English translators, in turn, rendered as thus creating the word . 7.2.4. English Plosives: /p b, t d, k g/
These three pairs of English plosives are quite unmarked (common), as in many languages, but learners of English as L2 with different linguistic backgrounds may encounter some problems in pronouncing them. Before citing some examples of pronunciation problems, it is necessary to point out that the voiceless plosives /p, t, k/ of English are aspirated i.e., [p, t, k] which means they are followed by a puff of air when released. English does have voiceless unaspirated variants of the aspirated ones only in post [s] initial clusters such as [sp__, st__, sk__]. Hence, the aspirated plosives in = [pn], = [tl] and = [kn] become unaspirated (or rather deaspirated) as in = [spn], = [stl] and = [skn]. Unfortunately, sounds that have only a phonetic occurrence in a language are not recognized phonologically by the brain of the native speaker. With some learners of English the presence or absence of aspiration may lead to phonological and/or phonetic accent. For instance, the Thai language has phonologically three-way plosive contrasts, namely voiced, voiceless unaspirated and voiceless aspirated as in the following examples in table 7.1, below.7
7
Handbook of the international phonetic association, Cambridge University
Press, 1999.
112
PRONUNCIATION IS IN THE BRAIN Voiced Plosives
Example
Meaning
[b]
[bā:n
to bloom
[d]
[dâ:n]
calloused
–
–
–
Voiceless Unaspirated Plosives [p]
[pā:n
birthmark
[t]
[tā:n
sugar palm
[k]
[kā:n
act
[p]
[pā:n
belligerent
[t]
[tā:n
alms
[k]
[kā:n
shaft
Voiceless Aspirated Plosives
Table 7.1. Three-way plosive distinction in Thai language
Thai learners of English are not expected to encounter pronunciation difficulties with the English aspirated plosives [p, t, k] except maybe in final position when Thai speakers pronounce them with no audible release whereas in English there is an option to release them (Kanokpermpoon, 2007). They may also have a perceptual problem in distinguishing the voiced plosive from the voiceless unaspirated plosive. On the reverse, English learners of Thai will encounter serious problems in mastering the production of the voiceless unaspirated plosives and distinguishing them from the voiced ones. Hispanic learners of English do not have a phonological problem with English plosives; nevertheless, they might demonstrate a certain degree of phonetic accent in rendering the aspirated plosives—especially in initial position—as unaspirated since their own voiceless plosives are unaspirated in nature. Because of the high frequency occurrence of initial plosives in English, a rendition of English by someone who fails to produce the aspiration gives the overall pronunciation a distinct phonetic accent. This has been typical of some of my Hispanic students.
CHAPTER 7
113
Arabic has all the plosives of English8 except the voiceless bilabial plosive /p/. Its absence is by far the most significant phonological problem for Arab learners of English inasmuch as segmental phonology is concerned. The failure to produce a /p/ and replace it with a /b/, almost exclusively, riddles the Arabic rendition of English with /b/s. It creates an unmistakable phonological and phonetic accent. Also noticeable with Turkish, Farsi and Greek learners’ rendition of English is the replacement of the velar plosives /k/ and / g/ with palatal plosives // or // for the voiceless and /Ɉ/ for the voiced. Although such a shift does not add up to a phonological accent, it does, however, color the rendition of English discourse with an distinct palatalized impression which, subsequently, manifests a readily distinguishable phonetic accent that runs throughout their discourse in English. Such palatal plosives are most distinct among Iranian speakers of English because of the palatal plosives in Farsi. Just to demonstrate the Farsi substitution of English velar plosives with palatal plosives, an acquaintance of mine used to pronounce the title of a cultural festival in the village of Skokie/Illinois ‘Coming Together in Skokie’ as [cmn tdr n scoci] for [km tgr n skoki]. Although there is no phonological accent, the phonetic one is more than a mouthful. 7.2.5. Labio-Dental Fricatives /f, v/
This pair of sounds can be a source of difficulty for learners of English of different linguistic backgrounds. For instance, in the Tagalog language of the Philippines, this pair is missing, therefore, the pair is replaced with /p/ and /b/.9 The substitution of such a pair in combination with the substitution of the English pair /, / with /t, d/ results in a very serious phonological accent in the rendition of English by Tagalog speakers. Typically, 8
It does not have /g/, but it is widely found in local vernacular Arabic
and many dialects. 9
For an interesting impersonation of English by a native speaker of Taga-
log watch the video at: (http://www.highpoint-ieltsblog.com/2011/03/filipinopronunciation.html).
114
PRONUNCIATION IS IN THE BRAIN
when the people of the Philippines are asked about their identity as a people they reply ‘Pilipino’ not ‘Filipino’. In Farsi and Turkish, the English /v/ is realized differently; most likely as the labialvelar approximant [w] or the labialdental approximant []. When replaced with the former it may lead to a phonological accent in pronouncing as or as ; otherwise, it merely leads to a phonetic accent. In Assyrian (Modern Aramaic), /v/ has a wide range of phonetic realizations (phonetic variants) depending on different regional and tribal dialects as demonstrated in table 7.2, below. Realizations of /v/
Phonetic
[]
Labial-velar approximant10
[]
Labial-palatal approximant
[]
Labial-dental approximant
Description
Table 7.2. Different phonetic realizations of the English phoneme /v/ in Assyrian.
7.2.6. The Affricates /ʧ ʤ/
Although these two sounds in English do not have specific letters to signal them, they are in fact quite common sounds and can be a cause of mispronunciation for L2 learners of English. It has already been pointed out earlier that Germans have difficulty with /ʤ/, which they replace with its voiceless counterpart /ʧ/. The French do not have this pair, therefore, they substitute the sounds with their fricative counterparts /, ʒ/. For Arab learners of English, the difficulty of such sounds may be relative depending on which Arabic dialect is in the background. For instance, speakers of eastern Arabic dialects such as Iraq, Saudi Arabia and the Gulf have much less difficulty 10
The feature “labial” with [w and with [ should really be “bilabial’
because both lips are involved in conjunction with the tongue configuration in the velar region for the former and the palatal region for the latter.
CHAPTER 7
115
in pronouncing /ʧ / and /ʤ/ and / as opposed to the speakers of Western dialects, especially Lebanon, Syria and Egypt. In the first two countries, they are usually replaced with // and /ʒ/, respectively, whereas in Egypt the /ʤ/ tends to become a voiced velar plosive /g/ and the /ʧ / is non-existent, and if attempted it will become a //. Interestingly, the pair /ʧ ʤ/ does not occur in Greek; therefore, Greek learners of English tend to replace the pair with the alveolar affricates /ʦ ʣ/. For instance, a Greek learner of English is expected to pronounce as [tʦ and as [ʣʣ . It is a substitution that generates a distinct phonetic accent. 7.3. CONCLUDING REMARKS
Because English is the most widely used language throughout the world, it has been used as the basis for reflecting the broad variety of phonological and phonetic accents that learners of English as L2 demonstrate in their rendition of its consonants. This, however, should not conceal the fact that English learners of other languages encounter a broad array of phonological and phonetic accents. Another aspect of the English consonantal system that should be highlighted is that most consonants of English are phonetically unmarked (common). This fact suggests that English learners may encounter their most salient difficulties when tackling the marked (uncommon) sounds of other languages such as the gutturals and emphatic of Arabic, retroflexes of Hindi, palatal plosives of Farsi and above all, the clicks in some African languages.
CHAPTER 8: EXAMPLES OF CROSS-LANGUAGE ACCENT-CAUSING VOWELS 8.1. SALIENT FEATURES IN GENERAL VOWEL DESCRIPTION
Foremost among the phonetic features used for vowel description are the terms quality and quantity. Quality is defined in terms of tongue-height, tongue-position and lip-shape, etc. and their combined acoustic impact on the ears of the listener. As for quantity, it is defined in terms of shortness vs. length and/or laxness vs. tenseness. A more comprehensive option for vowel systems identification has been introduced in the form of centripetal vs. centrifugal dichotomy which affords a more general model (Odisho, 1992). Because of some differences between the GAE and RP vowel systems and the manner in which those differences are transcribed, the RP system will be used to demonstrate the nature of the centripetal system. Generally speaking, this system tends to have a schwa vowel, [], with the rest of the vowels displaying different degrees of quantitative and/or qualitative values in unstressed positions. In figure 8.1, below, the vowel symbols indicate only differences in the quality of RP vowels, while in 8.2 the vowels are impressionistically marked phonetically for full-length and the short are left unmarked with schwa being the shortest and most reduced in quality. In the description and transcription of RP vowels Gimson’s (1967) model is followed because his transcription of individual vowels indicates both quality and quantity features unlike Daniel Jones’ model (1956) where the emphasis is primarily on quantity (i.e., either short or long). Gimson’s, in my opinion, is the most innovative and practical system especially in teaching advanced and efficient cross-language comparative pronunciation with emphasis on accent reduction. In this study, the only divergence from Gimson is in the identification of the English vowel []. Almost all those who have dealt 117
118
PRONUNCIATION IS IN THE BRAIN
with this vowel have identified it as short. However, as a phonetician and native speaker of Arabic, I beg to differ with this identification because the Arabic vowel الفُمد, which is the long counterpart of فَتحة, sounds phonetically, especially in non-emphatic contexts,1 almost identical with English [] as in the English and Arabic words in table 8.1, below. Arabic words
Meaning
Arabic words
Meaning
English words
هَم بَت َسم فَت َسد
worry
هام بات سام فات ساد
important
stayed
poisonous
elapsed
prevailed
with Fatha
decision poison weaken dam
with Alif
with []
Table 8.1. Words matching Arabic الفُمدwith English vowel [].
Impressionistically, the Arabic words in column #3 and the English ones in #5 are pronounced almost identically. Thus, in contexts other than with Arabic emphatics, the gutturals and , الفُمدhas a phonetic variant that is identical in vowel quality with English [] especially when the pronunciation adheres to that of Modern Standard Arabic and Classical Arabic. It is true though that فَتحةis not absolutely exactly of the same vowel quality of الفُمد, but the latter is identical with English [] in both quality and quantity. This is the rationale for assigning the length mark to the English vowel []. Marking such quantity and quality differences is extremely important especially for learners of English whose native languages tend to have centrifugal systems. Essentially, the centrifugal system has the propensity of avoiding two inclinations. First, it avoids any quantity (length) contrasts. Second, it leans in the direction of avoiding vowel reduction especially in the form of a schwa. Doubtless, there will 1
Not adjacent to < >طُصُضُظ.
CHAPTER 8
119
be some difference in the quantity of the unstressed vowels versus the stressed ones, but this is hardly a striking difference. According to such templates for vowel identification, English will be a typical representative of a centripetal system as opposed to Spanish, which will be a typical representative of a centrifugal system (figure 8.3) with Arabic (figure 8.4) falling half-way between the two. Notice that the vowels for Arabic indicate major quantitative contrasts coupled with a limited degree of quality difference which is not readily perceptible by people with no professional phonetic experience.
Figure 8.1. The English RP simple vowel system—a typical centripetal one.
120
PRONUNCIATION IS IN THE BRAIN
Figure 8.2. The English RP simple vowel system—a typical centripetal one with relative length marks.
Figure 8.3. The Spanish vowel system—a typical centrifugal one.
CHAPTER 8
121
Figure 8.4. The Arabic vowel system—halfway between centripetal and centrifugal.
Any cross-language teaching of pronunciation to avoid phonological and phonetic accent can hardly be effective if the approach handles the vowels individually and as decontextualized segments. An efficient mastery of a different vowel system cannot be attained without a holistic approach to the teaching of the systems through their most characteristic features. Even minor phonetic differences in a given language may have phonological weight in the target language. Thus, in cross-language teaching of pronunciation any phonetic differences matter—if not to avoid being caught in the phonological trap then at least for dodging a phonetic accent. To demonstrate, those learners with a native centripetal vowel system should be trained on both the avoidance of schwas and vowel quality and quantity reduction when learning languages with centrifugal systems. Conversely, learners with centrifugal systems should be trained on the production of schwas and vowel quality and quantity reduction in appropriate contexts and conditions. 8.2. THE VOWEL SYSTEM OF ENGLISH
Obviously, English being the native language of several countries and also the most widely used international language, one
122
PRONUNCIATION IS IN THE BRAIN
expects some differences, or perhaps even some major differences, within its different standard varieties. To avoid drifting sidewise into minutia, the focus will be on GAE simply because it is gradually becoming more commonly used internationally. Occasional mention of RP will be made when and where necessary. 8.2.1. Simple Vowels of General American English
There are some noteworthy differences between the vowels systems of GAE and RP. Generally speaking, the two systems are more different in the domain of diphthongs than in simple vowels. With regard to simple vowels, there are some differences in the qualitative and qualitative values assigned to the same phonetic symbols. For instance, in GAE the vowel phoneme // as in and // as in are used to indicate simple vowels, whereas for RP the quality of vowels in the preceding two words is diphthongal and is marked by the symbols /ei/ and /ou/, respectively. However, a more striking difference is in the absence of the vowel quality [o] = [ɒ] as in the RP rendition of the words , etc. Another diversion in GAE away from RP system, is the emergence of what are known as ‘r-colored vowels’ such as [ɚ and [ɝ also named as ‘unstressed schwar’ and ‘stressed schwar’, respectively.2 Table 8.2 below, represents GAE simple vowels with examples. The relative phonetic and/or phonological quantitative differences of vowels are marked with length mark [] and its absence indicates shortness. Highlighting such phonetic/phonological differences are absolutely essential in crosslanguage teaching of pronunciation without which learners can hardly avoid manifesting accent.
2
MacKay, 1978.
CHAPTER 8 Vowel Quality
Vowel Quantity
Symbol
Indicator
123 Example beat bit
bait bet
bat
father; shot
boot
book
but
boat
bought
about
ɚ ɝ
writer
bird
Table 8.2. Simple vowels of General American
Several points are worthy of consideration. First, the English vowel system, whether GAE or RP, is rich in quality compared to many vowel systems the most widespread of which being the five-vowel system.3 Second, to mark the quality and quantity differences sheds better light on the nature of the vowel systems a fact that is extremely significant to highlight in comparative phonetic and phonological studies especially when accent reduction and remediation are targeted. To demonstrate, if one has to phonetically transcribe the Spanish vowel in the word (without) compared to and of English, one has no choice but to mark the vowel in Spanish with halflength mark [i] because it is slightly different in both quality and length compared to English . English has to be transcribed as [] to set it apart from both English [], with a very short vowel, and Spanish [] with a half-long 3
Ladefoged and Maddieson, 1996.
124
PRONUNCIATION IS IN THE BRAIN
vowel. Third, in teaching comparative vowel systems such as English and Spanish, for instance, it is very difficult to enable the learners on both sides to master each other’s vowels without highlighting the tiny phonetic differences of quality and quantity across the vowel systems. Typically, Russian, Italian, French, Japanese and Spanish learners of English seriously confuse the English vowels such as in vs. and vs. . They reduce each pair of words to one in the form of half-long vowels as [dim] and [pul]. As surveyed above, the differences in the vocalic systems of the two main varieties of English are many, but fortunately, they often result in phonetic accent between its native speakers; however, at times, they may result in phonological accent as in the examples in table 8.3, below: GAE
Pronunciation
RP
Pronunciation
[kd]
[kd]
[ht]
[ht]
[pt]
[pt]
[p]
[p]
Table 8.3. Phonetic differences in vowels between GAE and RP can be phonological.
Thus, a combination of phonetic and, in rare cases, phonological differences makes the study of vowels in RP and GAE an area worthy of attention in teaching pronunciation, especially when making a choice between the teaching of one or the other variety. 8.3. SELECTIONS OF CROSS-LANGUAGE ACCENT-CAUSING VOWELS
In this section, attempts will be made to identify some of the most salient and readily perceptible features of both phonological and phonetic accent of learners of English as L2 with different linguistic backgrounds. Evidently, there will be limitations on what this section will include partly because of my limited linguistic knowledge and experience with languages and partly to avoid repetition of similar information. Additionally, at the end of each sub-section, the experience of learners will be reversed, i.e. native English speakers learning other languages.
CHAPTER 8 8.3.1. Hispanic Learners of English Vowels
125
It has already been demonstrated that English and Spanish vowel systems are almost maximally contrastive, the former being a centripetal system and the latter a centrifugal one. Due to this radical difference, mastering the vowel system of English becomes the foremost phonological and phonetic difficulty for Hispanic learners of English. The difficulty is not simply attributed to differences in quality and quantity of vowels; rather, the dynamics that govern the qualitative and quantitative features of the vowels involved further complicate the problem. It is the combinations of both factors that jointly lead to serious phonological and phonetic accent for Hispanic learners of English. A simple analogy in this regard may be beneficial in helping the learner envisage the difference between the two vowel systems. If each vowel in English and Spanish is likened to a car and the slot of the vowel to a garage, then Spanish will have five garages with five cars in them, while English will have twelve garages with 12 cars in them. Thus, when a Hispanic intends to learn English, he, subconsciously, has no choice but to allow the parking of more than one car in each of his five garages as it is schematically represented in figure 8.5 below, where the arrows show which vowels in English may mistakenly be identified as the ‘same’ vowel in Spanish. In linguistic terms, Hispanic learners will take two or three vowels in English as one vowel. And this is the most prominent pronunciation problem for Hispanics embarked on learning English. To demonstrate, each pair of the English vowel contrasts such as in the pairs = [] vs. = [] or = [] vs. = [] will go into one slot of [] for the former pair and [] for the latter one. Most important of all in terms of both quality and quantity, neither [] and [] are [] nor [] and [] are []. Phonetically, as well as phonologically in this case, these are certain vowels that should be precisely and carefully taught in an English/Spanish cross-language teaching of pronunciation. These examples serve to highlight one of the most challenging— if not the most challenging—problems for L2 learners of English whose vowel systems are very limited in quality and quantity. At times, the confusion can be extremely embarrassing when some obscene or taboo words are involved such as vs.
126
PRONUNCIATION IS IN THE BRAIN
, vs. , vs. and 4 vs. .
Figure 8.5. Misidentification of English vowels by Hispanic learners.
The above schematic diagram indicates that the English vowel system is a complex one, whereas the Spanish system is a relatively simple one as demonstrated with the words in table 8.4, below.
4
In its vulgar sense. These are some of the embarrassing pairs that one
may hear in classroom situation.
CHAPTER 8
127
Vowel Grapheme
Vowel Phoneme
Example
/a/
/e/
or
/i/
/o/
/u/
Table 8.4. Simple Spanish five-vowel system.
In fact, some describe it as “the essence of simplicity and elegance” (Stockwell and Bowen, 1965). What drives the two systems further apart is the substantial difference in the dynamics of vowel reduction in English, especially with regard to word syllable structure and the location of primary stress within a word or stretch of words. Spanish has hardly any noticeable variation of vowel quality and quantity in different contexts and across its different dialects (Stockwell and Bowen, 1965). As for its diphthongs, it is quite natural for a simple vowel system to have a simple and basic combination of diphthongs. The most common diphthongs in Spanish are the following: Diphthong
Example
/ei/
ley, reina
/ai/
hay, taita
/oi/
soy, coy
/au/
auto, chao
Table 8.5. Simple Spanish diphthong system.
The limited variety of diphthongs in GAE compared to RP, especially with diphthongs in words ending with where the is deleted and a schwa [] is inserted as in = [hi] and = [], makes the transition of Hispanics into GAE somewhat easier. However, the above observation about the easier transition into GAE compared to RP should be considered with caution because one can unequivocally state that phonetically every single simple vowel or diphthong in the two languages is virtually different from what phonologically
128
PRONUNCIATION IS IN THE BRAIN
may be considered a counterpart. For instance, phonologically one tends to say that both languages have the /au/ diphthong; nevertheless, phonetically, this is not exactly accurate because the two constituents that make up the diphthong in each language are phonetically different in the first place. At least, one can state that the English diphthong /au/ is originally composed of an [] vowel gliding into an [] vowel, whereas the Spanish diphthong is the coalescence of the [a] and [u] vowels. This is a strictly phonetic assessment that many teachers of pronunciation may not be professionally qualified to be aware of. From this perspective, the intention should be to secure as near a native-like pronunciation as possible that will dispel any confusion in meaning on the part of both the listeners and speakers. 8.3.1.1. English Learners of Spanish Vowels
Generally speaking, English learners of Spanish are expected to reverse what Hispanic learners of English do. They have to learn to shrink the domain of vowel quality diversity and eliminate the influence of schwa completely. Fortunately for such learners, due to the much richer vowel inventory of English than its Spanish counterpart, the problems will be more of phonetic nature than phonological. Nevertheless, it is worth mentioning that the dominating tendency of vowel reduction in English is expected to be the culprit for many unwanted qualitative and quantitative changes in the proper production of Spanish vowels. This tendency, if left unchecked, can become the cause of a serious distortion of the overall rendition of the pronunciation of Spanish primarily in the form of imposing a stress-timed rhythm on a syllable-timed one for which Spanish is well-known. A native speaker of English should never allow himself to be misled in pronouncing the Spanish word = [kolor], which is orthographically identical with the English one, as [klɚ or [kl]. Unlike English, which has remarkable inconsistency of sound and orthography, Spanish has one of the highest consistencies among known languages. 8.3.2. Arab Learners of English Vowels
Arabic and English are two languages that are drastically different in language family, sound systems and orthographic systems.
CHAPTER 8
129
English belongs to the Indo-European family, while Arabic is a typical Semitic language. The sound systems of the two languages differ extensively in consonants, vowels, stress placement and the dynamics. Arabic, like English, is the native language of a large population inhabiting a very large area. Consequently, it has a wide range of different regional, social and ethnic dialects. Some familiarity with a few most salient linguistic characteristics of the dialects is important for any learner of Arabic because even the so-called Modern Standard Arabic (MSA) is regionally influenced by those dialects. In fact, one can easily distinguish among different standard varieties of Arabic such as Iraqi Standard Arabic, Egyptian Standard Arabic and Lebanese Standard Arabic etc. These standard variants are not only different in segmental (consonants and vowels) pronunciation, but also in lexicon and, at times, even in the overall rhythm and melody, especially if North African Arabic varieties are considered. One typical deviation of the dialects away from MSA is the enhancement of the basic three vowel-quality system into a five vowel-quality one by adding the mid vowels of [] and []. The three simple vowels of Arabic usually combine in producing the diphthongs [ai] and [au] as in table 8.6, below. Thus, the above enhancement of vowel quality, through the influence of the Arabic dialects, does, somewhat, help Arab learners of English in handling more English vowels. Nevertheless, the system remains restricted in quality compared to English. Essentially, it is a simple triangular system of maximally differentiated vowels of /, , /. As a corollary to the restricted vowel quality in Arabic, its diphthongs are also limited in number. This is why some linguists are reluctant to accept the existence of diphthongs in Arabic. They opt to identify them as vowel plus a semi-vowel of يj] or وw].
130
PRONUNCIATION IS IN THE BRAIN Word
Meaning
SA in IPA
DA in IPA
َبيت بَين لَيل َكيف َسيف َكون يَوم بَول لَون
house
[bajt]
[bet]
between
[bajn]
[ben]
night
[lajl]
[lel]
how
[kajf]
[kef]
sword
[sajf]
[sef]
universe
[kawn]
[kon]
day
[jawm]
[jom]
urine
[bawl]
[bol]
color
[lawn]
[lon]
Table 8.6. Vowel contraction in Arabic and creation of the mid vowels [] and [].
The nature of the problems facing Arab learners of the English vowel system is, overall, typical of a transition of speakers of non-centripetal vowel system to a centripetal one. Such learners of English are usually pressed for enhancement and diversification of their vowel quality range alongside the mastery of vowel reduction and schwa production. The most typical feature of the centripetal system is the existence of a schwa vowel []. The predominance of a schwa [] in English and its absence in Arabic is the main culprit behind the tendency of the so-called ‘word-deflation’ in English vs. ‘word-inflation’ in Arabic (Odisho, 2009) which is one of the most primary causes of accent by Arab speakers of English (Odisho, 2013). To demonstrate, the reader is referred to the manner in which the full name of the ex-President of U.S.A. William Jefferson Clinton is transliterated in Arabic. Its expected traditional Arabic transliteration would be وليام ُجي ِفرسون ُكلينتون5 instead of a more accurate rendition of ِول َيم. It is clear that in the English pronunciation of the President’s name there are no long vowels, whereas in its Arabic rendition there emerge five long vowels which, in turn, bring about a major shift in the 5
http://ar.wikipedia. org/wiki.
CHAPTER 8
131
rhythmic structure of the name and its pronunciation as visually demonstrated by the following stress pattern in English rendition vs. the stress pattern in its Arabic rendition with larger dots standing for the stressed syllable in each case. Such a shift in stress pattern is the most powerful source of accent generation by Arab speakers of English including those who are highly educated (Odisho, 2013). 8.3.2.1. English Learners of Arabic
English learners of Arabic bring with them a centripetal vowel system with a broad range of vowel qualities dominated by a schwa. Consequently, the first thing they have to do is to learn how to restrain their strong inclination toward vowel quality diversity. For instance, they have to eliminate the use of a schwa as well as any tendency in the direction of schwaization and vowel reduction. Equally importantly, they have to try carefully to maintain the vowel quality of Arabic vowels. They should not attempt to render the Arabic past tense forms, which are predominantly formed with triliteral consonantal roots and vocalized with three فَتحةvowels such as َكت//, َد َر // and = //, as // and //, respectively. Additionally, even the stressed vowel [a] may be replaced with English []. Such a substitution of vowels does not only result in a shift in vowel quality and relative quantity, but most importantly in the overall rhythm. In the case of Arabic learners with an RP English background, there is yet one more shift in vowel quality in the form of replacing some long vowels of Arabic by English diphthongs as in words with a long vowel in pre-‘r’ position of a syllable such as in = /safir /, َ
= / samir / and < بير ك > kabir which are pronounced as َ َ
/ sfi /, < مير س > = / smi/ and < بير ك > kbi respectiveَ َ ly. 8.4. CONCLUDING REMARKS
Vowels play a very significant role in generating accent in crosslanguage situations. There are four main reasons for this role. First, vowel systems across languages can be drastically different. Second, vowels do not have a well-defined contact area during their articulation; they are the result of a configuration ra-
132
PRONUNCIATION IS IN THE BRAIN
ther than actual contact. Usually, it is more difficult to form a configuration than to execute definite contact. Third, except for the configuration of the lips, the other two constituents of vowel production, namely location and size of the narrowing, do not yield themselves easily to precise visual, kinesthetic and proprioceptive assessment. Fourth, vowels carry the weight of stress within words and determine the nature of the overall rhythm. Unfortunately, in the traditional teaching of pronunciation, in general, and accent, in particular, there has been more emphasis on consonants than on vowels, and this is partially the reason behind the ineffective approach of many instructors, especially those who depend on the so-called ‘phonics approach’ to teaching pronunciation.
CHAPTER 9: EXAMPLES OF CROSS-LANGUAGE ACCENT-CAUSING SUPRASEGMENTALS 9.1. A DESCRIPTION OF THE MOST SALIENT FEATURES OF SUPRASEGMENTALS
For many people, especially those without any linguistic orientation, the most natural categorization of sounds is into consonants and vowels. However, the design of human speech is too structurally and systematically complex, intricate and diversified to be straitjacketed in the dichotomy of consonants and vowels, which usually represent short segments of speech. Obviously, the use of the attribute ‘short’ for some segments (individual segments) implies the presence of units, which represent ‘long’ segments (relevant to more than one segment or a stretch of segments), which are often known as suprasegmentals. Without suprasegmentals, segmental features alone would not suffice to carry the complex and open-ended communicative message of human speech. A combination of segmental features and multilength suprasegmentals generates tremendous structural and systemic diversity of sound units, which, in turn, account for the multi-layered and multidimensional construct of human speech. A relevant aspect of the study of suprasegmentals is to decide what a ‘long segment’ is. The response to such a question depends on the linguistic school and perspective one follows and the targeted linguistic refinement. In the early works of structural linguistics, the primary emphasis was on stress and intonation; duration (length) and juncture were also occasionally treated. In the tradition of prosodic analysis, of the London school of linguistics, the longer segments called prosodies are not necessarily confined to the traditional stress, intonation, duration and juncture. A prosody, according to prosodic analysis, may represent any feature that extends over more than one segment or pervades throughout a stretch of segments. For instance, the 133
134
PRONUNCIATION IS IN THE BRAIN
‘ized’-suffixed sound phenomena such as labialized, (lip rounding) nasalized, velarized, palatalized, pharyngealized may all be handled as prosodies. For a simple and straightforward illustration of the nature of prosody, let us examine the differences between the sounds of the following minimal pair: = [] vs. = []. A standard phonemic approach will identify three phonemes in each word with the difference being confined to the vowel element. Unlike the phonemic approach, prosodic analysis will identify three segmental units in each called phonematic units and one prosody represented by the lipspreading in the former and lip-rounding in the latter each of which pervades throughout the whole word; hence, both lip spreading and lip-rounding are prosodies or suprasegmentals. With this broadening of the domain of suprasegmentals, the phonetic and/or phonological relevance of suprasegmentals will not only be associated with syllables, as the shortest units in speech, and the sentence as the longest unit, but also will certainly include any stretch of speech or discourse. This trend is highly consistent with the recent emphasis on the study and teaching of language at discourse level. Consequently, the teaching of pronunciation with the inclusion of suprasegmentals, in general, and the inclusion of articulatory settings (Honikman, 1964) or phonetic settings (Laver, 1980; 1994), in particular, creates a primary shift in emphasis and direction. The significance becomes even greater at higher levels of proficiency acquisition. In special cases when accent acquisition and accent reduction are targeted, orientation in suprasegmentals and articulatory settings are extremely vital; in fact, they become inevitable for refined pronunciation. Generally speaking, the formal study of suprasegmentals receives attention only after having exhausted the study of the segmental features. In the beginning, the segmental elements (consonants and vowels) receive more attention because they are more tangible by virtue of their easily identifiable nature. They have become even more tangible and identifiable in languages that have been reduced to writing. In writing, especially an alphabetic system, the major target has been the assignment of symbols to segmental sounds only. Few languages have cared to incorporate prosodic features into their orthographies. Today no linguistic description and study of a given language is com-
CHAPTER 9
135
plete and coherent if its prosodic aspects are left untouched. This advancement in the description of languages has led to the application of the findings in different L1 and L2 language instruction situations. It is interesting to note about language education in the United States that second and foreign language education and instruction are far more linguistically geared than native language instruction (i.e., English language arts). Perhaps two reasons may account for all or part of the discrepancy. Firstly, native language acquisition is completed in a subconscious, effortless and automatic manner due to ample exposure to authentic context-embedded and situation-embedded language materials. Secondly, under normal native language acquisition by normal individuals, there does not seem to be much need for formal linguistic intervention and support. Unlike native language acquisition situation, second and foreign language learning situations, especially with adult learners, may require more formal and linguistics-based intervention strategies to better systematize and enhance the learning process. This latter fact may account for the use of more linguistics-oriented teaching materials and textbooks in ESL and bilingual education instruction than in English language arts instruction (at least in the U.S.). For instance, English vowels are taught as if they are five in number and occasionally six when is added. Consonant clusters (the so-called ‘blends’) may still be determined on the bases of letters rather than sounds. Letter (grapheme), sound (phoneme) and letter-name (nomeneme)1 identities are easily mistaken for each other. Phonics is a mere letter-based technique that fails to handle proper teaching of pronunciation. Many of those misconceptions are less frequently encountered in ESL and bilingual language materials perhaps because those two disciplines have developed in close connection with applied linguistics.
1
This term was coined after the patterns of phoneme and grapheme based
on the Latin root ‘nomen’ (name) to designate ‘letter-name’ (Odisho, 2004)
136
PRONUNCIATION IS IN THE BRAIN
9.2. STRESS AND RHYTHM
Stress may have different interpretations from the speaker’s or the listener’s standpoints. When the speaker’s activity in producing stressed syllables is in focus, stress may be defined in terms of greater effort that is exerted in the production of a stressed syllable as opposed to an unstressed one (Lehiste, 1970). When stress is defined from the listener’s standpoint, the claim is often made that stressed syllables are louder than the unstressed syllables (ibid). This is why Ladefoged tends to think that stress can always be defined in terms of something a speaker does while it is difficult to define it from the listener’s point of view (1982). To avoid those complications, it suffices to deal with stress in terms of greater or lesser physiological effort by the speaker and greater or lesser prominence by the listener who assesses prominence as the overall index of greater loudness, length and higher pitch. Stress is, hence, primarily the result of greater physiological effort exerted by the speaker at a certain point within a polysyllabic word and at repeated points within the flow of speech. A greater respiratory effort makes a given syllable more prominent and with the decrease in this effort, syllable prominence diminishes. A realistic division of the prominence continuum is to identify three degrees of prominence to be associated with the trichotomy of weakly stressed, medium stressed and strongly stressed syllables. However, the dichotomy of unstressed and stressed syllables has customarily been more dominant. The term ‘unstressed’ is only figuratively employed to subsume the first two degrees of stress because the term is literally meaningless—no portion of speech is produced without physiological effort; consequently, every portion should have some prominence. In other words, the unstressed syllables stand for the portions with minimum stress. It is possible to distinguish between stress assignment within a word and within a sentence because within the latter it is likely for words to undergo a shift in the location of stress or to emerge with partial stress only. Within a word a certain syllable sounds more prominent in relation to others, while in a sentence certain word or words sound more prominent in relation to the rest. The former is called word stress and the latter sentence stress.
CHAPTER 9
137
Languages differ in the manner they use word stress and sentence stress to signal linguistic/nonlinguistic variations. Some languages show a strong tendency to retain the stress on a certain syllable within the word regardless of the syllabic structure and the number of syllables. Obviously, in such cases stress becomes highly predictable. Czech words tend to have stress predominantly on the first syllable irrespective of the number of syllables (Ladefoged, 1982), whereas Turkush tends to place stress on the last syllable. In other languages, stress changes its place according to several factors: foremost of all are the number of syllables, their internal structure and arrangement within a word, the grammatical category of words and their status as native or loan words. To facilitate the rules of stress assignment, linguists use the classificatory terms of ultimate, penultimate, and antepenultimate to identify the structural location of stress. If no rules can be formulated or if the rules can only capture certain instances leaving the rest of the instances unaccounted for without some ad hoc rules, then the predictability of stress becomes less likely and its role as a distinctive feature between the lexical items more striking. If stress is highly predictable, its function is primarily that of determining the rhythm and the overall pronunciation though it still can have a demarcative function, i.e. it helps to signal the word boundary (Hyman, 1975). In languages whose stress placement resists straightforward predictability, the function of stress is no longer confined to pronunciation and demarcation; it can assume a wide and diversified range of lexical and grammatical functions. The distribution of stressed and unstressed syllables within a language determines its rhythm. The traditional view is that rhythm in languages follows the dichotomy of stress-timed or syllable-timed (Adams 1979; Dauer, 1983). However, there are some linguists who tend to think that the concept of a dichotomy is too rigid a characterization to realistically portray the nature of rhythm in human language. Ladefoged (1982) states: “Perhaps a better typology of rhythmic differences among languages would be to divide languages into those that have variable word-stress (such as English and German), those that have fixed word-stress (such as Czech, Polish and Swedish) and those that have fixed phrase-stress (such as French)”. But since there is
138
PRONUNCIATION IS IN THE BRAIN
more than one factor that determines the nature of rhythm in a given language there is no compulsion to have one or the other of the stress-timed or the syllable-timed rhythmical bases (O'Connor, 1973). However, one needs to understand what is meant by a stress-timed or syllable-timed rhythm. A stress-timed rhythm is the one in which stressed syllables tend to recur at regular intervals of time and the syllables vary considerably in length depending on whether they are stressed or unstressed. On the other hand, a syllable-timed rhythm is one in which each syllable tends to retain more or less the same duration regardless of stress (Adams, 1979; Ladefoged, 1982; Roach, 1983). In the stress-timed rhythm, only syllables receiving the primary stress stand out prominently, while the unstressed syllables are reduced and compressed in time to become far less prominent. Unlike such uneven distribution of prominence, in the syllable-timed rhythm, all syllables, stressed or unstressed, receive a relatively even prominence; syllables take approximately the same time, and the overall length of an utterance depends on the number of syllables involved. In other words, in this latter type of rhythm there is hardly any noticeable reduction in the prominence of the unstressed syllables. It is in light of the above-mentioned characteristics that English is said to have a typically stress-timed rhythm, whereas Spanish is said to have a typically syllable-timed rhythm. At this stage, it is interesting to consider the possibility of an underlying connection between stress and rhythm type and the type of the vowel system in a given language. In English, vowel quality and quantity fall heavily under the influence of stress and this interaction is part of the dynamics of the vowel system. The location of stress and its strength within the word or sentence greatly influence the vowels both qualitatively and quantitatively. In syllables with a primary stress, vowel quantity (length) reaches its maximum and its quality is very distinct. In syllables with a secondary stress or a weak stress, both quality and quantity of vowels are reduced drastically. In unstressed syllables, almost all the English vowels can be reduced in both quality and quantity to the shortest vowels namely, [] or [] (Dalbor, 1969; Ladefoged, 1982; Dale and Poms, 1985). Such a qualitative and quantitative process of vowel reduction is a typically characteristic feature of English, but very un-
CHAPTER 9
139
characteristic of a language such as Spanish. The above exposition seems to point in the direction of the plausibility of a connection between the centripetal vowel system and the stress-timed rhythm type, on the one hand, and a centrifugal vowel system and a syllable-timed rhythm type, on the other hand. In case of Spanish, vowels tend to retain their relative quality and quantity, regardless of stress, and if they never undergo any schwaization or even vowel reduction, how does one expect the syllables to be manifestly different in length and prominence? A univalent system of vowels should undoubtedly yield a temporally uniform and univalent type of syllables, which is typical of a syllable-timed rhythm as in Spanish. By contrast, with a multivalent system of vowels combined with a very pervasive tendency toward schwaization, one should expect multivalent types of syllables as is the case typically of English stress-timed rhythm. To paraphrase it differently, if the vowel system of a given language does not maintain long/short or tense/lax phonological contrasts, if it does not have a schwa as part of its phonological system, and if it does not tolerate schwaization or a tangible degree of vowel reduction, it implies the presence of a synchronic constraint on the extent to which stress can alter the quality and/or quantity of its vowels. It is true that stress in Spanish can somewhat change vowel quality and quantity, but the change will still be confined to the phonetic domain. Thus, in Spanish, vowels may phonetically be slightly longer/shorter or tenser/laxer, but the absence of phonological contrasts based on those features will deny the language the potential for creating syllables that are significantly different in length and prominence. This argument in favor of binding the rhythm type to the vowel system does not mean that the vowel system is the only factor that determines the rhythm type in languages; undoubtedly, other factors such as syllable structure (Dauer, 1983), fixed/variable word stress and word/phrase stress (Ladefoged, 1982) are relevant in this regard. The most significant conclusion drawn from the preceding discussion is that the nature of the vowel system should be a factor to be seriously reckoned with in the typological classification of speech rhythm, its study and teaching. Those modifications to the traditional syllabletimed and stress-timed rhythm types amount to a major change
140
PRONUNCIATION IS IN THE BRAIN
that should be seriously considered in the development of the approach and the techniques of teaching pronunciation and the remediation of accent. 9.3. TONE AND INTONATION
In speech, there is always a continuous change in the fundamental frequency, which is auditorily realized as pitch also known as the melody of speech. Languages use pitch in two essentially different ways. If it signals semantic differences between words, the languages are called tone languages. In Chinese, for instance, the basic unit < ma > may have more than one meaning depending on the rising, falling, falling-rising or level pitch it carries. It is the pitch difference (or toneme to be consistent with other ‘eme’-suffixed linguistic labels such as phoneme and grapheme) that triggers the semantic difference. Many of the African and Asian languages fall into this category. Thus, a given language, whose pitch pattern has no specific role in the semantic shaping of words, but is rather used to signal a combination of syntactic, semantic and attitudinal features of the utterance, is called an intonation language. 9.4. BASIC PITCH PATTERNS
Pitch patterns are very vividly explained and schematically represented in terms of pitch height and pitch direction (Laver, 1994). The labels used to refer to different pitch levels are high, mid, low, mid-high, mid-low etc. Pitch contour refers to the shape and direction of pitch yielding different shapes such as rise, fall, level, rise-fall, fall-rise etc. A combination of attributes from both pitch-height and pitch direction produce the basic pitch patterns whose recognition and production should be an essential goal in any program for training in phonetics, in general, and pronunciation, in particular. No teaching of tone and intonation will be effective without the mastery of the basic pitch patterns. An interesting aspect of the basic pitch patterns is the cross-linguistic commonality in their nuances. The falling pitch patterns, both low and high, have the general purpose of expressing an utterance with a sense of completeness so that the attention of the listener is no longer required inasmuch as that particular utterance is con-
CHAPTER 9
141
cerned. A high fall usually indicates a more vigorous and determinate notion of completeness and finality than the low fall does. The rising pitch patterns, unlike the falling patterns, imply a sense of the incompleteness of the utterance as if further information is expected from the speaker or a response is necessary on the part of the listener. The basic pitch patterns are:2
Figure 9.1. Basic pitch patterns in human speech in general.
9.5. CONSONANT CLUSTERS
Consonant clusters or so-called consonant blends should be distinguished from consonants that occur juxtaposed to each other. The former is a combination that should structurally belong to one syllable and is pronounced as one intact piece. The latter is a combination that is spread over two syllables. Take the word which has a combination of two consonants, 2
For a demonstration of tone patterns in Chinese go to section 14.2.1/f.
142
PRONUNCIATION IS IN THE BRAIN
but it is not a cluster because the belongs to the first syllable and belongs to the second. Compare the of with the
sounds in perception and recognition following the steps below: a) Demonstrate the pair vs. several times and highlight the difference to be picked up auditorily. It is quite acceptable to exaggerate the difference somewhat for the sake of clarity. b) Ask for volunteers from among the learners to do what you did. c) If you notice that learners are having difficulty or even doubt in telling the difference, move to the next step.
192
PRONUNCIATION IS IN THE BRAIN d) Move to the visual sensory modality. Place a flimsy piece of paper in front of your mouth and repeat the demonstration. The paper should flutter visibly with , but will hardly flutter in the case of . e) Ask for volunteers to do what you did. f) Arrange the class in pairs the members of which face each other while conducting the flimsy paper experiment
After such joint demonstrations, some of the learners should be able to perceive and recognize the difference because of the combined auditory and visual modalities of instruction. It is after this phase of orientation, that the instructor can begin the production phase of such phonetic materials from the targeted language such as [p:r] (odd numbers) vs. [p:r] (lambs) in Assyrian language. One can take the exercises one step further by involving the other two plosives of [t] vs. [t] and [k] vs. [k]. In summary, the above orientation began with listening to the targeted phonetic phenomenon. The auditory modality was supported by visual modality to help in two ways. First, elevate natural listening to discriminative listening. Second, use both listening and discriminative listening to activate the cognitive involvement of the brain in the process of internalizing the phonetic difference. 12.3.2. Developing Learning Strategies
Learning strategies are equally important to teaching strategies. When the instructor dominates the classroom situation and reduces the learners to mere listeners with minimum interaction, this is an old fashioned teaching strategy. There is always a wide variety of learning styles that should be considered and encouraged some of which will be considered below. 12.3.2.1. Discover Learning Strategies
When it comes to promoting learning strategies, the instructor has to discover the learning styles of as many learners as possible. To achieve this, he has to be very alert and observant. The discovery procedure usually takes place when the instructor presents one targeted problem with diverse auditory, visual and tactile-kinesthetic modalities and styles and monitors the reac-
CHAPTER 12
193
tion of students to each modality of the presentation. Once the instructor senses that a given learner responds to a given sensory modality, he should reinforce that in future attempts. 12.3.2.2. Consider Cultural Diversity of Learners
Foremost of the requirements for teaching pronunciation is the audible oral exercise that usually requires the presence of one or more persons. This face-to-face scenario can become intimidating, especially with some learners whose background culture is conservative and sensitive to errors and failures in public and in the presence of other learners, especially those who are strangers to them. In fact, this is even more relevant to female learners of those societies. The instructor should carefully consider the psychological and cultural differences between learners and encourage all of them to participate in class activities. This participation cannot be achieved without the promotion of mutual trust and respect for each other’s culture traditions as well as the tolerance of errors in early performance trials. 12.3.2.3. Encourage a Relaxed Attitude among Learners
Some aspects of teaching pronunciation can be intimidating as well as humorous. This is especially true with some very unfamiliar sounds. To have fun while experimenting with sounds should be part of the classroom culture of teaching pronunciation provided it is controlled fun lest it should overflow and interfere with class management. Learners should be instructed that failures and mistakes are part of the learning process. All learners should be encouraged to build up mutual trust between themselves. The more the interaction grows between the learners, the greater the discovery of each other’s learning styles. 12.3.2.4. Move between Individual and Group Learning Styles
It was pointed out previously that all learners have their own personal strategies of learning. In a classroom environment, however, personal strategies should be complemented with group strategies. It is natural to have group strategies because teaching and learning take place collectively in a classroom setting. One way of promoting alternative strategies to one’s individual style is through group and cooperative learning. This re-
194
PRONUNCIATION IS IN THE BRAIN
quires forming groups or teams led by facilitators from amongst the learners. This all should be done under the guidance and supervision of the instructor. 12.4. CONCLUDING REMARKS
In any successful classroom environment, there should be a balance between teaching and learning strategies because they are complementary in the mission of educating learners. Furthermore, a successful classroom environment should afford a threeway learning process: a) Learner from instructor; b) Learner from learner; and c) Instructor from learner. The first two learning styles are self-explanatory, whereas the third needs some further elaboration. With regard to the latter style, I am reflecting on my personal classroom experience which has been the richest source of gaining hands-on experience in instruction. There have been repeated instances when I corrected myself or improved my instruction because of direct or indirect feedback from learners. Let me cite the following example—not related to pronunciation but with a genuine linguistic connection. In one English language class, I was teaching spelling through a multisensory and multicognitive approach. I selected the word ‘grammar’ as an example of a commonly misspelled word often as ‘grammer’ or ‘gramar’, whereas its correct spelling is . I brought the following facts to the attention of the learners. I suggested the following mnemonics:
Remember 7 letters = ;
Remember 1 single letter plus 3 pairs = + 2r + 2a + 2m = grammar;
Remember mirror image: ram + mar;
Remember ‘ram’
At this stage I exhausted my sensory and cognitive mnemonic hints that help with the retention of the correct spelling of the word . All of a sudden, one of the learners, an artist, asked to go to the board to demonstrate an addition. He quickly drew two sketches of two rams facing each other as if preparing for a headbutt. His quick sketch was to reflect the mirror image of ‘ram’ and ‘mar’ in . It was an ex-
CHAPTER 12
195
cellent and creative idea coming from an artist. I learned that and added the head-butting rams to my sketches.
CHAPTER 13: EXEMPLARY APPLICATIONS OF ACCENT REMEDIATION TECHNIQUES 13.1. INTRODUCTORY REMARKS
In this chapter, we will try to apply the approach detailed in the previous chapters with as many cognitive, sensory and educational teaching and learning strategies as relevant to the sound or sound phenomenon selected for elaboration. To be sure, every sound or sound phenomenon can be a source of difficulty for some learners of a given L2. Consequently, there has to be a limited selection of cross-language problems that encounter large numbers of L2 learners. Due to the international popularity of English as an L2, the application of the strategies tends to lean strongly in that direction with a reversal of roles, i.e., with native speakers of English tackling other languages. In many instances, teaching a given sound whether a vowel or consonant implies teaching a natural class of sounds. For example, teaching an aspirated/unaspirated consonant pair implies teaching an aspirated vs. unaspirated class of consonants usually /p t k/ vs. /p t k/; nevertheless, the two classes can involve other plosives such as palatal as well as affricates, etc., as is the case in Modern Assyrian (Neo-Aramaic). It is relevant to point out that some selected sounds, such / / may be difficult for many L2 learners with a wide variety of linguistic backgrounds, whereas other sounds may be difficult for learners with more specific linguistic backgrounds such as the distinction of /b/ vs. /v/ for Hispanic and Filipino learners of English. 13.2. TECHNIQUES FOR TEACHING SELECTED CONSONANTS
The consonants or consonantal phenomena selected for demonstration include the labiodentals, especially /v, /; the interdentals pair / /; retroflex sounds for learners of sub-continental 197
198
PRONUNCIATION IS IN THE BRAIN
Indians languages and the reversal of retroflexion for most subcontinental Indians learning L2s. It is important to bring to the attention of the instructor and learner that this section will contain sets of strategies that are procedurally, though not specifically, typical of teaching most sounds according to a multicognitive and multisensory approach. This entails that not all sounds yield themselves to the same sensory modalities and teaching techniques; however, all difficult sounds require some sort of cognitive orientation. Consequently, most of the details expounded in this section will not be reproduced for all sounds taught. 13.3. TECHNIQUES FOR TEACHING LABIAL-DENTAL SOUNDS
The voiced labialdental fricative [v] and the voiced labialdental approximant [] are more marked (less common) sounds than the voiced bilabial plosive [b] and the labialvelar approximant [w]. However, because of the generic labial nature of all those sounds as well as the absence or presence of some of them in given languages, any instructor of English as L2 will come across scores of learners who will demonstrate serious difficulties in pronunciation amounting, at times, to phonological accent while others will indicate at least a certain degree of phonetic accent. For instance, Hispanics typically replace [v] with [b], whereas Persians, Turks and Assyrians, among others, replace [v] with either a [w] or a []. I have an Assyrian friend from Iran who pronounces and as ‘Wote’ and ‘Harward’. In some other Assyrian dialects the [w] may be replaced with a labialpalatal approximant []. If one works with Hispanic students, he will readily notice that the mispronunciation of [v] is pervasive even among some individuals whose proficiency in English is otherwise excellent. Obviously, the primary reason is the absence of a /v/ sound in the phonology of Spanish. Pedagogically, the persistence of the problem even with welleducated Hispanics with high competency in English may be attributed to two factors. a) The mispronunciation has not received much attention from instructors at an early stage of learning; b) The instructors did not follow some effective techniques in teaching it. To put it more bluntly, the instructors did not have the know-how of effective remediation of learners’ mispronunciations. Most probably, if ever, they might have fol-
CHAPTER 13
199
lowed the mechanical procedure of ‘repeat-after-me’ which is often less effective with adults due to psycholinguistic insensitivity or deafness explained in earlier chapters. In what follows, some strategies are put forth to develop an effective set of procedures to overcome the problem of /b/ vs. /v/. The strategies typically reflect different cognitive and sensory modalities. a) Cognitive Orientation: Prepare the learners mentally (cognitively) to recognize the existence of the problem and its seriousness because it leads to serious phonetic and phonological accent. The preparation requires the following steps: 1) Instruct them to be ready to accept the problem and be willing to pay utmost attention. 2) Tell them they will certainly manage the pronunciation. 3) Tell them to watch your facial gestures, especially those of the mouth and recognize the difference in the pronunciation of [b] vs. [v]. In fact, to dramatize the postural difference in the articulation of the two sounds, you may call the [v posture a ‘dogface’ because when one assumes the posture, one looks like an angry dog ready to bite. In contrast, you may call the [b posture a ‘tight-lip face’ one since the lips have to come together tightly for the sound. The dramatization of the articulatory facial postures for the sounds oftentimes functions as a humorous, albeit robust and concrete mnemonic to remind the learners of the required articulatory differences. 4) Demonstrate the pronunciation of the sounds in selected minimal pairs of words for which the difference in meaning is easily noticeable and, perhaps, even funny or embarrassing, such as vs. ; vs. or vs. (Figure 13.1).
200
PRONUNCIATION IS IN THE BRAIN
Figure 13.1. Notice the difference when /v/ is replaced with /b/. 5) Use colors and pictures or any other audio-visuals to highlight the difference that results from substituting one sound for the other. 6) Ask learners to watch carefully your facial gestures, especially your mouth and lips, while you slowly and distinctly demonstrate the production of the two sounds. To put it differently, ask them to watch the dogface posture for [v] and the tight-lip-face posture for [b]. 7) While you do all the above, carefully watch the facial gestures of the learners. If you notice that learners’ faces seem attentive and serious then you have to be sure that the learners are in a mode of thinking. In other words, they are trying to cognitively grasp the difference between the two sounds.
b) Auditory Orientation: Go back to the minimal pairs, number each member of the pair as #1 and #2 then produce each member of the pair and ask learners to identify the word as #1 or #2. Do this demonstration with your mouth covered with a piece of carton to prevent lip reading and easy guessing. Another major difference between the two sounds is that [v] being a fricative sound is sustainable (can be prolonged), while [b] being a stop is unsustainable (cannot be prolonged). If some learners still experience some difficulty in perceiving and recognizing the difference between the sounds, then go to the next step. c) Visual Orientation: Remove the carton and pronounce the two sounds quite consciously while exaggerating the bilabial (upper & lower lips or the so-called tight-lip-face) posture for [b] and the labial-dental (lower lip and the upper teeth or the so-
CHAPTER 13
201
called dogface) posture for [v]. Put the learners in pairs facing each other and ask each member of the pair to perform the articulatory postures for the two sounds, while the other learner is observing. Allow them to turn-take on this performance. d) Kinesthetic/Proprioceptive Orientation: Ask the learners to carefully watch your demonstration of the two sounds with distinct performance of their articulatory postures. Stick with one of the sounds and repeat its articulatory posture while saying the name of the sound (i.e., its letter-name) then do the same with the other one. In other words, pronounce the sound [v] or [b] several times followed by or . Ask them to impersonate what you have been doing with emphasis on the need to develop a kinesthetic (tactile) and proprioceptive (inner) sensing of the articulatory contacts made for [v] and [b]. To rephrase the latter statement, learners must be asked to sense the contact of the two lips for [b] and the contact of the upper teeth and the lower lip for [v]. e) Cognitive Reinforcement and Internalization: The initial cognitive orientation is considerably reinforced by the following three sensory modalities of auditory, visual and kinesthetic/proprioceptive. The activity and performance conducted via each sensory modality plays a certain role in the joint reinforcement of the articulatory difference between the two sounds. Once the brain receives the input through each sensory modality it begins to process it and develop the impressions required for the neuronization (i.e., imprinting in neurons) of the two sounds as two different entities. f) Follow-up Procedures: 1. Obviously, the manner in which human memory functions should be taken into consideration. To put it differently, human internalization of an impression may just be for a short time or for a long time. Those different times were technically known earlier on as the sensory memory, shortterm memory and long-term memory. 2) It is quite natural and normal for the learner to be able to correctly articulate the targeted sound, but then forget it in a split second. This situation may be highly indicative of the
202
PRONUNCIATION IS IN THE BRAIN behavior of a sensory memory mode (i.e., perceive, produce and forget). 3) The above situation certainly tells you that more rehearsal is definitely needed which is often the case. 4) Return to more auditory, visual and kinesthetic/proprioceptive rehearsal to at least transform the impression of the sound into the short-term memory during which it undergoes a ‘remember-forget-remember’ process. This is an important stage in the learning/acquisition of the targeted sound because it, most likely, indicates the initial stages in the cognitive internalization of the sound. In other words, learner is actively engaged in a conscious and cognitive processing of the sound. 5) If after an active day of remember-forget-remember, the learner comes back the next day and has forgotten the targeted pronunciation do not panic. All that the learner needs is a refreshment of memory. 6) Once you notice that the learner produces the sound incorrectly, but then he instantaneously realizes the mispronunciation and rectifies it immediately, you should relax because the learner is most likely at the final stage of the correct internalization of the sound in the long-term memory. 7) After this stage, what the learner needs is more occasional rehearsal and practice to finally subconsciously internalize the sound for immediate and automatic retrieval.
13.4. TECHNIQUES FOR TEACHING INTERDENTAL FRICATIVES / /
These two sounds are extremely rare in many languages throughout the world; consequently they are typically classified as marked (unfamiliar; uncommon) sounds. Their rarity may be attributed to their association with the clinical speech problem known as alveolar lisp according to which individuals, especially young children, manifest the symptoms of such pronunciation problem by replacing the [s, z] with [, ], respectively. What matters here is the fact that in spite of the rarity of this pair, it is, nonetheless, of high frequency of occurrence in English and this renders the pair a major pronunciation problem for a large
CHAPTER 13
203
number of L2 learners of English from a wide variety of linguistic backgrounds. It has already been pointed out, in 7.2.1 above, that one of the most interesting aspects of the mispronunciation of this pair is that it is realized differently depending on the native language of the learner and its phonological system. Its replacement is usually with far more unmarked (common) pairs of sounds such as /t, d/ or /s, z/. Surprisingly, when teaching this pair of sounds, one can come across individual learners who are capable of pronouncing the pair, but are psychologically reluctant to do that in conversation because they feel they are lisping. This is a psychological observation that should be taken into consideration. Since the orientation for those who replace / / with the plosive /t, d/ will be different from the orientation of those who replace them with the fricatives /s, z/ there will be two sets of procedures of orientation. Below are, first, the orientation procedures and techniques for /t, d/ substitution. a) Cognitive Orientation: The instructor has to ask learners to carefully watch the articulatory postures and facial gestures concomitant with the production of [] and []. Carefully and dramatically emphasize how the tip of the tongue slightly sticks out at the biting edge of the upper incisors in the production of both of them. It is helpful to remind learners that in many cultures, children tend to stick their tongue out as a gesture of mocking. The articulatory postures for [] and [] are similar to the mocking gesture except in a very moderate and gentle manner. Just to exaggerate the visual, auditory and kinesthetic differences between [, ] and [t, d], carefully and slowly demonstrate the production of the latter pair to highlight the several sensory differences. Cognitively, the purpose of this comparison is to encourage learners to consciously think about the physical and articulatory production of the two targeted sounds and their unwanted substitutions. b) Auditory Orientation: You need to prepare suitable lists of minimal pairs for both[] vs. [t] and [] vs. [d] to orient the learners on the perception, recognition and production of the two pairs of sounds. The initial group of minimal pairs should be carefully selected not only to highlight the pronunciation differences, but also to highlight the semantic differences that result from the mispronunciation. One way to better high-
204
PRONUNCIATION IS IN THE BRAIN
light the pronunciation differences is to select monosyllabic words. Table 13.1, below represents some such minimal pairs: Word []
Mispronunciation
[t]
[]
[d]
Table 13.1. The [t, d] rendition of the English interdentals [, ] by different learners of English.
To further highlight the semantic difference ensuing from the substitution, select a couple of minimal pairs for pictorial (visual, figure 13.2) difference such as vs. . Bring to the attention of learners, that unfortunately the production of [] and [] may sound like a speech defect (alveolar lisp), but that is how the two sounds are in English.
Figure 13.2.Visual and semantic difference when substituting /t /for //.
c) Visual Orientation: In order to warm up the learners for the visual orientation, go back to the actual production of the
CHAPTER 13
205
pair /, / vs. /t, d/. This time use a paper to block the visual channel of guessing of the sounds through lip-reading. Next, remove the paper and pronounce [] and [] with a clear facial posture showing the tip of the tongue at the biting edge of the incisors. Then do the articulatory postures for [t] and [d] while drawing the attention of the learners to the disappearance of the tip of the tongue. d) Tactile-Kinesthetic Orientation: To help learners with this type of sensory orientation all that the instructor has to do is to direct learners to put the tip of the tongue at the biting edge of the incisors and repeat the contact for [, ] several times until the brain makes an impression of the contact. The brain may forget the sensation minutes, hours or days later, but it is easy to recall the impression with a couple of maneuvers. Word []
Mispronunciation [s]
[]
[z]
(letter)
(v.)
Table 13.2. The [s, z] rendition of the English interdentals [, ] by different learners of English.
After completing all the cognitive and sensory orientations for /, / vs. /t, d/, move to contrasting /, / with /s, z/. This contrast may be more challenging than that of /t, d/ simply because in this case there is only one distinctive feature of place of articulation that separates them (i.e., interdental vs. alveolar) instead of two in the case of /t, d/ (i.e., place of articulation: interdental
206
PRONUNCIATION IS IN THE BRAIN
vs. alveolar coupled with manner of articulation: fricative vs. plosive). The instructor should use the general procedures in a, b, c and d above with emphasis on the place of articulation using the minimal pairs in table 13.2, above. In addition to the above guidelines and examples, instructor should attempt to demonstrate the considerable difference in meaning when the /, / are replaced with /s, z/ similar to the example of vs. . A good example would be the contrast between 1 and .
Figure 13.3.Visual and semantic difference when substituting /s/ for //.
After the different cognitive and sensory orientations, the instructor has to put his work to the test through designing and implementing some perception and cognition exercises leading finally to production in the following manner: a) Perception: Demonstrate the /, / pair several times for perception by learners with /t, d/ substitution; b) Recognition: Move to the recognition assessment by numbering the sounds /t, d, , / as 1, 2, 3 and 4; produce a list of ten sets of the four sounds arranged in a random order; pronounce the ten sets slowly and methodically while asking the learners to identify the sounds according to their assigned numbers on the worksheet; for example, if the sounds were pronounced with this order: /t, , d, / then the set of numbers should be: 1, 4, 2, 3; collect the worksheets to assess the general accuracy. If the matching is considerably satisfactory, proceed to the production phase, if not repeat the exercise asking learners to print their names on the worksheets. The latter addition of
1
Point to the head while pronouncing it.
CHAPTER 13
207
names will make the learners more alert and attentive; besides, it will also identify the learners who need more orientation. c) Production: Ask for several volunteers to demonstrate the pair /, / while in their seats; if all successful, ask the same volunteers to appear before the class and demonstrate the sounds while the rest of the learners listen to the demonstrations and watch the facial gestures of the performers; ask the volunteers to produce minimal pairs that you provide such as vs. and vs. and so on with contrasts of /, / vs. /s, z/. 13.5. TECHNIQUES FOR TEACHING TENSE (LONG) VS. LAX (SHORT) VOWELS
Obviously, the features of quantity and quality are in many instances too intertwined to be separated and autonomously evaluated and described. The case of the English vowels in < sit > vs. < seat > is typical in this regard. Even though many authors and in many instances, handle the relationship of those two vowels as short vs. long, the relationship is too complex to be glossed over as short vs. long; it involves a feature of lax vs. tense accompanied by a difference in quality. This complex feature combination becomes an instructional reality when adult L2 learners of English in whose languages there are no short vs. long or lax vs. tense vowel distinctions. Typically in this regard are speakers of Spanish, Italian, Russian, Greek, Filipino, among many other languages. The first time I came across such a major pronunciation difficulty for Hispanic learners of English was some three decades ago when I was teaching an ESL class at Loyola University Chicago. In order to demonstrate to my students the so-called short vs. long vowel contrast in English, I wrote the words and on the board and asked the best student in class to pronounce them. To my utter surprise, I heard him produce the same pronunciation for both in the form of [ʧp]. I asked him to repeat the pronunciation and the result was no more than [ʧp] for both. As a linguist, I realized for the first time that apparently there was only one version of an vowel in Spanish and that the language does not have a [] consonant sound. As an instructor, I paid utmost attention to such problems. In this case, the incident became a focus of part of my future research.
208
PRONUNCIATION IS IN THE BRAIN
Let us consider more examples of teaching Hispanics the vowel system of English. a) Cognitive Orientation: In an experiment conducted with adult Hispanic learners of English at a beginning proficiency level, the learners were asked to pronounce the following minimal pairs:
The overwhelming characteristic of most of the tokens obtained was the failure to distinguish the vowel difference within each minimal pair. The vowel that dominated those words sounded shorter than the English long one and longer than the short one although it shared a degree of tenseness with the English long one. The failure to distinguish minimal pairs based on vowel quantity is certainly the most major pronunciation problem for Hispanic learners of English for two reasons. First, they have no cognitive perception and recognition of the quality and quantity differences between the two sets of English vowels because such differences do not exist in Spanish. Second, without the realization of such a vocalic feature, thousands of words in English may be semantically confused some of which may be socially very embarrassing because of obscene or vulgar connotations. This is why some Spanish, French or Russian learners of English avoid words such as < beach, sheet, winner, keys> etc. Right at the outset, the instructor should bring this major vocalic difference to the attention of the learners highlighting the considerable phonetic and semantic differences. He should also prime them for the cognitive realization of the difference through the following steps: 1) Cite some monosyllabic minimal pairs and highlight the difference in both pronunciation and meaning such as: vs. vs.
CHAPTER 13
209
vs. 2) Demonstrate the above minimal pairs again and highlight the difference especially in the shape of the lips which tend to be more stretched sidewise in the latter group of words than the former. 3) Try to insert the Spanish word (without) in between the English minimal pair and to further underscore the phonetic difference. 4) Transcribe the difference between the three words phonetically in the following format: English: = [sn]; Spanish: = [sin]; English = [sn].
The Spanish vowel should be transcribed as [] with [i] indicating the quality of tenseness and the single dot indicating the medium or half-length somewhere between the two English vowels [] and [i]. The transcription is meant to signal both quality and quantity (length) differences. Schematically, the instructor can express the quantity (length) and quality (tense vs. lax) by the thickness of the dark band as demonstrated below: English Vowels Schematic Representation = [p] = [p]
The expected Spanish rendition of (without) will look like the following which is shorter than the long English vowel, but slightly less tense. On the flip side, it is longer than the short English one though somewhat more tense. The following strategies are suggested to handle such vocalic multiple-feature differences. b) Perception: 1) Produce a simulated Spanish vowel in the context of, for example, and transcribe it as [n]. Model this simulated pronunciation several times while learners are listening.
210
PRONUNCIATION IS IN THE BRAIN 2) Embed the third item (i.e., the simulated Spanish examples) in between the English minimal pair of = /sn/ and = /sn/. It is very convenient to select monosyllabic words to avoid any perceptual interference with bisyllabic or multisyllabic words. Number the three items as 1, 2 and 3 as in table 13.3.
English
Transcription
Spanish
English
Transcription
[t]
[t]
[t]
[bt]
[bt]
[bt]
[st]
[st]
[st]
[pl]
[pl]
[pl]
[rʧ
[rʧ
[rʧ
Word
#1
Rendition #2
Word
#3
Table 13.3. Simulated Spanish rendition of English minimal pairs with [] and [] vowels. Model each of the above triplets very carefully and ask students to listen, sense and reflect on the process. 3) Cite more English minimal pairs such as: [i]
[i]
and simulate the expected mispronunciations by Hispanic learners i.e., [bd], [tn] and [lp] 4) Encourage learners to practice what Catford calls ‘silent introspection’ with emphasis on sensing the vowels rather than hearing them. According to Catford, whenever one makes sounds aloud, the auditory impression tends to mask or override the sensations of muscle movements and other proprioceptive sensations. Certainly, student awareness of these proprioceptive sensations can be a useful adjunct to making the articulatory adjustments necessary for learning the articulation of new sounds (Catford, 1994).
CHAPTER 13
c) Recognition: 1) Select the triplet [bt], [bt], [bt] from table 13.3 in Perception and number the items 1, 2, and 3. Record them randomly each of which repeated twice in, at least, ten to fifteen attempts. Play the recordings back one attempt at a time with a few seconds of pause between each attempt and ask the learners to mark the items as 1, 2 or 3 on a specially prepared worksheet. 2) Give the learners the key to the correct answers, ask them to identify the errors and notice the tokens which were with the highest percentage of inaccuracy. The results may be very significant for further design of exercises and drills. If learners misidentify the first token, such a result is expected because the first token may be uttered more emphatically by the reader prior to settling to a normal mode of pronunciation. Besides, with the first token, learners are, so to say, taken by surprise when they have not yet developed a mental or psycholinguistic yardstick for the estimated evaluation of the samples. 3) Ask learners to return all the worksheets of the first trial then ask them to prepare for a repetition of the exercise. Usually, a second and a third trial are much better than the first one; more exposure creates more familiarity and both lead to more confidence and better focus. 4) Select as many minimal pairs as in perception and mark the columns as #1 and # 2 then continue to pronounce the items arbitrarily and ask learners to identify them as #1 or #2. 5) Create carrier sentences and ask learners to place the appropriate member of a minimal pair as in perception/3 in the blank then read the sentences once or twice: Example: select either mill or meal and place it in the sentences below: A
is the place we turn wheat into flour.
A breakfast is a
in the morning.
211
212
PRONUNCIATION IS IN THE BRAIN d) Production: 1) Ask volunteers to repeat the modeling of items in perception/2 above after as many times as necessary. The repetition should be instantaneous. 2) When some learners excel in the impersonation or production of the targeted sounds, allow them to replace you, as the instructor, in modeling. All learners should to be given the opportunity to participate. Preferably, learners should model while seated in their places among the students; this setting creates a more learner-friendly situation.
13.6. TECHNIQUES FOR TEACHING VOWEL REDUCTION
What is vowel reduction, anyway? A reduced vowel becomes shorter in length (quantity) and/or moves in the direction of a neutral vowel, typically a schwa []. Vowel reduction is used here to refer to the overall diminishing in the length (longer to shorter) and quality (from markedly distinct features to less distinct ones) of the vowel. For instance, the change of [] in strong form of < than > = [] into [], its weak form, is a change from a longish and more distinct vowel into a very short and qualitatively less distinct one. Earlier on, the English vowel system was identified as a typical centripetal system as opposed to the Spanish centrifugal one. Since vowel reduction in English leads predominantly to schwa-type vowels, English is best described as a schwa vowel system, whereas Spanish, without any vowel reduction, is best described as a schwaless vowel system. At this juncture, the question of what the characteristics of a schwa are becomes unavoidable. Auditorily, the schwa is the most obscure vowel. Visually, the lips are in the most neutral position: not spread, not rounded with medium opening. Kinesthetically, the whole vocal musculature is relaxed with no sensation of an extreme forward, backward or upward movement of the body of the tongue. Features such as those make the schwa a very elusive sound and a difficult one to teach. With the elusive articulatory maneuvers involved in the production of a schwa, the primary dependence will be on the auditory channel in the form of: (perceive) listen, recognize and produce. Nevertheless, in order to maximize the effectiveness of the instruction, all the other sensory channels have to be brought into play. This should be
CHAPTER 13
213
reinforced by considerable non-linguistic and non-verbal activities and by contextualizing the schwa in words and longer pieces of speech. The typical action word to describe the instructor’s productive action will be ‘demonstrate’; the learner’s action word will be ‘impersonate’. The word imitate will be avoided because it usually connotes mechanical action whereas impersonate implies watching, contemplating and purposeful production. Indeed, vowel systems that tend to be tense (centrifugal) are usually without a schwa unit; therefore, its speakers face great difficulty in neutralizing the vowels. This simply implies that teaching the production of a schwa requires concentrated sensory efforts until the sound is internalized cognitively. Below are some perceptual orientations learners have to go through. a) Perception: 1) Demonstrate to learners a posture of relaxing the muscles as opposed to tensioning them. Muscle tension is most visible in the form of stretched neck muscles, which if detected it means the whole musculature is inappropriate for lax sound production of which a schwa is most typical. 2) Once the relaxed position is maintained, proceed to demonstrate short intermittent moaning sounds as if feeling pain which tend to be typical of schwa vowel postures. 3) If you, as an instructor, are a native or near-native speaker of English who has mastered the articulatory process of vowel reduction or schwaization in English, try to demonstrate the so-called hesitation vocal gesture (Delattre, 1965) which happens to be very much a schwa-like sound. 4) Conduct a demonstration of CV type of nonsensical syllables with the following vowel qualities: [] as in the English : [l l l] [a] as in Arabic الNo): [la la la] Then pronounce the three syllables with each vowel in two formats: firstly, change the stress from the first syllable through the third maintaining the vowel quality and with no temptation, whatsoever, of reducing it to a schwa; secondly,
214
PRONUNCIATION IS IN THE BRAIN do the same but this time reducing the unstressed vowels to schwas. Repeat as demonstrated:
lalala
lalala
lalala
lall
llal
llla
5) Shift from nonsensical syllables to real words that are typical of schwa articulation such as banana and America. 6) Model the pronunciation banana emphasizing the reduction of the vowels of the first and third syllables and accentuating, and somewhat exaggerating, the length and quality of the vowel in the second syllable. 7) Model the pronunciation of banana in English, e.g., [bnn]2 and compare it to a simulated pronunciation of it in Spanish, e.g., [banana]. It is perfectly acceptable to somewhat over-accentuate (exaggerate) the differences between the two pronunciations to attract the attention to them. This over-accentuation is similar to caricaturish drawing in which the distinctive features are exaggerated to attract attention. However, once the objective has been attained, instructor should stop the exaggerated versions and end his demonstration with normal and natural modeling of pronunciation 8) To reinforce the auditory channel with the visual one, transcribe the English pronunciation of banana phonetically and portray the syllabic structure of the word with various schematic diagrams that visualize the difference such as below,
2
Or [bnn] as in RP.
CHAPTER 13
The schematic representation can be in any form or shape as long as it visually demonstrates the reduced shapes versus the augmented one, the latter of which often standing for the accentuated syllables. Model the syllabic pronunciation repeatedly. 9) For further visualization, demonstrate vowel reduction and accentuation through nonverbal gestures. To demonstrate the English pronunciation of [b n n] nonverbally, take one short step, followed by a long step and then another short one; the size of the steps indicates the strength of the stress and degree of prominence of the vowel. For the demonstration of the Spanish pronunciation, take three steps, which are, more or less, of the same medium size although the stressed syllable will slightly affect the quantity (length) and prominence of the vowel. Visually, the Spanish pronunciation of will schematically appear as follows.
215
216
PRONUNCIATION IS IN THE BRAIN b) Recognition: 1) Number [b n n] and [ba na na] as #1 and #2, model them several times and ask learners to identify them as 1 or 2. 2) Repeat Recognition#1 above then ask learners to identify the word in its reduced (English) form or unreduced (Spanish) version.
c) Production: Return to activities in the perception stage and redo them. Then ask learners to reproduce everything seen, heard or perceived in the following manner: 1) Initially, any attempt at production by learners should be instantaneously after the instructor’s modeling with no background noise or speech separating the modeling from production. This procedure secures the recency effect of the input in the learner’s short-term memory circuitry for instant retrieval and reproduction. 2) If the instructor feels the learners still experience difficulty in recognizing and producing a schwa or a schwa-like vowel viz., [ , which is “acceptable for some native speakers of English” (Whitley, 1986:58), instructor should sustain the modeling. If the failure continues, the instructor should not persist. A break should be taken to let the learners relax. The physical and psycholinguistic distancing of self from the practice and drilling helps both the instructor and the learner to start afresh—the instructor with a more energetic and enthusiastic attitude and the learner with a more determined, thoughtful and hopeful attitude. 3) The instructor should encourage learners to think of the targeted sounds and the overall practice for their mastery. This reflective thinking can go on anytime during the practice and drilling or the distancing-of-self breaks. Techniques 1 (immediate repetition) and 2 (a break), above, should apply in the orientation of any aspect of pronunciation. 4) If the instructor feels learners need a refreshment of memory, he should afford them with more opportunity to internalize the production of a schwa through comparison
CHAPTER 13
217
with other non-reduced vowels in syllables similar to those in Perception/4:
lalala
lalala
lalala
lall
llal
llla
5) Place the word “banana” in short appropriate sentences, portray the English rendition in broad phonetic transcription, model the pronunciation and ask learners to produce after. Example: Barbara eats a banana a day. [brbr
ts
bnn
de]
6) Select other appropriate English sentences, such as the one below, model them and ask learners to reproduce them. Example: I can see a banana in the tree. [ai
kn si
bnn
n
tri]
13.7. TECHNIQUES FOR TEACHING ACCENTUATION (STRESS)
Before understanding the nature of rhythm and its teaching, one has to be aware of the premise on which rhythm is based— stress. The traditional approach to teaching stress has been through the auditory modality and its predominant technique has long been, and unfortunately still is, repeat-after-me. This exclusive auditory sensory modality in teaching stress may not be effective with all learners. Besides, in some very traditional methods of teaching, stress and rhythm are hardly ever tackled. To help more learners master the process of stress perception, recognition and production effectively and efficiently, the instructor has to resort to a cognitive approach that is implemented through a wide variety of activities and exercises based on diverse sensory modalities—auditory, visual and kinesthetic/ proprioceptive. Occasionally, multisensory modalities are jointly invoked. Once again, one has to be reminded of the triangular base of teaching pronunciation in the form of perception, recognition
218
PRONUNCIATION IS IN THE BRAIN
and production. These three sub-processes are interrelated and one leads to the other. In the following sections, the techniques of teaching stress perception and recognition will be clustered together, while its production will be handled somewhat separately since the techniques applicable to teaching the first two processes may not be applicable to the teaching of the third and vice versa. a) Perception: 1) Take a two-syllable word such as and demonstrate it with accent on the first syllable and then with accent on the second syllable as exhibited below. Also to highlight the grammatical and/or semantic role of stress inform the learners that the first rendition of indicates a ‘noun’, whereas the latter indicates a ‘verb’.
pro duce
pro duce
2) Ask whether any of the learners wants to demonstrate the difference. If so, let them demonstrate the difference. Such exercises are the initial exposure to the phenomenon of stress. You should expect to have learners in your class who do not know what is going on and what the difference is between the above two demonstrations of stress. Personally, I am confident that such learners will always be there. Once upon a time, I had been one such learner and I always encountered such learners in my classes as an instructor. 3) Go one step further to impress the difference on the learners and try again by tapping on a desk or table—one strong tap followed by a weak tap for and reverse the order for . 4) To dramatize the difference and capture the attention of the learners, grab an empty carton box or can and beat the rhythm of the two words on it. The visualization that goes on with the latter demonstration is of great help to reinforce the retention of the beats and the overall stress location.
CHAPTER 13
219
b) Recognition: 1) Assign colors to the stressed and unstressed syllables. Produce a prepared poster with the targeted two words. Tell learners that the red-colored syllable is the stressed one in each case and demonstrate the difference. 2) To do it differently but still visually, you may write/print the stressed syllable in larger size as in,
pro duce
produce
then take a big step followed by a small step forpro duce, followed by a small step and a big one for produce. The arrows mark the size of steps 3) One important set of features for the learners to observe and impersonate are the facial and body gestures that accompany the stressing of a syllable. Stressing a syllable immediately implies the exertion of greater physiological effort. This additional effort reveals itself physically in the overall body gestures of the speaker. Such features may be in the form of a downward head movement, rising of an eyebrow or both, or simply a sudden movement with the hand or arm. 4) For an auditory checking on stress recognition, take another noun-verb contrast of English Such as: vs. , number them #1 and #2, consecutively and read them randomly asking the learners to identify them as #1 or #2. 5) The learners will be taken one step further in the recognition challenge if they are given three nonsense syllables to be demonstrated three times each with stress on a different syllable as below,
220
PRONUNCIATION IS IN THE BRAIN Mark the triplets as #1, #2 and #3 and then read them randomly asking the learners to identify them as #1, #2 or #3. One can also tap the three stress triplets on a desk, empty can or on anything that resounds and ask the learners to identify the location of the strongest stress within each triplet. If you have a small wooden hammer or a gavel, tap or beat the same stress triplets then ask learners to recognize the strongest beat in each case. In both of the above demonstrations (reading or tapping) there are certainly visual signs to indicate the stressed syllable. Bring the latter fact to the attention of the learners. 6) Repeat the beating (tapping) of the stress triplets mentioned above putting more emphasis on the movement of the hand or the gavel so that the movement will visually attract the attention of learners then ask them to identify the stressed syllable. One can also use the overhead projector or power point slides of the stress triplets telling the learners that red-colored syllables are the stressed ones. 7) Finally, you can ask for volunteers to demonstrate the stress triplets.
c) Production: Obviously, production is the last phase in the training of learners to master the dynamics of stress. It is in the production phase that learners will be far more able to sense stress dynamically and proprioceptively because they will actually be physically producing the stress either through nonlinguistic activities or through authentic linguistic ones as follows: 1) Foremost in this regard is for the instructor to demonstrate any stress performance that he intends to ask learners to execute through tapping or beating of syllables in doublets, triplets or even quadruplets. It is preferable for both instructor and learner to use the hand or a small hammer or gavel to perform the demonstration. It is easier for the learner to begin with two-syllable structures and proceed further with multi-syllable structures. Ask for a volunteer to perform the beating of two syllables once with the first syllable receiving the strong stress and then the second syllable receiving it as in:
CHAPTER 13
Gradually, the performance should include more learners with more syllables and beats as in:
or
While all this tapping or beating of syllables goes on, you may ask the learners to notice the difference in the physical effort exerted by the hand of the performer. The physical difference is visually quite noticeable. In fact, the performer may be asked to intentionally exaggerate or intensify the physical effort or bodily gesture that accompanies the production of a stressed syllable. 2) After one is done with the non-linguistic exercises, it is time to move to the linguistic ones especially with counting numbers beginning with ‘one two’ and moving to ‘one two three’ etc… as follows:
/one two/ / / Reverse the stress.
/one two/ / /
221
222
PRONUNCIATION IS IN THE BRAIN Move to three units advancing the stress every time:
Move to four units advancing the stress every time:
3) In English, the most appropriate materials to teach stress placement and its linguistic significance at word level are the noun-verb doublets such as: in table 13.4, below, Noun
Verb
subject
subject
record
record
contract
contract
perfect
perfect
present
present
insult
insult
Table 13.4. Sets of noun-verb contrasts signaled by stress placement Pronounce pairs such as the above slowly and somewhat emphatically (i.e., with somewhat more determined articulation) and ask the whole class to repeat after you as a chorus. Later on, ask for individual volunteers to come to the front and repeat the above list after you. At this stage, the instructor may carry the performance a step further by asking individual learners to produce the pairs without being modeled.
CHAPTER 13
223
If there are still more learners who still experience some difficulty, you have to be patient with them and be ready to repeat the performance and the demonstrations. In such instances, you should proceed in the following manner. First, demonstrate the examples with somewhat more emphatic articulation in the hope of highlighting the difference between them even further. In fact, you may select pairs of words in English which contrast not only grammatically, such as the above noun-verb, but also semantically and orthographically such as vs. and vs. . Second, demonstrate them in conjunction with somewhat more visible facial and body gestures. Third, as pointed out in the early stages of teaching production, learners should produce the targeted modeling instantaneously after you with no background noise or speech separating the modeling from production to trigger the recency effect and its impact on memory. Fourthly, ask the learner to produce a series of repetitions for a given item of demonstration. These repetitions may sound mechanical in nature, but they may have a positive impression on the brain inasmuch as their long-term retention is concerned. After all, learning by human beings is a process of transforming mechanical habits into cognitive ones.
13.8. CONCLUDING REMARKS
There are no solid and comprehensive rules that predict and capture stress placement in English words for two main reasons. First, English is a stress-timed language with mostly unpredictable stress placement (MacKay, 1978, 150). Second, there is a high percentage of foreign loanwords in English that maintain their own stress patterns and create exceptions to the native rules. This makes the problem of mastering stress placement for learners of English as L2 a major difficulty and source of accent. It is, therefore, an area which is worthy of the attention of any instructor teaching English or any other language as L2. The instructor has to be professionally qualified to teach this very essential linguistic aspect in cross-language teaching. As has been discussed earlier on, instruction can begin as a phonetic practice which gradually should lead to authentic linguistic materials. The most appropriate linguistic materials would be the
224
PRONUNCIATION IS IN THE BRAIN
so-called strong vs. weak forms the majority of which fall under the categories of prepositions, conjunctions, auxiliary verbs, etc. Word
Strong Form
Weak Form (most frequent)
[]
[] or even [] or []
[]
[]
[]
[]
The noun-verb category of words which are only distinguished grammatically by the placement of stress such as the ones cited in section 13.7. Additional excellent materials that serve the purpose of teaching stress placement in English words is through grammatical derivations. One of the most typical examples cited by teachers of English to demonstrate the variability of stress and rhythm in English is the word . Notice the following changes in primary stress indicated by the bold syllables in conjunction with the large dots as marked below.
Because of the highly unpredictable nature of stress placement in English, some phoneticians and instructors of pronunciation advise learners of English to listen carefully to the native speakers or materials recorded by them or use a pronunciation dictionary. This is not a bad advice, but it is a privilege that is not attainable by all learners of English. Consequently, some, regardless of how few, of the most common and effective rules of stress assignment in English should be taught systematically. However, foremost in importance is that stress and stress placement should be taught because many learners in many languages are not really aware of the existence of stress as a linguistic dynamic or may not have the auditory sensitivity to pick up the stress on their own.
CHAPTER 14: TIPS FOR ACCENT REDUCTION AND ACCENT DETECTION 14.1. INTRODUCTORY REMARKS
Accent reduction has already been defined and elaborated on (4.5.2) as an attempt to bridge the pronunciation gap between the native speaker of a given language and someone who is attempting to learn that given language as L2. The only further addition to the theme of ‘accent reduction’ in this chapter is to provide some key tips to help learners focus on the points which are considered significant in its reduction. With regard to accent detection in the context of this book, two points are worthy of consideration. First, how does one discover that he has an accent? Second, how can accent lead one to the identification of the linguistic affiliation of the speaker? In response to the first question, one has to bear in mind that perhaps one of the most controversial aspects of pronunciation at large is that many people do not realize they have an accent or admit to having it because they hear themselves through their own native linguistic filter or prism. There are two ways to convince one to admit having an accent. First, the instructor or coach brings it to one’s attention and advises him to work on it. Second, it is self-discovered after having ample exposure to the target language especially in cases of full immersion coupled with experience in speech production and pronunciation. The latter case applies to my personal attempt at accent reduction. Nobody brought my accent in English to my attention; I was fortunate to be able to discover it and work on it diligently. I pointed out in chapter 1, that my so-called ‘Iraqi English’ was distinctly accented with the linguistic substrata of my three native languages. I did not have any problem with consonants except for replacing the English approximant with my tap and trill s. I had some problems with the vowels, especially 225
226
PRONUNCIATION IS IN THE BRAIN
with the neutralization of unstressed ones. My most serious problem involved accentuation and rhythm at large. I also have to admit that I did not discover my accent right away; it took about six months to gradually realize that I had a few fundamental problems. I had to spend long days, weeks and months paying intensive attention to native speakers of English and practicing what I listened to. I certainly reduced my accent considerably in segmental as well as suprasegmental constituents. Specifically, I have to emphasize that I improved my stress placement and accentuation in a very marked way. In spite of all that improvement coupled with decades of full immersion in English as an adult in the United States, I still feel I have an accent; however, it hardly ever interferes with communication. Briefly, I have overcome almost all sources of phonological accent, but there are still some residues of phonetic accent that pop up here and there. 14.2. TIPS FOR ACCENT REDUCTION
There are three general routes that lead to accent reduction. First, try to improve pronunciation as much as possible by first mastering the basic phonological differences between L1 and L2, including those pertaining to segmental elements (consonants, vowels) and other suprasegmentals (stress, rhythm and intonation and/or tones) to avoid semantic confusion. This is one way to prove the functionality of the phonological vs. phonetic accent dichotomy and why educationally the former should receive more attention. Second, work on the most striking phonetic differences that do not result in semantic change, but do generate different degrees of noise ranging from slight to substantial which indirectly interfere with comprehension. Third, raise the competency level in other aspects of language including, morphology, syntax, lexicon and idiomatic expressions to facilitate comprehension. 14.2.1. Tackle the most Salient Phonological Problems
Because the phonological deficiencies in one’s L2 proficiency are more perceptually outstanding and semantically distracting, they should be identified first and given the priority in tackling. Naturally, if the individual is not able to identify such deficien-
CHAPTER 14
227
cies, a competent instructor or a seasoned supervisor has to do that. In any case the following are some of the most significant phonological accent-causing problems that should receive priority. a) There is a significant number of consonantal elements generating phonological accent across languages. Just to cite some examples, Arabic poses a set of several really challenging sounds that learners of Arabic usually encounter. They include, the uvulars [], the pharyngeals [ ] and the emphatic [ ]. For instance, /q/ = [q] is overwhelmingly confused with /k/ = [k] by non-Arabs. For Germans, Arabs, Filipinos and Greeks the sounds of /ʤ/, /p/ /f/ and // are, respectively, quite demanding. The pair / / is absent in many languages; consequently, it is a fundamental source of both phonological and phonetic accent; frequently, the pair is replaced with either /s, z/ or /t, d/. What is of utmost importance is not just the ability to accurately pronounce such sounds in isolation (out of linguistic context), but also in proper conversational context and in lengthy discourse. In my teaching career, I have come across many learners who master the execution of the phonologically alien sounds in isolation, but occasionally fail when a lengthy and indepth conversation is sustained. This happens when the alien sound has not yet become a cognitive entity stored in the subconscious; consequently, the speaker may become vulnerable to a phonological lapse or pull in the direction of L1. b) Care must be taken of the most significant vocalic elements generating phonological accent such as the absence of short (lax) vs. long (tense) vowels of English in many languages including Italian, French, Russian, Tagalog and Spanish, among others. For instance, a Mexican must distinguish between an English short/lax vowel [] and a long/tense vowel [i] in order not to confuse a = [bd] (for a contract) with a = [bd] (on a rosary). There is no place, whatsoever, for a typical Mexican [bid] rendition in English. Equally, a native Italian should be able to distinguish between = [pl] and = [pul] as there is no place for his rendition of those two English words as [pul]. As for individual vowels one can readily notice the absence of close front rounded vowel /y/ of French and German in English and many other
228
PRONUNCIATION IS IN THE BRAIN
languages. The very English vowel // has no counterpart in many languages; it is, therefore, replaced with /a/ or //, etc. c) Seriously consider the most significant consonant clusters generating phonological problems for speakers of languages without clusters or with very limited word-initial, middle or word-final clusters. Typically, such problems are observed with learners of English of Japanese, Korean, Chinese, Hispanic, Italian, Arabic linguistic backgrounds, among others. Just for the sake of example, any Japanese should really work extremely hard on moving his pronunciation of = [f] far away from the traditional Japanese rendition of [fut-ta-bo-]. d) Carefully abide by the stress-placement differences between L1 and L2. Differences in stress placement can be a major source of both phonological and phonetic accent. For instance, one should be aware of the fact that the difference in stress locations between some English words determines whether they are verbs or nouns/adjectives such as with , etc. Also, the learner should be aware of the fact that regardless of the difference in spelling between and , it is the stress location that triggers the semantic difference between them. e) Observe the rhythm differences between L1 and the targeted L2. The shift from stress-timed rhythm to syllable-timed one or vice versa is one of the most significant markers of accent. Although the accent tends to be more phonetic than phonological, but the phonetic noise that ensues causes serious pronunciation distortion that obscures meaning. If, for instance, a Spanish speaker imposes his syllable-timed rhythm on the stresstimed rhythm of English the overall rhythmical chunking of the target language and tempo changes. In many instances, a Hispanic speaker, whose overall linguistic competency in English is quite proficient except for the retention of his syllable-timed rhythm, may still sound somewhat unintelligible to a native speaker of English. It is like speaking a ‘morse-code’ rhythm with a ‘machine-gun’ rhythm. f) Tone and intonation are equally important like stress and rhythm. They are extremely significant when the individual involved is moving from a tone language, such as Chinese, to an intonation language such as English or vice versa. For instance, Mandarin Chinese has four tones in the form of: high level, rise,
CHAPTER 14
229
fall-rise and fall. To demonstrate the lexical function of such tones, one has to use the most common monosyllabic word in Mandarin, which conveys the following meanings with the four tones: /ma/ = (mother); /ma/ = (hemp); /ma/ = (horse) and /ma/ = (scold) (http://mandarin.about.com/od/pro nunciation/a/tones.htm). Although both tone and intonation are based on pitch, the difference is in the use of pitch. In the intonation languages, pitch tends to be the feature of a phrase or sentence to signal attitudinal, emotional and, at times, grammatical differences, while tone tends to be the feature of a word or syllable to signal lexical and/or grammatical differences. Thus, the difference between tonal and intonational languages constitutes a major difference in the overall melody of a given language. This is why adult learners of intonational languages, such as English, experience serious difficulty when learning tonal languages. By the same token, people with tonal languages encounter similar level of difficulty when embarking on learning intonational languages. Each group struggles to restrain the strong drive in the direction of pitch orientation in their languages when working on their L2s. Failing to restrain that drive results in serious overall melody change coupled with lexical, grammatical and attitudinal alterations. 14.2.2. Tackle the most Salient Phonetic Problems
Obviously, the instructor has to point out to learners which problems rise to the level of a phonological accent and which ones remain at the phonetic level between the two languages involved. Between any two languages, there are innumerable phonetic differences some more prominent than others. Let us take English and Greek. In Greek, the English postalveolar affricates [ʧ ʤ , as in and , are absent and are usually replaced with their alveolar counterparts [ʦ ʣ . The substitution does not confuse meaning; it simply renders unfamiliar the words in which the sounds occur. Although for English learners of Spanish the confusion between two s as in = [kao] (dear) and = [karo] (cart/car) constitutes a phonological problem, for Spanish learners of English the problem remains phonetic. Most typical for speakers of sub-continental Indian language, retroflexion is the primary source of their phonetic accent; however, it is so pervasive that
230
PRONUNCIATION IS IN THE BRAIN
it can seriously interfere with their overall comprehension by the natives of the language they are targeting. The comprehension is further hampered with the syllable-timed stress type that most Indians languages have. On the flip side, it is worth pointing out that for English speakers learning other languages, it makes a substantial difference in their phonetic rendition of those languages if they can replace their approximant with a tap, flap, or trill according to the specific language they are targeting. 14.2.3. Improve other Linguistic Skills
Developing high competency in all linguistic systems including syntax, morphology, lexicon and idiomatic expressions alleviates the negative impact of both phonological and phonetic accent in two ways. First, it affords the person using L2 greater sentential and discourse fluency. Second, the resulting fluency increases the likelihood of predicting the meaning of the mispronounced words by the native listener from the context of the sentence or discourse produced. This is exactly what most educated adults attempting communication in L2 make recourse to. It was mentioned earlier on that retroflexion permeates throughout the native languages of sub-continental India. Consequently, when speaking or learning other languages, the strong retroflexion colors their rendition of any L2 they attempt to speak. It, thus, generates serious phonetic noise to the ear of the natives of the attempted L2. One strategy to mitigate the pervasive interference of retroflexion noise in their L2s is through enhancing the competency in syntax, morphology, lexicon, idiomatic expressions, etc. They will all contribute toward a better conveyance of meaning to their listeners regardless of the phonetic interference. 14.3. ACCENT DETECTION
Research and investigation in the domain of automatic speech and accent detection and recognition systems using advanced software is in considerable progress (Zheng et al, 2005). However, machine accent detection is not the focus here; rather, the focus is on live accent detection in real-life situations. Stated differently, the attention is focused on man-to-man accent detec-
CHAPTER 14
231
tion not machine-to-man. There are different scenarios for the detection of the presence of accent by different individuals ranging from an ordinary unsophisticated person to a professional one. Let us shed some light on these scenarios. 14.3.1. Accent Detection by Ordinary Individuals
This is a scenario which involves ordinary native speakers of a given language listening to others who are speaking their language as L2. The listeners in this case may range from illiterate to educated and to highly educated individuals, but none of whom has professional orientation in linguistics, in general, and phonetics and pronunciation, in particular. All that they have is a sensitive ear to their language that helps them to innately assess the accuracy of the pronunciation of their L1 by an L2 speaker. Often, they perceptually, as well as cognitively, sense the accent, but cannot describe it or pinpoint the details. The assessment of accent by some of them might be very general and as simple as saying: “He has an accent.” Others might say: “The vowels sound somewhat off the normal.” A few might be fairly detailed in their commentary by identifying certain vowels, consonants or even stress and rhythmic performances that did not sound accurate and/or native-like. 14.3.2. Accent Detection by Professionals
In this context, professionals represent those individuals who possess knowledge and experience in the nature and structure of human language with reasonable exposure to linguistics and phonetics. They understand human language as sets of structures and systems and the manner in which they collaborate accurately to generate meaning. With specific focus on pronunciation, they possess enough expertise to make very refined judgment as to the source of mispronunciation and accent as well as the recommendation and techniques to eliminate the source of the problem. However, to be able to achieve all that, the individual has to have knowledge about the structures and systems of the two or more languages involved. This knowledge does not imply knowing the two languages fluently or even fairly; however, he must have a certain degree of experience and exposure to the two languages in action and real-life situations.
232
PRONUNCIATION IS IN THE BRAIN
Personally, I do not consider Spanish a language that I know, but I have had ample exposure to it in the community and as well as in my classes. For a period of three decades, I accumulated rich knowledge and experience about Hispanic learners of English. I identified almost all the major pronunciation errors they made in English. This authentic interaction with Hispanic learners of English encouraged me to pursue my interest in documenting my experience in a book1 which covered a wide range of pronunciation problems they encounter in learning English and the tips to enable them overcome them. It was through such educational exchange with my students that I enriched my own knowledge and field experience in teaching pronunciation. Their paramount difficulty in distinguishing English from or from helped me identify the vowel system of Spanish as a centrifugal one in which the vowels tend to be almost ‘frozen’ in their quality and quantity. This forces them cognitively and articulatorily to cluster the two vowels of each pair into one vowel of their own that is neither of the two English vowels. It took me a long time to discover the problem, analyze it and design effective strategies and exercises to help them perceive the difference, recognize it and execute it comfortably. 14.3.3. Telling the Linguistic Background of a Speaker through Accent
Some people, especially linguists or those who have a passion for languages, manifest a hobby or skill at trying to identify the linguistic background of a speaker of a second language (L2) by simply listening to the person without even having any prior knowledge of the native language of the speaker. This is not an easy task because it needs talent, linguistic skill and experience; most likely, a combination of all three. Personally, I have tried this hobby and I succeeded at times and failed at others. Such a linguistic skill can work as a two-edged sword for professional
1
Linguistic tips for Latino learners and teachers of English, 2007.
CHAPTER 14
233
orientation of under-cover agents as well as for those who hunt for them. More will be said on this in due course. This is not an easy task because it needs thorough knowledge about the overall pronunciation of the L2 that is spoken and the L1 of the speaker; additionally, it requires knowledge about the two phonologies and at times even more refined phonetics. Let us take English as the attempted L2 by an X language speaker. One of the foremost prerequisites to make a correct prediction is to be aware of some of the most salient phonological features of the X language and their expected reflexes in the rendition of English. Even with this apparent thorough knowledge, one can go wrong for two primary reasons. First, some typical phonological and/or phonetic features may be shared by more than one language. Thus, one tends to think that the language of the speaker is X but it turns out to be Y or Z. Let me cite one example of how I went wrong in my linguistic identification of one of my students. It was the first day of a class and I wanted to get as much background information about my students as possible. Everyone began to give a brief summary of his/her background. In the short presentation of one of the students, I heard a couple retroflex s, so, I said: “Are you from India or Pakistan?” She responded: “I am from Guatemala.”
Of course, I was mute and I apologized. Deep in my heart, as a linguist, I was embarrassed for the glaring misidentification. She identified herself as a native speaker of Spanish, but where did she get that retroflex ? I still do not know to this day. My only explanation at the time was that the retroflex might have been part of a linguistic substratum of a local Native American language or an idiosyncratic feature of her overall speech. To add more specifics to this linguistic encounter, I have to admit that besides the hearing of the ‘retroflex’ , the student was of a dark complexion very much like many sub-continental Indians. The complexion enticed me to make the judgment that turned out to be wrong. So since then I taught myself a lesson: “Don’t be fooled by the complexion.” Obviously, my attempts at accent detection have not always been disappointing. On several occasions I was right on
234
PRONUNCIATION IS IN THE BRAIN
target and then the question came: “How did you know I was X?” Let me cite some such accent detection anecdotes. Years back, in one instance, I was involved in the registration for the new semester courses. One course was canceled, but was still on the list. Some students came complaining and yelling: “Why is it still on the list?” One of the students who desperately needed the course was the most vocal. I heard typical repeated alveolo-palatal [] and [] sounds in place of the English [s] and [z]. I wanted to cool her down and, I said to her: “Are from Greece or Cyprus?” She abruptly said: “How did you know?” I said: “Register for my course and I will tell you later.”
No doubt, my course was relevant to what she was planning to study. In another instance, I was in hospital for a surgery. A day after surgery, I was allowed to take some clear soup. When it was brought to my room the nurse said: “You hab to pinish all of it.” Naturally, her sentence contained several pronunciation hints that pointed in the direction of her Tagalog language background. Three hints were outstanding, namely, [b] for [v], [p] for [f] besides her unaspirated [t] and [p]. When she came back to check on me, I greeted her in Tagalog. She was surprised and said: “Are you ‘Pilipino’?”2 I said: “No, but I thought you were. I read your name tag.”3
Just recently, I greeted one of my new neighbors simply to strike a friendly relationship. There was work done in his garage door. I said: “You must be busy.” He replied: “Yes, I am changing my garage door.”
2 3
Remember, not ‘Filipino’
Of course, I gave a fake answer to avoid commenting on her accent.
CHAPTER 14
235
Once I heard [ʦʣ instead of [ʧʤ] and [gʣ instead of [gʤ , I guessed he was Greek. I said: “Tikanis?” (How are you?). He responded: “Kala” (Good).
And he was Greek. It took one little indicator of accent to identify him linguistically. 14.3.4 Hiding an Agent through Hiding an Accent
In this section, the aim is not to detect an accent, but rather to hide it lest it should be exposed with unknown consequences. It was made clear earlier on that accent reduction is professionally very beneficial for instructors of languages, actors, newscasters, spontaneous interpreters and under-cover agents; however, in the real world, there is one difference between the under-cover agents and the rest of the professionals who can tolerate making mistakes or even blunders, but under-cover agents are not supposed to. Any inappropriate linguistic gaffe can cost them dearly. No doubt, linguistic talent, language competency, in general, and pronunciation, in particular, are indispensible prerequisites for someone who is willing to serve as under-cover agent or the more crudely named profession of ‘spying’. Grosjean (2010-a) wrote a compact, but very linguistically and culturally rich article concerning the linguistic and cultural prerequisites for this very risky and mentally demanding profession of spying, especially as sleeper agent (or deep cover agent). A sleeper agent can be either the native of the same country that is spied on or a native of the spying country who has been embedded in the target country for a long time. In the case of a native citizen, there is no problem of native language and culture competency and native unaccented pronunciation because he belongs to that linguistic/cultural community. The lack of native or even native-like linguistic and cultural qualifications may be the source of problem for the non-native agent. This is so because in spite of the very disciplined linguistic and cultural orientation of the person, he may inadvertently and subconsciously reveal a linguistic or cultural hint, regardless of how minor, that is inconsistent with the language and culture of the targeted country. As the intruder country does its best to groom
236
PRONUNCIATION IS IN THE BRAIN
its spies, so does the targeted country in priming its counterintelligence cadre. Spy trackers should supposedly be coached to watch for the tiniest linguistic, cultural and behavioral indications to be able to detect ‘deeply covered’ agents. The above discussion brings forth the very delicate responsibility of preparing an under-cover agent. Naturally, if the agent is a native speaker then accent becomes a non-issue. For the non-native agent achieving native proficiency in L2 is a rarity except if he is embedded as a very young person so that he will have ample opportunity for full immersion in the target language and culture. In the absence of the latter condition, the candidate will come from amongst adults. Even with adults, the younger the better for mastering an L2 or C2. Inasmuch as pronunciation is concerned the target will be near-native proficiency as the native proficiency is a rarity. Even near-native proficiency is not an easy task to achieve in pronunciation as early exposure to the sound system seems to be a prerequisite unlike the rest of the linguistic systems of lexicon, morphology and syntax in which many educated non-natives can outperform many natives. Once again, international figures such as Joseph Conrad, Krishna Menon and Henry Kissinger achieved the highest levels of native competency in all linguistic systems of English except for pronunciation. In the case of the latter two, whom I have heard speaking, it has been a phonetic accent (no interference with meaning) rather than a phonological one (interference with meaning). It is assumed that individuals recruited for spying must demonstrate prowess in many skills, foremost of which are expected to be competency in L2. This implies extensive orientation in pronunciation with particular focus on areas that trigger the most revealing phonological accent followed by areas of less revealing phonetic accent. To achieve this goal, the individual should go through a very stringent preparation and orientation under the supervision of highly gifted linguists and phoneticians whose focus would be all the steps highlighted in section 14.2 perhaps with greater depth and rigor. The greater in-depth and rigorous orientation of spies is needed because of the risky nature of the assignment. Greater in-depth orientation implies not only preparation in linguistic and phonetic nature of the targeted language, but also in C2
CHAPTER 14
237
orientation. “If a person is claiming to be a native speaker, there may be occasional language oddities or incongruous phrasing that can alert you to this person not being a native speaker but trying to hide this fact” (http://www.wikihow.com/Spot-a-Spy). This implies that any L1 and C1 residues seeping accidentally through the adopted L2 and C2 can become potential traps. At the highest level of orientation, the person may need a certain degree of theoretical familiarization with linguistics and the dynamics of speech production and pronunciation. Considerable difference is expected in the interpretation of speech acts and linguistic facts between an ordinary person, a person with some exposure to linguistics and a specialist. Consider this example that I recently had to comment on. One of my friends who is a geologist, thought that another friend of ours was the speaker of certain Neo-Aramaic (Modern Assyrian) dialect because all his s sounded fully retroflex instead of the regular tap or trill s of the language. I said: ‘No, he was not.” Then, I explained to him that our friend has a speech defect that made his s sound as if they were retroflex. This indicates the huge gap between the impression of a lay person and the reality of a specialist. There is no doubt, whatsoever, that the theoretical knowledge and expertise in any field facilitates the mastery of applied skills and the creativity in them. One last piece of advice from a linguist to whoever is willing to consent to the mission of an under-cover agent: beware your subconscious when handling the target language (L2) and target culture (C2). It has been emphatically stressed throughout this book that both native language (L1) and native culture (C1) are the creation of the brain and are stored in the subconscious. This is why Grosjean highlights the fact that “Sleeper agents have to ‘speak monolingually’ at all times, even when the situation is conducive to code-switching” (Grosjean, 2010-b). Certainly, Grosjean inspires me to stress the fact that sleeper agents should also behave monoculturally lest they should be snared because of falling victim to their cognitively deep-seated C1. Let me cite the following anecdote from years back when two undercover Iraqi policemen tricked two Iranian smugglers into admission and arrested them by manipulating the linguistic/cultural subconscious of the smugglers:
238
PRONUNCIATION IS IN THE BRAIN “Half a century ago when we used to travel from Kirkuk in the north of Iraq to Baghdad in the center, we had the option of taking a very slow train that took twelve (12) hour or the option of a bus that took six hours, but was vulnerable to getting lost due to the lack of modern roads with proper signs. Once I took the bus and it pulled out in the morning and continued for two (2) hours after which it stopped at a small town and picked up two more passengers and continued the journey. The two passengers were dressed in folk attire and carrying midsize bags on the back. After an hour or so, and to the surprise of all passengers, the driver announced that he had lost his way and had been circling around during the last hour. The diver was embarrassed to find himself back at the same town where he first stopped and picked up two passengers. We were all frustrated, but we had no better choice. Just a coincidence, two more passengers were picked up who were dressed in pants and jackets. They proceeded to the last seats on the bus which were empty. The bus drove for about ten (10) minutes and all of a sudden I heard someone from the back seats loudly uttering a few words which I knew they were not Arabic, Turkmeni or Assyrian because I spoke those languages. They did not sound Kurdish as I had some impression of how Kurdish language sounds. When the few words were spoken out, I saw only two people from the middle seats turn their head to look back at the source of the words. It was one of the last two passengers who was the speaker. In a split second, they both dashed forward in the direction of the first two passengers who looked back, yelled out some obscenities and handcuffed them saying that they were Iranian smugglers whom they have been tracking for hours. They then ordered the driver to return to the same town once again where the police station was.”
I discovered later that the last two passengers were undercover policemen and had lost the track of those two Iranian smugglers, but they had a feeling that they had taken the bus. Ironically, the bus lost its destination and by a stroke of chance was back where it was two hours ago. Apparently, the two undercover policemen intentionally went to the back of the bus and sound-
CHAPTER 14
239
ed out loudly a phrase in the Farsi language. Of all the passengers, only the two smugglers turned their head decidedly and looked back because they thought they were welcomed by an acquaintance. Certainly, by turning their heads to the source of the Farsi language greeting, they did not behave consciously; rather, it was the two experienced policemen that played a linguistic trick on them through manipulating the unintentional, subconscious and reflexive response of the two native speakers of Farsi language. It was their linguistic and/or cultural subconscious that betrayed them. In sum those two smugglers were caught because they failed to behave as Arabic-speaking monolinguals, as Grosjean phrased it, and hide their identity; hence, under-cover agents should be more careful than those two smugglers in order not to take the bait and uncover themselves under the influence of the subconscious. 14.4. CONCLUDING REMARKS
No one likes to have an accent; however, accent is usually part and parcel of attempting an L2 when one is no longer a child. Some adults stay with their accent throughout their life, but others try to reduce it as much as possible. The degree of success in accent reduction is contingent on many factors including age—the younger the more successful—linguistic talent, enthusiasm, professional orientation, etc. If all factors are available, it is the professional instruction and orientation that eventually make the difference in the degree of success. Doubtless, not everyone is qualified to conduct the professional orientation in matters of language mastery and accent reduction. Any person delegated with this responsibility must have the qualifications of a linguist and phonetician or be a ‘linguistically’ gifted individual. From the linguistic perspective, all attempts at accent reduction or accent faking have no negative intent except for spying which is a double-edged sword: good for the spying agent, but bad for the spied on. If one accepts this type of professional service and wants to avoid the risky consequences, he has to learn how to restrain the subconscious linguistic and cultural dominance of L1 and C1 in situations where L2 and C2 are needed.
REFERENCES Abercrombie, D. (1967). Elements of general phonetics. Edinburgh: Edinburgh University Press. Adams, C. (1979). English speech rhythm and the foreign learner. The Hague: Mouton. Aitchison, Jean (1996). Seeds of language: language origin and evolution. New York: Blackwell. Anderson, John R. (1980). Cognitive psychology and its implications. San Francisco: W.H. Freeman & Company. Arnold, Magda B. (1984). Memory and the brain. Hillsdale, New Jersey: LEA Publishers. Baddeley, Alan D. (1976). The psychology of memory. New York: Basic books, Inc. Publishers. ——— (1993). Your memory: a user’s guide. London: Multimedia Books Limited. Bates, Elizabeth (1999). On the nature and nurture of language. Frontiere della Biologia. The Brain of Homo Sapiens. Rome: Giovanni Trecanni. Beck, Douglas L., and Flexer, Carol (2011). Listening is where hearing meets brain. http://www. hearingreview. com/issues/ articles/2011-02_02.asp. Best, Catherine T. (1991). The emergence of native-language phonological influences in infants: a perceptual assimilation model, Haskins laboratories status report on speech research, SR·107/108, 1-30. Best, Catherine T, McRoberts, Gerald W. and Goodell, Elizabeth (2001). Discrimination of non-native consonant contrasts varying in perceptual assimilation to the listener’s native phonological system. Journal of the Acoustic Society of America, Vol.109, 2, 775–794.
241
242
PRONUNCIATION IS IN THE BRAIN
Borden, Gloria J. and Harris, Katherine S. (1980). Speech science primer: physiology, acoustics and perception of speech. Baltimore: Williams & Wilkins. Bourne, L.E, Dominowski, R.L., Loftus, E.F. and Healy, A.F. (1986). Cognitive processes. Englewood Cliffs, New Jersey: Prentice-Hall. Caplan, David (1995). Language and the Brain, Vol. 4, No. 4, 1–4. Carpenter, Harry (2004). The Genie within: your subconscious mind. San Diego: Anaphase II Publishing. Catford, J.C. (1977). Fundamental problems in phonetics. Edinburgh: Edinburgh University Press. Catford, J.C. (1994). A practical introduction to phonetics. Oxford: Clarendon Press Chechik, Gal, Meilijson, Isaac and Ruppin, Eytan (1999). Neuronal regulation: a mechanism for synaptic pruning during brain maturation. Journal of Neuronal Computation, Vol. 11, 2061–2080. Chomsky, N. and Halle, M. (1968). The sound pattern of English. New York: Harper & Row. Christensen, Ken Ramshøj (2001). The Co-evolution of language and the brain: a review of two contrastive views (Pinker & Deacon). Grazer Linguistische Studien, Vol. 55, 1–20. Croom, Christopher, (2003). Language origins: did language evolve like the vertebrate eye, or was it more like bird feathers? http://www.csa.com/discoveryguides/lang/gloss_f.php Clancy, Barbara and Finlay, Barbara (2001). Neural correlates of early language learning. In Language development: the essential readings (Mike Tomasello & Elizabeth Bates, eds.), 307– 330, Wiley-Blackwell. Cummins, Jim (1979). Cognitive/academic language proficiency, linguistic interdependence, the optimal age question and some other matters. Working papers in bilingualism, 19, 121-129. ——— (1984). Bilingualism and special education: issues in assessment and pedagogy. Clevedon: Multilingual Matters. Dalbor, J. (1969). Spanish pronunciation: theory & practice. New York: Holt, Rinehart & Winston. Dale, P. & Poms, L. (1985). English pronunciation for Spanish speakers. Englewood Cliffs, New Jersey: Prentice-Hall.
REFERENCES
243
Daniloff, Raymond G. (1973). Normal articulation processes. (Fred D. Minifie, Thomas J. Hixon, Frederick Williams, David J. Broad, eds.). Normal aspects of speech, hearing and language, 169–209. Englewood Cliffs, New Jersey: PrenticeHall. Dauer, R. (1983). Stress-timing and syllable-timing reanalyzed. Journal of Phonetics, 11, 51–62. Deacon, Terrence (1997) The Symbolic species: the co-evolution of language and the human brain. London: Penguin Books. de Houwer, Annick, (1990). The Acquisition of two languages from birth: a case study, Cambridge Studies in Linguistics. Delattre, P. (1965). Comparing the phonetic features of English, German, Spanish and French. Heidelberg: Julius Groos Verlag. Eimas, P.D. (1978), Developmental aspects of speech perception. (In R. Held, H. W. Leibowitz, & H. L. Teuber, eds.), Handbook of sensory physiology, vol. 8. Berlin: Springer. Esling, John H. and Wong, R. F. (1983). Voice quality settings and the teaching of pronunciation. TESOL Quarterly, Vol. 17, 89–95. Fleischhacker, Heidi (2000). Cluster-Dependent Epenthesis Asymmetries (http://www. linguistics.ucla.edu/people/grads/ fleischhacker/uclawpl.pdf). Fox, Barbara, J. and Hull, Marion A. (2002). Phonics for the teacher of reading. Upper Saddle River/New Jersey: Merrill. França, Aniela Improta (2006). Introduction to neurolinguistics. In: Textos em Psicolingüística, Ingrid Finger; Carmen Matzenauer(Org.). Pelotas: Editora da Universidade Católica de Pelotas, Vol.. 1, 1–52. Gardner, Howard. (1983). Frames of mind: the theory of multiple intelligences. New York: Basic Books. ——— (1993). Creating minds: an anatomy of creativity seen through the lives of Freud, Einstein, Picasso, Stravinsky, Graham, and Gandhi. New York: Basic books. Gimson, A. C. (1967). An introduction to the pronunciation of English. London: Arnold. Gopnic, A., Meltzoff, A., Kuhl, P. (1999). The scientist in the crib: what early learning tells us about the mind. New York: Harper Collins Publishers
244
PRONUNCIATION IS IN THE BRAIN
Grosjean, Franҫois (2010-a). Bilingual: life and reality. Harvard University Press. ——— (2010-b). http://www.guardian.co.uk/education/2010. Hadlich, Roger, Holton, James and Montes, Matias (1968). A drillbook of Spanish pronunciation. New York: Harper & Row, Publishers. Handbook of the international phonetic association (1999). Cambridge: Cambridge University. Hiiemae, Karen M. and Palmer, Jeffrey B. (2003). Tongue movements in feeding and speech. Critical Reviews in Oral Biology & Medicine. Vol. 14, 413–429. Hyman, Larry (1975). Phonology. New York: Holt, Rinehart & Winston. Jabbari, Ali A., van de Weijer, Jeroen, Safari, Parvin and Falaknaz, Farane (2012). Journal of Teaching Language Skills, 3/4, 59–76. Jakobson, R., Fant, G. & Halle, M. (1969). Preliminaries to speech analysis. Cambridge, Massachusetts: M.I.T. Press. Johnson, J. S. and Newport, E. (1989). Critical period effects in second language learning: the influence of maturational state on the acquisition of English as a second language, Cognitive Psychology, Vol. 21, 60–99. Joseph, Rhawn (2011). Origin of thought: consciousness, language, egocentric speech and multiplicity of mind. Journal of Cosmology, Vol. 14. Kanokpermpoon, Monthon (2007). Thai and English consonantal sounds: a problem or a potential for EFL learning? ABAC Journal, Vol. 27-.1, 57–66. Kelly, Spencer, Özyürek, Asli and Maris, Eric (2009). Two sides of the same coin: speech and gesture mutually interact to enhance comprehension. Psychological Science, Vol. 20,1–8. Kenyon, J. S. and Knott, T. A. (1953). A Pronouncing dictionary of American English. Springfield, Massachusetts: G & C Merriam Company. Kissin, Benjamin (1986). Conscious and unconscious programs in the brain. New York: Plenum Medical Book Company. Ladefoged, P. (1982). A course in phonetics. New York: Harcourt Brace Jovanovich. Ladefoged, P and Maddieson, Ian. (1996). Sounds of the world’s languages. Cambridge, Massachusetts: Blackwell.
REFERENCES
245
Larson, Christian D. (1912) Your forces and how to use them (http://www.sacred-texts.com/nth/yfhu/yfhu02.htm). Laver, John (1980). The phonetic description of voice quality. Cambridge: Cambridge University Press. ——— (1994). Principles of phonetics. Cambridge: Cambridge University Press. Lehiste, I. (1970). Suprasegmentals. Cambridge, Massachusetts: M.I.T. Lenneberg, Eric H. (1967). Biological foundations of language. New York: John Wiley & Sons. Levitt, Robert A. (1981). Physiological psychology. New York: Holt, Rinehart & Winston. Loftus, Elizabeth (1980). Memory: surprising new insights into how we remember and why we forget. Reading, Mass.: AddisonWesley Publishing Co. Lowie, W. and Bultena, Sybrine (2007). Articulatory settings and the dynamics of second language speech production (http://www.phon.ucl.ac.uk/ptlc/proceedings/ptlcpaper_15e.pdf) MacKay, I. R. A. (1978). Introducing practical phonetics. Boston: Little Brown & Company. McKinney, James C. (2005). The diagnosis and correction of vocal faults: a manual for teachers of singing and for choir directors. Long Grove/Illinois: Waveland Press. Martinet, André (1964) Elements of general linguistics. London: Faber & Faber. Miller, Cynthia. Brain for success. (http://brainforsuccess.com/ howyourbrain work.html). Morley, Joan (1991). The pronunciation component in teaching English to speakers of other languages. TESOL Quarterly. Vol. 25, 481–520. O’Connor, J. D. (1973). Phonetics. London: Penguin Books. Odisho, Edward Y. (1977). Arabic /q/: a voiceless unaspirated uvular plosive. Lingua, Vol. 42, 343–347. ——— (1977). The opposition/ t / vs. / th / in Neo-Aramaic. Journal of the International Phonetic Association, Vol. 7, 79– 83. ——— (1979/a). An emphatic alveolar affricate, Journal of the International Phonetic Association, Vol. 9, 67–71. ——— (1979/b). Consonant clusters and abutting consonants. System, Vol. 7, 205–210.
246
PRONUNCIATION IS IN THE BRAIN
——— (1981). Teaching Arabic emphatics to the English learners of Arabic. Papers and Studies in Contrastive Linguistics, Vol. 13, 275–280. ——— (1988a). The sound system of modern Assyrian (NeoAramaic). Wiesbaden: Otto Harrassowitz Velag. ——— (1988b). Sibawaihi’s Dichotomy of ‘majhūra’ and ‘mahmūsa’ Revisited. Al-arabiyya. Vol. 21, 81–90. ——— (1990) Phonetic and phonological description of the labio-palatal and labio-velar approximants in Neo-Aramaic. Wolfhart Heinrichs (ed). Studies in Neo-Aramaic, 29–33. Atlanta, Georgia: Scholars Press. ——— (1992). Transliterating English in Arabic. Zeitschrift für arabische Linguistik, Vol. 24, 21–34. ——— (2002). Bilingualism: a salient and dynamic feature of ancient civilizations. Mediterranean Language Review, Vol. 14, 71–97. ——— (2003). Techniques of teaching pronunciation in ESL, bilingual and foreign language classes. München: Lincom-Europa. ——— (2004). A linguistic apoproach to the application and teaching of the English language. New York: Edwin Mellen Press. ——— (2005). Techniques of teaching comparative pronunciation in Arabic and English. New Jersey: Gorgias Press. ——— (2007/a). A Multisensory, Multicognitive Approach to Teaching Pronunciation. Revista de Estudos Linguisticos da Universidade do Porto, Vol. 2, 3–28. ——— (2007/b). Linguistic tips for Latino learners and teachers of English. New Jersey: Gorgias Press. ——— (2010). An Aerodynamic, Proprioceptive and Perceptual Interpretation of Sibawayhi’s Misplacement of /ط/ and /ق/ with Majhūra Consonants. Zeitschrift für Arabische Linguistik, Vol. 52, 39–52. ——— (2013). Some primary sources of accent generation in the pronunciation of English by native Arabs. Nicht nur mit Engelzungen (Beiträge zum semitischen Dialektologie: Festschrift für Werner Arnold zum 60 Geburstag, eds. R. Kuty, U Seeger and Sh. Talay, 265–274). Harrassowitz Verlag, Wiesbaden. Pennington, Martha C. and Richards, Jack C. (1986). Pronunciation revisited. TESOL Quarterly, Vol. 20/2, 207–225.
REFERENCES
247
Petitto, Laura-Ann (2002) http://www.dartmouth.edu/~news/rel eases/ 2002/nov/ 110402a.html. Pinker, S. (1994). The language instinct. New York: Harper Perennial Modern Classics. Port, Robert F. (2007).The graphical basis of phones and phonemes. Murray Munro and Ocke-Schwen Bohn, eds. 349– 365) Second-language speech learning: the role of language experience in speech perception and production. Amsterdam: John Benjamin Publishing Co. Repetti, Lori (2012). Consonant-final loanwords and epenthetic vowels in Italian. Catalan Journal of Linguistics, Vol. 11, 167–188. Roach, Peter (1983). English phonetics and phonology. Cambridge: Cambridge University Press. Selinker, Larry (1972). Interlanguage. International Review of Applied Linguistics, Vol. 10, 209–231. Shiver, Elaine (2001). Brain development and mastery of language in the early childhood years. (http://www.idra. org/IDRA_Newsletter/April_2001_Self_ Renewing_Schools_Early_ Childhood/Brain_Development_and_Mastery_of_Language_in_ the_Early_Childhood_Years/) Stockwell, R. P. and Bowen, J. (1965). The sounds of English and Spanish. Chicago: The University of Chicago Press. Tortora, G. and Grabowski, S. (1996). Principles of anatomy and physiology,(8th ed.), New York: HarperCollins College Publishers. Vicentini, Alessandra (2003). The economy principle in language: notes and observations from early Modern English grammars. (http://www.ledonline.it/mpw/.) Werker, J.F. (1995). Exploring developmental changes in crosslanguage speech perception. (Lila R. Gleitman and Mark Liberman, eds.). An invitation to cognitive science. Cambridge, Massachusetts: MIT Press. Werker, Janet F. and Tees, Richard C. (2002). Cross-language speech perception: evidence for perceptual reorganization during the first year of life. Infant Behavior & Development, Vol. 25, 121–133. Werker, J.F., Yeung, H.H., & Yoshida, K.A. (2012). How do infants become experts at native-speech perception? Current Directions in Psychological Science, Vol. 21/4, 221–226.
248
PRONUNCIATION IS IN THE BRAIN
Wesson, Kenneth A. (Neurosciencehttp://www.sciencemaster.com/ columns/wesson/ wesson_part_03.ph). Whitley, M. S. (1986). Spanish/English contrasts: a course in Spanish linguistics. Washington, D.C.: Georgetown University Press. Zero to Three. How the brain develops (2009). (https://www. childwelfare.gov/pubs/issue_briefs/brain_development/how.cfm). Zheng, Yanli, Sproat, Richard, Guy, Liang, Shafranz, Izhak, Zhouz, Haolang, Suz, Jurafsky, Dan, Starr, Rebecca, Yoon, Su-Youn (2005). Accent detection and speech recognition for Shanghai-accented Mandarin. (http://www.iscaspeech.org/ archive/ interspeech_2005/i05_0217.html).