We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Twenty-five years ago, the publication of an article by Pallier, Colomé, and Sebastián-Gallés (2001) launched a new and rapidly evolving research program on how second language (L2) learners represent the phonological forms of words in their mental lexicons. Many insights are starting to form an overall picture of the unique difficulties for establishing functional and precise phonolexical representations in L2; however, for the field to move forward it is pertinent to outline its major emerging research questions and existing challenges. Among significant obstacles for further research, the paper explores the current lack of theoretical agreement on the concept of phonolexical representations and the underlying mechanism involved in establishing them, as well as the variable use of the related terminology (e.g., fuzziness and target-likeness). Methodological challenges involved in investigating phonological processing and phonolexical representations as well as their theoretical implications are also discussed. To conclude, we explore the significance of L2-specific phonological representations for the bottom-up lexical access during casual, conversational speech and how our emerging knowledge of L2 lexical representations can be applied in an instructional setting as two potentially fruitful research avenues at the forefront of the current research agenda.
Study abroad is typically viewed as a catalyst for pronunciation learning because it affords learners both massive amounts of L2 input and abundant opportunities for meaningful L2 use. Yet, even in such an environment, there is substantial variability in learning trajectories and outcomes. The nature of the target structure is also a powerful determinant of learning; some structures seem to develop effortlessly, whereas others do not improve much at all. Additionally, study abroad research brings to light the important issue of speaker identity, as learners often make decisions about how they want to sound and what pronunciation features they will adopt. This chapter examines developmental time frames, trajectories, and turning points in the phonetics and phonology of L2 learners in a study abroad context. We also describe how learners acquire the regional pronunciation variants of their host communities considering the phonetics of the target feature and learners’ attitudes and beliefs. We argue that study abroad should be situated within a dynamic, longitudinal, and context-dependent view of phonetic and phonological learning.
Researchers in bilingualism seek to identify factors that are associated with specific features of bilingual speech. One such predictive factor is language dominance, typically understood as the degree to which one of the languages of a bilingual is more often and more proficiently used. In this chapter we review landmark studies that demonstrate the power of language dominance in predicting fine-grained phonetic and phonological characteristics of speech production and on the perceptual and processing abilities in one or both languages of bilinguals. We then critically examine the construct of dominance and identify ways that dominance can be and has been measured, as well as challenges inherent in the measurement of dominance. We follow demonstrating the dynamic character of dominance by reviewing research on dominance switches and shifts. This is followed by a review of extant studies on language dominance in bilingual speech production, perception, and processing in both languages. We conclude with four areas where research can be fruitfully directed.
The Automatic Selective Perception (ASP) model posits that listeners make use of selective perceptual routines (SPRs) that are fast and efficient for recovering lexical meaning. These SPRs serve as filters to accentuate relevant cues and minimize irrelevant information. Years of experience with the first language (L1) lead to fairly automatic L1 SPRs; consequently, few attentional resources are needed in processing L1 speech. In contrast, L2 SPRs are less automatic. Under difficult task or stimulus conditions, listeners fall back on more automatic processes, specifically L1 SPRs. And L2 speech perception suffers where there is a mismatch between the L1 and the L2 phonetics because L1 SPRs may not extract the important cues needed for identifying L2 phonemes. This chapter will present behavioral and neurophysiology evidence that supports the ASP model, but which also indicates the need for some modification. We offer suggestions for future directions in extending this model.
This chapter provides a thorough, up-to-date review of the literature on the phonetics and phonology of early bilinguals. It pulls together studies from a range of bilingual settings, including bilingual societies and heritage language contexts. While the chapter mostly reviews evidence from adolescent and adult participants, it also makes reference to the child bilingualism literature, where appropriate. The chapter first reviews studies on the accents of early versus late bilinguals, followed by a discussion of the various explanatory accounts for the observed differences between these two groups. Subsequently, the critical significance of early linguistic experience on bilingual speech patterns is considered, with particular reference to the evidence from childhood overhearers and international adoptees. The following sections then review studies comparing simultaneous and early sequential bilinguals, and those exploring the role of language dominance, continued use, the language of the environment in bilinguals’ pronunciation patterns, and the role of sociolinguistic factors in early bilingual speech patterns. The chapter concludes with suggestions for future research.
This chapter reviews evidence that the orthographic forms (spellings) of L2 sounds and words affect L2 phonological representation and processing. Orthographic effects are found in speech perception, speech production, phonological awareness, and the learning of words and sounds. Orthographic forms facilitate L2 speakers/listeners – for instance in lexical learning – but also have negative effects, resulting in sound additions, deletions, and substitutions. This happens because L2 speakers’ L2 orthographic knowledge differs from the actual working of the L2 writing system. Orthographic effects are established after little exposure to orthographic forms, are persistent, can be reinforced by factors other than orthography, including spoken input, and are modulated by individual-level and sound/word-level variables. Future research should address gaps in current knowledge, for instance investigating the effects of teaching interventions, and aim at producing a coherent framework.
This chapter outlines studies within the domain of speech perception by bilingual adult listeners. I first discuss studies that have examined bilinguals’ perception of L1 and/or L2 speech segments, as well as those that have tested perception of unfamiliar, non-native speech segments. In turn, I examine each of the factors that are known to affect bilinguals’ perception of speech, which include age of L2 acquisition, effects of L1:L2 usage as they pertain to language dominance and proficiency, and short-term contextual effects on speech perception. I also provide an overview of the literature on bilinguals’ perception of suprasegmentals. Finally, I explore what I think are some of the crucial questions facing the field of bilingual speech perception.
This chapter examines the acquisition of vowels and consonants in perception and production during typical second language (L2) acquisition by sequential, dominant bilinguals. The acoustic and articulatory studies reviewed serve to illustrate general patterns of L2 segmental learning with a focus on four principal themes, each structured around one or two main research questions: 1) cross-linguistic influence (How does a sequential, dominant bilingual’s first language shape L2 segmental perception and production?); 2) development (Are there universal patterns to L2 segmental speech development? Are certain vocalic and consonantal phenomena acquired more easily?); 3) inter- and intralearner variability (What are the sources of differences in perception and/or production between learners and for the same learner over time and in different communicative contexts?); and 4) training effects (What are the effects of training on learning? How do variables such as task and stimuli type condition effectiveness?).
Evidence that listeners attend to and track subphonemic phonetic details is indicated by listeners’ ability to reliably connect subphonemic variation and, often, socio-indexical associations in ways that align with the patterns realized in production. Bilinguals are presented with the task not only of associating within-language variation (e.g., social group X is connected to a particular range of phonetic realizations within language Y) but also of attending to how ethnolects and bilingually accented speech index social categories across languages. Having access to multiple languages also gives bilingual speakers a larger repertoire with which to index language- and community-specific social meaning. This chapter outlines the linguistic structures bilinguals may connect across their languages and then presents a specific exemplar model, noting the opportunities within the model’s structure for bilingual dynamics. The heterogeneity of bilingual individuals and speech communities is necessarily addressed, as this dynamic adds to the complexity and intrigue of studying bilingual populations.
In this chapter, we thoroughly describe the L2LP model, its five ingredients to explain speech development from first contact with a language or dialect (initial state) to proficiency comparable to a native speaker of the language or dialect (ultimate attainment), and its empirical, computational, and statistical method. We present recent studies comparing different types of bilinguals (simultaneous and sequential) and explaining their differential levels of ultimate attainment in different learning scenarios. We also show that although the model has the word “perception” in its name, it was designed to also explain phonological development in general, including lexical development, speech production, and orthographic effects. The chapter demonstrates that the L2LP model can be regarded as a comprehensive theoretical, computational, and probabilistic model or framework for explaining how we learn the phonetics and phonology of multiple languages (sequentially or simultaneously) with variable levels of language input throughout the life span.
While the pathogenesis of sudden sensorineural hearing loss is thought to be localised to the cochlea, recent microRNA findings suggest a neuro-topic localisation in some patients. This study distinguishes if neural and non-neural groups differ in hearing recovery.
Methods
Neural-type hearing loss was defined as a presenting word recognition score less than 60 per cent, with a word recognition score reduction greater than 20 per cent than expected based on the averaged pure tone audiometry. Hearing recovery was defined as an improvement of greater than or equal to 10 decibels in pure tone audiometric thresholds.
Results
Eight of 12 and 24 of 36 of neural and non-neural hearing loss patients demonstrated hearing recovery, respectively. The affected ear's word recognition score (per cent) change with treatment were different between the neural and non-neural groups (46.9 ± 29.8 vs 3.2 ± 25.8 (p < 0.0001)).
Conclusion
The hearing recovery rate in neural and non-neural hearing loss groups was similar. Patients with neural-type hearing loss demonstrated greater word recognition score recovery post treatment than those in the sensory group.
Spoken language is a rapidly unfolding signal: a complex code that the listener must crack to understand what is being said. From the structures of the inner ear through to higher-order areas of the brain, a hierarchy of interlinked processes transforms the acoustic signal into a linguistic message within fractions of a second. This Element outlines how we perceive speech and explores what the auditory system needs to achieve to make this possible. It traces a path through the system and discusses the mechanisms that enable us to perceive speech as a coherent sequence of words. This is combined with a brief history of research into language and the brain beginning in the nineteenth century, as well as an overview of the state-of-the-art neuroimaging and analysis techniques that are used to investigate phonetics in the brain today. This title is also available as Open Access on Cambridge Core.
Ideophones – imitative words using the stream of speech to simulate/depict the rise and fall of sensory perceptions and emotions and temporal experiences of completiveness, instantaneousness, and repetitiveness – have been characterized as semantically empty and context-dependent. The research reported here tested a simple schematic for the semantic categories of Pastaza Kichwa ideophones by tracking neurological responses to ideophones categorized as VISUAL, MOTION, and SOUND. Seventeen native speakers of Pastaza Kichwa listened to audio clips of ideophones extracted from sentential contexts. Subjects’ neural activity was assessed using functional near-infrared spectroscopy. Results demonstrate that these posited semantic categories activate areas of the brain associated with visualization, motion, and sound processing and production, respectively. This suggests that these ideophones convey semantic information related to these concepts, independent of context. This contrasts with what would be expected by theories suggesting that ideophones on their own are semantically empty. The data give rise to questions regarding whether any language contains only sound ideophones that do not carry additional sensory information and whether ideophones in previous studies treated strictly as sound ideophones might require greater specification of their semantics, specifically from a multisensorial perspective.
Spoken threats are a common but linguistically complex language crime. Although threatening language has been examined from different linguistic perspectives, there is limited research which critically addresses how people perceive spoken threats and infer traits such as threat and intent from speakers' voices. There is also minimal linguistic research addressing differences between written and spoken threats. By specifically analysing threats delivered in both written and spoken modalities, as well as integrating perceptual phonetic analysis into discussions on spoken threats, this Element offers perspectives on these two under-researched areas. It highlights the dangers of assuming that the way in which someone sounds correlates with, for example, their intention to commit harm, and explores potential problems in assuming that written and spoken threats are equivalent to one another. The goal of the Element is to advance linguistic knowledge and understanding around spoken threats, as well as promote further research in the area.
Silent pauses are a natural part of speech production and have consequences for speech perception. However, studies have shown mixed results regarding whether listeners process pauses in native and non-native speech similarly or differently. A possible explanation for these mixed results is that perceptual consequences of pauses differ depending on the type of processing that listeners engage in: a focus on the content/meaning of the speech versus style/form of the speech. Thus, the present study examines the effect of silent pauses of listeners’ perception of native and non-native speech in two different tasks: the perceived credibility and the perceived fluency of the speech. Specifically, we ask whether characteristics of silent pauses influence listeners’ perception differently for native versus non-native speech, and whether this pattern differs when listeners are rating the credibility versus the fluency of the speech. We find that while native speakers are rated as more fluent than non-native speakers, there is no evidence that native speakers are rated as more credible. Our findings suggest that the way a non-native accent and disfluency together impact speech perception differs depending on the type of processing that listeners are engaged in when listening to the speech.
The present study investigates the interplay between proficiency and empathy in the development of second language (L2) prosody by analyzing the perception and processing of intonation in questions and statements in L2 Spanish. A total of 225 adult L2 Spanish learners (L1 English) from the Northeastern United States completed a two-alternative forced choice (2AFC) task in which they listened to four utterance types and categorized them as either questions or statements. We used Bayesian multilevel regression and drift diffusion modeling to analyze the 2AFC data as a function of proficiency level and empathy scores for each utterance type. We show that learner response accuracy and sensitivity to intonation are positively correlated with proficiency, and this association is affected by individual empathy levels in both response accuracy and sentence processing. Higher empathic individuals, in comparison with lower empathic individuals, appear to be more sensitive to intonation cues in the process of forming sound-meaning associations, though increased sensitivity does not necessarily imply increased processing speed. The results motivate the inclusion of measures of pragmatic skill, such as empathy, to better account for intonational meaning processing and sentence comprehension in second language acquisition.
Ideally, comprehensible second language (L2) speech would be seen as acceptable speech. However, the association between these dimensions is underexplored. To investigate the relationship between comprehensibility and “academic acceptability,” defined here as how well a speaker could meet the demands of a given role in an academic setting, 204 university stakeholders judged L2 speech samples elicited from a standardized English test used for university admissions. Four tasks from 100 speakers were coded for 13 speech stream characteristics. Judgments for comprehensibility and acceptability correlated strongly (r = .93). Linear mixed-effects models, used to examine judgments across all tasks and separately for each task, indicated that while random intercepts (i.e., speaker ability, listener severity) explained a substantial amount of total variation (32–44%) in listener judgments compared to speech characteristic fixed effects (8–21%), fixed effects did account for variation in speaker random effects (reducing variation compared to intercept-only models by 50–90%). Despite some minimal differences across task types, the influence of speech characteristics across both judgments was mostly similar. While providing evidence that comprehensible speech can indeed be perceived as acceptable, this study also provides evidence that speakers demonstrate both consistent and less consistent performance, in reference to speech stream production, across performances.
This chapter first notes progress in finding ‘new’ sounds in languages and advances in (computer-related) instrumentation, which underpinned discoveries in phonetics and experimental phonology, 1950-2000, in three areas.
1. speech production, e.g., control of air-streams, phonation, articulation, hidden articulation (underarticulation), ‘covert contrast’, and co-articulation.
2. advances in speech acoustics, e.g., models of the relationship between articulation and the resulting sound (FFT, LPC, ‘Distinctive Regions and Modes Theory’).
3. new findings in speech perception:
a) context-sensitivity of perceptual cues;
b) categorical perception;
c) variability and the search for invariance, including the ‘quantal nature of speech’; the (connectionist) ‘Fuzzy Logical Model of Phoneme Identification’; models of speech recognition from acoustic data: LAFS (‘Lexical Access From Spectra’) and ‘TRACE’; exemplar-based approaches to phonology; the probabilistic turn in Laboratory Phonology; evidence of the importance of frequency effects and the lexical status of stimuli in perception; holistic models of speech perception, including the integration of multiple cues, which combine expectations with bottom-up acoustic processing; ‘the phoneme restoration effect’ with deleted phonemes; statistical, pattern-matching approach to speech perception; evidence that phonological representations may be rich in phonetic detail and memorized from experience.
Work of this sort might unite experimental/theoretical phonetics/phonology, in the 21st c.
Non-native speech is difficult for native listeners to understand. While listeners can learn to understand non-native speech after exposure, it is unclear how to optimize this learning. Experimental subjects transcribed non-native speech and were paid either a flat rate or based on their performance. Participants who were paid based on performance demonstrated improved performance overall and faster learning than participants who were paid a flat rate. These results suggest that exposure alone is not sufficient to optimize learning of non-native speech and that current models of this process must be revised to account for the effects of motivation and incentive.
This Element focuses on phonetic and phonological development in multilinguals and presents a novel methodological approach to it within Complex Dynamic Systems Theory (CDST). We will show how the traditional conceptualisations of acquisition with a strong focus on linear, incremental development with a stable endpoint can be complemented by a view of language development as emergent, self-organised, context-dependent and highly variable across learners. We report on a longitudinal study involving 16 learners with L1 German, L2 English and L3 Polish. Over their ten months of learning Polish, the learners' perception and production of various speech sounds and phonological processes in all of their languages were investigated. Auditory and acoustic analyses were applied together with group and individual learner statistical analyses to trace the dynamic changes of their multilingual phonological system over time. We show how phonetic and phonological development is feature-dependent and inter-connected and how learning experience affects the process.