To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This paper brings new evidence for prosodic correspondence, where prosodic units (e.g. main-stressed nuclei and prominent syllables) of morphologically related forms are compared. Since prosodic correspondence was formalized in Crosswhite's (1998) analysis of Chamorro, it has received almost no empirical discussion. I argue that Tgdaya Seediq (Austronesian, Atayalic) has vowel alternations that should be analyzed using prosodic correspondence. In Seediq, unsuffixed and suffixed forms tend to share the same stressed syllable nucleus. This vowel matching pattern cannot be explained as surface harmony, but it can be explained as the result of a constraint enforcing vowel identity of main-stressed nuclei in morphologically related forms. Unlike the categorical alternations analyzed by Crosswhite (1998), Seediq vowel matching is gradient and only emerges on a statistical level. Nevertheless, prosodic correspondence appears to be active in the synchronic grammar of Seediq; in a production experiment, speakers applied vowel matching to novel forms and even over-generalized it to environments not predicted by lexical statistics. Vowel matching is modeled in Maximum Entropy Harmonic Grammar (Goldwater & Johnson 2003), a stochastic variant of OT. I use prosodic correspondence to enforce vowel matching, and Zuraw's (2000, 2010) dual listing approach to capture the discrepancy between lexical and experimental results.
The goal of this article is to provide a balanced assessment of the significance autism has for the scientific study of language. While linguistic profiles in autism vary greatly, spanning from a total absence of functional language to verbal levels within the typical range, the entire autism spectrum is robustly characterized by lifelong disabilities in intersubjective communication and persistent difficulties in adopting the perspective of other people. In that sense, autism constitutes a unique profile in which linguistic competence is dissociated from communication skills. Somewhat paradoxically, autism is often mentioned to underscore the importance of mind reading for language use and of intersubjective communication for the emergence of language. Yet experimental studies on pragmatics in autism indicate that many pragmatic processes unfold without adopting one's conversational partner's perspective. Moreover, the patterns of language acquisition and learning in autism represent a strong challenge to the central role constructionist theories assign to socio-communicative skills. Data on autism thus force a reconsideration of the a priori conceptual boundaries on language learnability that shape the foundational debates between constructionist and nativist linguistic theories.
This article is an analysis of the claim that a universal ban on certain (‘anti-markedness’) grammars is necessary in order to explain their nonoccurrence in the languages of the world. Such a claim is based on the following assumptions: that phonological typology shows a highly asymmetric distribution, and that such a distribution cannot possibly arise ‘naturally’—that is, without a universal grammar-based restriction of the learner’s hypothesis space. Attempting to test this claim reveals a number of open issues in linguistic theory. In the first place, there exist critical aspects of synchronic theory that are not specified explicitly enough to implement computationally. Second, there remain many aspects of linguistic competence, language acquisition, sound change, and even typology that are still unknown. It is not currently possible, therefore, to reach a definitive conclusion about the necessity, or lack thereof, of an innate substantive grammar module. This article thus serves two main functions: acting both as a pointer to the areas of phonological theory that require further development, especially at the overlap between traditionally separate subdomains, and as a template for the type of argumentation required to defend or attack claims about phonological universals.
What psycholinguistic mechanisms shape the emergence of Creole languages, and are these processes unique or universal across human language evolution? In this exploration, determiner-noun fusion (DNF) in Haitian Creole takes center stage, challenging assumptions about the sole role of substrate influence. By analyzing DNF patterns in Haitian Creole and comparing them to those in Mauritian Creole, the chapter reveals how statistical learning – hallmarks of word segmentation – plays a pivotal role. These findings align Creole emergence with broader linguistic processes, refuting claims of a “break in transmission.” This chapter bridges Creole linguistics and psycholinguistics, providing support for the Uniformitarian Principle and reshaping the debate on Creole emergence.
Advances in natural language processing (NLP) and Big Data techniques have allowed us to learn about the human mind through one of its richest outputs – language. In this chapter, we introduce the field of computational linguistics and go through examples of how to find natural language and how to interpret the complexities that are present within it. The chapter discusses the major state-of-the-art methods being applied in NLP and how they can be applied to psychological questions, including statistical learning, N-gram models, word embedding models, large language models, topic modeling, and sentiment analysis. The chapter concludes with ethical discussions on the proliferation of chat “bots” that pervade our social networks, and the importance of balanced training sets for NLP models.
Music & spoken language share many features by combining smaller units (e.g., words, notes) into larger structures (e.g., sentences, musical phrases). This hierarchical organization of sound is culturally contingent & communicates meaning to listeners. Comparisons of music & language from a cognitive neuroscience perspective provide several insights into commonalities & differences between these systems, how they are represented in the brain. The cognitive neuroscience research of music & language, emphasizes the pitfalls & promises identified, including (1) the apparent acoustic & structural similarities between these systems, (2) how both systems convey meaning to listeners, (3) how these systems are learned over the course of development, & (4) the ways in which experience in one domain influences processing in the other domain. We conclude that searching for similarities in how these complex systems are structured (e.g., comparing musical syntax to linguistic syntax) represents a pitfall that researchers should approach with caution. A promising approach in this area of research is to examine how general cognitive mechanisms underlie the learning & maintenance of both systems
The study of individuals with hippocampal damage and amnesia provides a compelling opportunity to directly test the role of declarative memory to communication and language. Over the past two decades, we have documented disruptions in discourse and conversation as well as in more basic aspects of language in individuals with hippocampal amnesia including at the word, phrase, and sentence level across offline and online language processing tasks. This work highlights the critical contribution of hippocampal-dependent memory to language and communication and suggests that hippocampal damage or dysfunction is a risk factor for a range of language and communicative disruptions even in the absence of frank disorders of amnesia or aphasia. This work also raises questions about the reality and utility of the historical distinction between communication and language in defining cognitive-communication disorders as individuals with isolated memory impairments show deficits that cut across both communication and language.
Bridging theory and practice in network data analysis, this guide offers an intuitive approach to understanding and analyzing complex networks. It covers foundational concepts, practical tools, and real-world applications using Python frameworks including NumPy, SciPy, scikit-learn, graspologic, and NetworkX. Readers will learn to apply network machine learning techniques to real-world problems, transform complex network structures into meaningful representations, leverage Python libraries for efficient network analysis, and interpret network data and results. The book explores methods for extracting valuable insights across various domains such as social networks, ecological systems, and brain connectivity. Hands-on tutorials and concrete examples develop intuition through visualization and mathematical reasoning. The book will equip data scientists, students, and researchers in applications using network data with the skills to confidently tackle network machine learning projects, providing a robust toolkit for data science applications involving network-structured data.
Non-native languages tend to be acquired through a combination of explicit and implicit learning, where implicit learning requires coordination of language information with referents in the environment. In this study, we examined how learners use both language input and environmental cues to acquire vocabulary and morphology in a novel language and how their language background influences this process. We trained 105 adults with native languages (L1s) varying in morphological richness (English, German, Mandarin) on an artificial language comprising nouns and verbs with morphological features (number, tense, and subject-verb [SV] agreement) appearing alongside referential visual scenes. Participants were able to learn both word stems and morphological features from cross-situational statistical correspondences between language and the environment, without any instruction. German-speakers learned SV agreement worse than other morphological features, which were acquired equally effectively by English or Mandarin speakers, indicating the subtle and varied influence of L1 morphological richness on implicit non-native language learning.
Language models can produce fluent, grammatical text. Nonetheless, some maintain that language models don’t really learn language and also that, even if they did, that would not be informative for the study of human learning and processing. On the other side, there have been claims that the success of LMs obviates the need for studying linguistic theory and structure. We argue that both extremes are wrong. LMs can contribute to fundamental questions about linguistic structure, language processing, and learning. They force us to rethink arguments and ways of thinking that have been foundational in linguistics. While they do not replace linguistic structure and theory, they serve as model systems and working proofs of concept for gradient, usage-based approaches to language. We offer an optimistic take on the relationship between language models and linguistics.
Understanding the mechanisms of major depressive disorder (MDD) improvement is a key challenge to determining effective personalized treatments.
Methods
To identify a data-driven pattern of clinical improvement in MDD and to quantify neural-to-symptom relationships according to antidepressant treatment, we performed a secondary analysis of the publicly available dataset EMBARC (Establishing Moderators and Biosignatures of Antidepressant Response in Clinical Care). In EMBARC, participants with MDD were treated either by sertraline or placebo for 8 weeks (Stage 1), and then switched to bupropion according to clinical response (Stage 2). We computed a univariate measure of clinical improvement through a principal component (PC) analysis on the variations of individual items of four clinical scales measuring depression, anxiety, suicidal ideas, and manic-like symptoms. We then investigated how initial clinical and neural factors predicted this measure during Stage 1 by running a linear model for each brain parcel’s resting-state global brain connectivity (GBC) with individual improvement scores during Stage 1.
Results
The first PC (PC1) was similar across treatment groups at stages 1 and 2, suggesting a shared pattern of symptom improvement. PC1 patients’ scores significantly differed according to treatment, whereas no difference in response was evidenced between groups with the Clinical Global Impressions Scale. Baseline GBC correlated with Stage 1 PC1 scores in the sertraline but not in the placebo group.
Using data-driven reduction of symptom scales, we identified a common profile of symptom improvement with distinct intensity between sertraline and placebo.
Conclusions
Mapping from data-driven symptom improvement onto neural circuits revealed treatment-responsive neural profiles that may aid in optimal patient selection for future trials.
Recent evidence from cross-situational learning (CSL) studies have shown that adult learners can acquire words and grammar simultaneously when sentences of the novel language co-occur with dynamic scenes to which they refer. Syntactic bootstrapping accounts suggest that grammatical knowledge may help scaffold vocabulary acquisition by constraining possible meanings, thus, for children, words and grammar may be acquired at different rates. Twenty children (ages 8 to 9) were exposed in a CSSL study to an artificial language comprising nouns, verbs, and case markers occurring within a verb-final grammatical structure. Children acquired syntax (i.e., word order) effectively, but we found no evidence of vocabulary learning, whereas previous adult studies showed learning of both from similar input. Grammatical information may thus be available early for children, to help constrain and support later vocabulary learning. We propose that gradual maturation of declarative memory systems may result in more effective vocabulary learning in adults.
Describe the challenges children face in learning language; understand key features of child language development; explain the strategies children use to learn sounds, words, and grammar.
Studies investigating phonological processing indicate that words with high regularity/consistency in pronunciation or high frequency positively impact reading speed and accuracy. Such effects of consistency and frequency have been demonstrated in Japanese kanji words and are known as consistency and frequency effects. Using a mixed-effects model analysis, this study reexamines the two effects in Chinese–Japanese second-language (L2) learners with two different L2 proficiency levels. The two effects are robustly replicated in oral reading tasks; in particular, the performance of intermediate learners is similar to that of Japanese semantic dementia patients, whose reading accuracy is affected by sensitivity to the statistical properties of words (i.e., reading consistency and lexical frequency). These results are explained by the interaction between semantic memory and word statistical properties. Moreover, the interaction highlights the important consequences of statistical learning underlying L2 phonological processing.
In today’s insurance market, numerous cyber insurance products provide bundled coverage for losses resulting from different cyber events, including data breaches and ransomware attacks. Every category of incident has its own specific coverage limit and deductible. Although this gives prospective cyber insurance buyers more flexibility in customizing the coverage and better manages the risk exposures of sellers, it complicates the decision-making process in determining the optimal amount of risks to retain and transfer for both parties. This article aims to build an economic foundation for these incident-specific cyber insurance products with a focus on how incident-specific indemnities should be designed for achieving Pareto optimality for both the insurance seller and the buyer. Real data on cyber incidents are used to illustrate the feasibility of this approach. Several implementation improvement methods for practicality are also discussed.
Connectionist networks consisting of large numbers of simple connected processing units implicitly or explicitly model aspects of human predictive behavior. Prediction in connectionist models can occur in different ways and with quite different connectionist architectures. Connectionist neural networks offer a useful playground and ‘hands-on way’ to explore prediction and to figure out what may be special about how the human mind predicts.
The suffixing bias (the tendency to exploit suffixes more often than prefixes to express grammatical meanings) in languages was identified a century ago, yet we still lack a clear account for why it emerged, namely, whether the bias emerged because general cognitive mechanisms shape languages to be more easily processed by available cognitive machinery, or if the bias is speech-specific and is determined by domain-specific mechanisms. We used statistical learning (SL) experiments to compare processing of suffixed and prefixed sequences on linguistic and non-linguistic material. SL is not speech-specific, and we observed the suffixing preference only on linguistic material, suggesting its language-specific origin. Moreover, morphological properties of native languages (existence of grammatical prefixes) modulate suffixing preferences in SL experiments only on linguistic material, suggesting limited cross-domain transfer.
The present study examined whether length of bilingual experience and language ability contributed to cross-situational word learning (XSWL) in Spanish-English bilingual school-aged children. We contrasted performance in a high variability condition, where children were exposed to multiple speakers and exemplars simultaneously, to performance in a condition where children were exposed to no variability in either speakers or exemplars. Results revealed graded effects of bilingualism and language ability on XSWL under conditions of increased variability. Specifically, bilingualism bolstered learning when variability was present in the input but not when variability was absent in the input. Similarly, robust language abilities supported learning in the high variability condition. In contrast, children with weaker language skills learned more word-object associations in the no variability condition than in the high variability condition. Together, the results suggest that variation in the learner and variation in the input interact and modulate mechanisms of lexical learning in children.
In Chapter 13 we will discuss how to produce compression schemes that do not require a priori knowledge of the generative distribution. It turns out that designing a compression algorithm able to adapt to an unknown distribution is essentially equivalent to the problem of estimating an unknown distribution, which is a major topic of statistical learning. The plan for this chapter is as follows: (1) We will start by discussing the earliest example of a universal compression algorithm (of Fitingof). It does not talk about probability distributions at all. However, it turns out to be asymptotically optimal simultaneously for all iid distributions and with small modifications for all finite-order Markov chains. (2) The next class of universal compressors is based on assuming that the true distribution belongs to a given class. These methods proceed by choosing a good model distribution serving as the minimax approximation to each distribution in the class. The compression algorithm for a single distribution is then designed as in previous chapters. (3) Finally, an entirely different idea are algorithms of Lempel–Ziv type. These automatically adapt to the distribution of the source, without any prior assumptions required.
Prediction and classification are two very active areas in modern data analysis. In this paper, prediction with nonlinear optimal scaling transformations of the variables is reviewed, and extended to the use of multiple additive components, much in the spirit of statistical learning techniques that are currently popular, among other areas, in data mining. Also, a classification/clustering method is described that is particularly suitable for analyzing attribute-value data from systems biology (genomics, proteomics, and metabolomics), and which is able to detect groups of objects that have similar values on small subsets of the attributes.