We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In most social psychological studies, researchers conduct analyses that treat participants as a random effect. This means that inferential statistics about the effects of manipulated variables address the question whether one can generalize effects from the sample of participants included in the research to other participants that might have been used. In many research domains, experiments actually involve multiple random variables (e.g., stimuli or items to which participants respond, experimental accomplices, interacting partners, groups). If analyses in these studies treat participants as the only random factor, then conclusions cannot be generalized to other stimuli, items, accomplices, partners, or groups. What are required are mixed models that allow multiple random factors. For studies with single experimental manipulations, we consider alternative designs with multiple random factors, analytic models, and power considerations. Additionally, we discuss how random factors that vary between studies, rather than within them, may induce effect size heterogeneity, with implications for power and the conduct of replication studies.
The conclusions close the manuscript and make four points. First, they review the macro-level observational expectations tested in Part II, and how my findings, obtained through a triangulation of different techniques, allow for a comprehensive picture of how war affected state formation throughout the entire region. Second, they bring together all case studies in Part III, noting how the historical evidence collected fits the expectations of the theory at a micro-level—e.g., considering the behavior of individual actors and the effects of narrow events like battles within wars—and does so with out-and-out consistency—i.e., case by case, almost without exception. Third, they reflect upon the scope of the theory, discussing many other cases that could be explained by the long-term effects of war outcomes. This discussion covers many regions and time periods, showing that classical bellicist theory not only can travel, but can also solves logical problems and empirical puzzles highlighted by previous scholarship. Finally, the conclusions suggest many lines of enquiry for future research that the book leaves open.
High variability phonetic training using perceptual tasks such as identification and discrimination tasks has often been reported to improve L2 perception. However, studies comparing the efficacy of different tasks on different measures are rare. Forty-four Catalan/Spanish bilingual learners of English were trained with identification or categorical discrimination tasks and were tested on both measures. Results showed that both methods were successful in improving the identification and discrimination of English vowels. Training with nonword stimuli generalized to new nonwords and real word stimuli, and improvement was maintained four months later. Cross-task effects may be related to the categorical nature of the discrimination task, which may entail a level of processing similar to that of identification training. Interestingly, whereas identification training improved identification more than discrimination training, discrimination training did not enhance discrimination more than identification training. This asymmetry may be explained by task differences in the amount and type of feedback used.
Population-based structural health monitoring (PBSHM) systems use data from multiple structures to make inferences of health states. An area of PBSHM that has recently been recognized for potential development is the use of multitask learning (MTL) algorithms that differ from traditional single-task learning. This study presents an application of the MTL approach, Joint Feature Selection with LASSO, to provide automatic feature selection. The algorithm is applied to two structural datasets. The first dataset covers a binary classification between the port and starboard side of an aircraft tailplane, for samples from two aircraft of the same model. The second dataset covers normal and damaged conditions for pre- and postrepair of the same aircraft wing. Both case studies demonstrate that the MTL results are interpretable, highlighting features that relate to structural differences by considering the patterns shared between tasks. This is opposed to single-task learning, which improved accuracy at the cost of interpretability and selected features, which failed to generalize in previously unobserved experiments.
Young children today are exposed to masks on a regular basis. However, there is limited empirical evidence on how masks may affect word learning. The study explored the effect of masks on infants’ abilities to fast-map and generalize new words. Seventy-two Chinese infants (43 males, Mage = 18.26 months) were taught two novel word-object pairs by a speaker with or without a mask. They then heard the words and had to visually identify the correct objects and also generalize words to a different speaker and objects from the same category. Eye-tracking results indicate that infants looked longer at the target regardless of whether a speaker wore a mask. They also looked longer at the speaker’s eyes than at the mouth only when words were taught through a mask. Thus, fast-mapping and generalization occur in both masked and not masked conditions as infants can flexibly access different visual cues during word-learning.
Inductive reasoning involves using existing knowledge to make predictions about novel cases. This chapter reviews and evaluates computational models of this fundamental aspect of cognition, with a focus on work involving property induction. The review includes early induction models such as similarity coverage, and the feature-based induction model, as well as a detailed coverage of more recent Bayesian and connectionist approaches. Each model is examined against benchmark empirical phenomena. Model limitations are also identified. The chapter highlights the major advances that have been made in our understanding of the mechanisms that drive induction, as well as identifying challenges for future modeling. These include accounting for individual and developmental differences and applying induction models to explain other forms of reasoning.
With the aid of recently proposed word embedding algorithms, the study of semantic relatedness has progressed rapidly. However, word-level representations are still lacking for many natural language processing tasks. Various sense-level embedding learning algorithms have been proposed to address this issue. In this paper, we present a generalized model derived from existing sense retrofitting models. In this generalization, we take into account semantic relations between the senses, relation strength, and semantic strength. Experimental results show that the generalized model outperforms previous approaches on four tasks: semantic relatedness, contextual word similarity, semantic difference, and synonym selection. Based on the generalized sense retrofitting model, we also propose a standardization process on the dimensions with four settings, a neighbor expansion process from the nearest neighbors, and combinations of these two approaches. Finally, we propose a Procrustes analysis approach that inspired from bilingual mapping models for learning representations that outside of the ontology. The experimental results show the advantages of these approaches on semantic relatedness tasks.
This chapter illustrates how to apply Bayesian reasoning when analyzing more than one case. The same principles that govern Bayesian updating with multiple pieces of evidence apply to Bayesian inference with multiple cases.
We reflect on the relative ‘success’ versus ‘failure’ of psychology as a research field, and we challenge the widelybheld notion that we are in a reproducibility (or replication) crisis. At the centre of our discussion is the question: does psychology have a future, qua science, if the phenomena it studies are changing all the time and contingent on fleeting contexts or historical conditions? This chapter describes how there is only a reproducibility crisis if we adopt assumptions and expectations that enact a substance ontology. In contrast, we describe how variability is to be expected if we adopt a process ontology. We argue that the way out of the current ‘crisis’ is therefore not necessarily more methodological and experimental rigour, but a fundamental shift in what we should expect from psychological phenomena. We call for a prioritization of understanding the ways in which phenomena are socially situated and context-contingent, rather than an unrealistic need to replicate.
This chapter discusses the practice of measurement in psychological research. Here, where we cast doubt on the basic assumptions and endeavours underlying the act of measuring in mainstream psychology. Next, we introduce the processual alternative, which stresses the study of activity as situated and coupled with an environment. This chapter explains how a process approach to ‘measurement’ is thus fundamentally different from the standard one, and can remedy existing issues related to non-ergodicity and the ecological fallacy. These ideas are illustrated by means of the concept of intelligence, which is undoubtedly one of psychology’s show-pieces of measurement.
Woolcock focuses on the utility of qualitative case studies for addressing the decision-maker’s perennial external validity concern: What works there may not work here. He asks how to generate the facts that are important in determining whether an intervention can be scaled and replicated in a given setting. He focuses our attention on three categories: 1) causal density, 2) implementation capability, and 3) reasoned expectations about what can be achieved by when. Experiments are helpful for sorting out causally simple outcomes like the impact of deworming, but they are less insightful when there are many causal pathways, feedback loops, and exogenous influences. Nor do they help sort out the effect of mandate, management capacity, and supply chains, or the way results will materialize – whether some will materialize before others or increase or dissipate over time. Analytic case studies, Woolcock argues, are the main method available for assessing the generalizability of any given intervention.
Learning new words and, subsequently, a lexicon, is a time-extended process requiring encoding of word-referent pairs, retention of that information, and generalization to other exemplars of the category. Some children, however, fail in one or more of these processes resulting in language delays. The present study examines the abilities of children who vary in vocabulary size (including both children with normal language (NL) and late talking (LT) children) across multiple timescales/processes – known and novel word mapping, novel word retention, and novel noun generalization. Results indicate that children with lower language skills suffer from deficits in quick in-the-moment mapping of known words compared to their NL peers, but age and vocabulary size rather than normative vocabulary ranking or NL/LT status better predicts performance on retention and generalization processes. Implications for understanding language development as a holistic process with multiple interacting variables are discussed.
Constructivist approaches to language acquisition predict that form-function mappings are derived from distributional patterns in the input, and their contextual embedding. This requires a detailed analysis of the input, and the integration of information from different contingencies. Regarding the acquisition of morphology, it is shown which types of information leads to the induction of (lexical) categories, and to paradigm building. Regarding the acquisition of word order, it is shown how languages with fixed or variable word order profit from stable syntactic hyperschemas, but require a more detailed analyses of the form-function contingencies to identify the underlying, more specific semantic, syntactic and morphological patterns. At a theoretical level, it is shown how findings from acquisition and processing converge into new linguistic theories that aim to account for regular as well as irregular phenomena in language.
Understanding, categorizing, and using implementation science theories, models, and frameworks is a complex undertaking. The issues involved are even more challenging given the large number of frameworks and that some of them evolve significantly over time. As a consequence, researchers and practitioners may be unintentionally mischaracterizing frameworks or basing actions and conclusions on outdated versions of a framework.
Methods:
This paper addresses how the RE-AIM (Reach, Effectiveness, Adoption, Implementation, and Maintenance) framework has been described, summarizes how the model has evolved over time, and identifies and corrects several misconceptions.
Results:
We address 13 specific areas where misconceptions have been noted concerning the use of RE-AIM and summarize current guidance on these issues. We also discuss key changes to RE-AIM over the past 20 years, including the evolution to Pragmatic Robust Implementation and Sustainability Model, and provide resources for potential users to guide application of the framework.
Conclusions:
RE-AIM and many other theories and frameworks have evolved, been misunderstood, and sometimes been misapplied. To some degree, this is inevitable, but we conclude by suggesting some actions that reviewers, framework developers, and those selecting or applying frameworks can do to prevent or alleviate these problems.
Machine-learning algorithms can be viewed as stochastic transformations that map training data to hypotheses. Following Bousquet and Elisseeff, we say such an algorithm is stable if its output does not depend too much on any individual training example. Since stability is closely connected to generalization capabilities of learning algorithms, it is of interest to obtain sharp quantitative estimates on the generalization bias of machine-learning algorithms in terms of their stability properties. We describe several information-theoretic measures of algorithmic stability and illustrate their use for upper-bounding the generalization bias of learning algorithms. Specifically, we relate the expected generalization error of a learning algorithm to several information-theoretic quantities that capture the statistical dependence between the training data and the hypothesis. These include mutual information and erasure mutual information, and their counterparts induced by the total variation distance. We illustrate the general theory through examples, including the Gibbs algorithm and differentially private algorithms, and discuss strategies for controlling the generalization error.
A grand challenge in representation learning is the development of computational algorithms that learn the explanatory factors of variation behind high-dimensional data. Representation models (encoders) are often determined for optimizing performance on training data when the real objective is to generalize well to other (unseen) data. This chapter provides an overview of fundamental concepts in statistical learning theory and the information-bottleneck principle. This serves as a mathematical basis for the technical results, in which an upper bound to the generalization gap corresponding to the cross-entropy risk is given. When this penalty term times a suitable multiplier and the cross-entropy empirical risk are minimized jointly, the problem is equivalent to optimizing the information-bottleneck objective with respect to the empirical data distribution. This result provides an interesting connection between mutual information and generalization, and helps to explain why noise injection during the training phase can improve the generalization ability of encoder models and enforce invariances in the resulting representations.
Problems in learning that sights, sounds, or situations that were once associated with danger have become safe (extinction learning) may explain why some individuals suffer prolonged psychological distress following traumatic experiences. Although simple learning models have been unable to provide a convincing account of why this learning fails, it has recently been proposed that this may be explained by individual differences in beliefs about the causal structure of the environment.
Methods
Here, we tested two competing hypotheses as to how differences in causal inference might be related to trauma-related psychopathology, using extinction learning data collected from clinically well-characterised individuals with varying degrees of post-traumatic stress (N = 56). Model parameters describing individual differences in causal inference were related to multiple post-traumatic stress disorder (PTSD) and depression symptom dimensions via network analysis.
Results
Individuals with more severe PTSD were more likely to assign observations from conditioning and extinction stages to a single underlying cause. Specifically, greater re-experiencing symptom severity was associated with a lower likelihood of inferring that multiple causes were active in the environment.
Conclusions
We interpret these results as providing evidence of a primary deficit in discriminative learning in participants with more severe PTSD. Specifically, a tendency to attribute a greater diversity of stimulus configurations to the same underlying cause resulted in greater uncertainty about stimulus-outcome associations, impeding learning both that certain stimuli were safe, and that certain stimuli were no longer dangerous. In the future, better understanding of the role of causal inference in trauma-related psychopathology may help refine cognitive therapies for these disorders.
Most theories and hypotheses in psychology are verbal in nature, yet their evaluation overwhelmingly relies on inferential statistical procedures. The validity of the move from qualitative to quantitative analysis depends on the verbal and statistical expressions of a hypothesis being closely aligned – that is, that the two must refer to roughly the same set of hypothetical observations. Here, I argue that many applications of statistical inference in psychology fail to meet this basic condition. Focusing on the most widely used class of model in psychology – the linear mixed model – I explore the consequences of failing to statistically operationalize verbal hypotheses in a way that respects researchers' actual generalization intentions. I demonstrate that although the “random effect” formalism is used pervasively in psychology to model intersubject variability, few researchers accord the same treatment to other variables they clearly intend to generalize over (e.g., stimuli, tasks, or research sites). The under-specification of random effects imposes far stronger constraints on the generalizability of results than most researchers appreciate. Ignoring these constraints can dramatically inflate false-positive rates, and often leads researchers to draw sweeping verbal generalizations that lack a meaningful connection to the statistical quantities they are putatively based on. I argue that failure to take the alignment between verbal and statistical expressions seriously lies at the heart of many of psychology's ongoing problems (e.g., the replication crisis), and conclude with a discussion of several potential avenues for improvement.
In this chapter, there are two types of probabilities that can be estimated: empirical probability and theoretical probability. Empirical probability is calculated by conducting a number of trials and finding the proportion that resulted in each outcome. Theoretical probability is calculated by dividing the number of methods of obtaining an outcome by the total number of possible outcomes. Adding together the probabilities of two different events will produce the probability that either one will occur. Multiplying the probabilities of two events together will produce the probability that both will occur at the same time or in succession. As the number of trials increases, the empirical probability and theoretical probability converge.
It is possible to build a histogram of empirical or theoretical probabilities. As the number of trials increases, the empirical and theoretical probability distributions converge. If an outcome is produced by adding together (or averaging) the results of events, the probability distribution is normally distributed. Because of this, it is possible to make inferences about the population based on sample data – a process called generalization. The mean of sample means converges to the population mean, and the standard deviation of means (the standard error) converges on the value.
Philosophers speculated that assocations might be crucial in guiding behavior, but it was Pavlov who first explored this question experimentally. He discovered that dogs would begin to salivate to stimuli that preceded the delivery of food; the greater the contiguity of these stimuli with food, the stronger the conditioning. He also discovered fundamental properties of conditioning including extinction, inhibition, second-order conditioning, counterconditioning, discrimination, and generalization. Watson and Raynor extended his research by showing that conditioning also occurred in humans; they showed that fear could be conditioned in an infant who became famous as Little Albert. Subsequent research used random and unpaired control groups to determine if a change in behavior was truly due to conditioning, or to sensitization or pseudoconditioning. Three responses proved particularly useful in this later research: fear conditioning (CER) and taste-aversion learning in rats, and the galvanic skin response (GSR) in humans.