We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
A popular refrain in many countries is that people with mental illnesses have “nowhere to go” for care. But that is not universally true. Previously unexplored international data shows that some countries provide much higher levels of public mental health care than others. This puzzling variation does not align with existing scholarly typologies of social or health policy systems. Furthermore, these cross-national differences are present despite all countries’ shared history of psychiatric deinstitutionalization, a process that I conceptualize and document using an original historical data set. I propose an explanation for countries’ varying policy outcomes and discuss an empirical strategy to assess it. The research design focuses on the cases of the United States and France, along with Norway and Sweden, in order to control for a range of case-specific alternative hypotheses. The chapter ends with brief descriptions of contemporary mental health care policy in each of the four countries examined in this book.
Information on the time spent completing cognitive testing is often collected, but such data are not typically considered when quantifying cognition in large-scale community-based surveys. We sought to evaluate the added value of timing data over and above traditional cognitive scores for the measurement of cognition in older adults.
Method:
We used data from the Longitudinal Aging Study in India-Diagnostic Assessment of Dementia (LASI-DAD) study (N = 4,091), to assess the added value of timing data over and above traditional cognitive scores, using item-specific regression models for 36 cognitive test items. Models were adjusted for age, gender, interviewer, and item score.
Results:
Compared to Quintile 3 (median time), taking longer to complete specific items was associated (p < 0.05) with lower cognitive performance for 67% (Quintile 5) and 28% (Quintile 4) of items. Responding quickly (Quintile 1) was associated with higher cognitive performance for 25% of simpler items (e.g., orientation for year), but with lower cognitive functioning for 63% of items requiring higher-order processing (e.g., digit span test). Results were consistent in a range of different analyses adjusting for factors including education, hearing impairment, and language of administration and in models using splines rather than quintiles.
Conclusions:
Response times from cognitive testing may contain important information on cognition not captured in traditional scoring. Incorporation of this information has the potential to improve existing estimates of cognitive functioning.
Several methods used to examine differential item functioning (DIF) in Patient-Reported Outcomes Measurement Information System (PROMIS®) measures are presented, including effect size estimation. A summary of factors that may affect DIF detection and challenges encountered in PROMIS DIF analyses, e.g., anchor item selection, is provided. An issue in PROMIS was the potential for inadequately modeled multidimensionality to result in false DIF detection. Section 1 is a presentation of the unidimensional models used by most PROMIS investigators for DIF detection, as well as their multidimensional expansions. Section 2 is an illustration that builds on previous unidimensional analyses of depression and anxiety short-forms to examine DIF detection using a multidimensional item response theory (MIRT) model. The Item Response Theory-Log-likelihood Ratio Test (IRT-LRT) method was used for a real data illustration with gender as the grouping variable. The IRT-LRT DIF detection method is a flexible approach to handle group differences in trait distributions, known as impact in the DIF literature, and was studied with both real data and in simulations to compare the performance of the IRT-LRT method within the unidimensional IRT (UIRT) and MIRT contexts. Additionally, different effect size measures were compared for the data presented in Section 2. A finding from the real data illustration was that using the IRT-LRT method within a MIRT context resulted in more flagged items as compared to using the IRT-LRT method within a UIRT context. The simulations provided some evidence that while unidimensional and multidimensional approaches were similar in terms of Type I error rates, power for DIF detection was greater for the multidimensional approach. Effect size measures presented in Section 1 and applied in Section 2 varied in terms of estimation methods, choice of density function, methods of equating, and anchor item selection. Despite these differences, there was considerable consistency in results, especially for the items showing the largest values. Future work is needed to examine DIF detection in the context of polytomous, multidimensional data. PROMIS standards included incorporation of effect size measures in determining salient DIF. Integrated methods for examining effect size measures in the context of IRT-based DIF detection procedures are still in early stages of development.
A method is discussed which extends canonical regression analysis to the situation where the variables may be measured at a variety of levels (nominal, ordinal, or interval), and where they may be either continuous or discrete. There is no restriction on the mix of measurement characteristics (i.e., some variables may be discrete-ordinal, others continuous-nominal, and yet others discrete-interval). The method, which is purely descriptive, scales the observations on each variable, within the restriction imposed by the variable's measurement characteristics, so that the canonical correlation is maximal. The alternating least squares algorithm is discussed. Several examples are presented. It is concluded that the method is very robust. Inferential aspects of the method are not discussed.
It is common practice in IRT to consider items as fixed and persons as random. Both, continuous and categorical person parameters are most often random variables, whereas for items only continuous parameters are used and they are commonly of the fixed type, although exceptions occur. It is shown in the present article that random item parameters make sense theoretically, and that in practice the random item approach is promising to handle several issues, such as the measurement of persons, the explanation of item difficulties, and trouble shooting with respect to DIF. In correspondence with these issues, three parts are included. All three rely on the Rasch model as the simplest model to study, and the same data set is used for all applications. First, it is shown that the Rasch model with fixed persons and random items is an interesting measurement model, both, in theory, and for its goodness of fit. Second, the linear logistic test model with an error term is introduced, so that the explanation of the item difficulties based on the item properties does not need to be perfect. Finally, two more models are presented: the random item profile model (RIP) and the random item mixture model (RIM). In the RIP, DIF is not considered a discrete phenomenon, and when a robust regression approach based on the RIP difficulties is applied, quite good DIF identification results are obtained. In the RIM, no prior anchor sets are defined, but instead a latent DIF class of items is used, so that posterior anchoring is realized (anchoring based on the item mixture). It is shown that both approaches are promising for the identification of DIF.
Problem solving has been recognized as a central skill that today’s students need to thrive and shape their world. As a result, the measurement of problem-solving competency has received much attention in education in recent years. A popular tool for the measurement of problem solving is simulated interactive tasks, which require students to uncover some of the information needed to solve the problem through interactions with a computer-simulated environment. A computer log file records a student’s problem-solving process in details, including his/her actions and the time stamps of these actions. It thus provides rich information for the measurement of students’ problem-solving competency. On the other hand, extracting useful information from log files is a challenging task, due to its complex data structure. In this paper, we show how log file process data can be viewed as a marked point process, based on which we propose a continuous-time dynamic choice model. The proposed model can serve as a measurement model for scaling students along the latent traits of problem-solving competency and action speed, based on data from one or multiple tasks. A real data example is given based on data from Program for International Student Assessment 2012.
This article presents a joint modeling framework of ordinal responses and response times (RTs) for the measurement of latent traits. We integrate cognitive theories of decision-making and confidence judgments with psychometric theories to model individual-level measurement processes. The model development starts with the sequential sampling framework which assumes that when an item is presented, a respondent accumulates noisy evidence over time to respond to the item. Several cognitive and psychometric theories are reviewed and integrated, leading us to three psychometric process models with different representations of the cognitive processes underlying the measurement. We provide simulation studies that examine parameter recovery and show the relationships between latent variables and data distributions. We further test the proposed models with empirical data measuring three traits related to motivation. The results show that all three models provide reasonably good descriptions of observed response proportions and RT distributions. Also, different traits favor different process models, which implies that psychological measurement processes may have heterogeneous structures across traits. Our process of model building and examination illustrates how cognitive theories can be incorporated into psychometric model development to shed light on the measurement process, which has had little attention in traditional psychometric models.
The paper addresses three neglected questions from IRT. In section 1, the properties of the “measurement” of ability or trait parameters and item difficulty parameters in the Rasch model are discussed. It is shown that the solution to this problem is rather complex and depends both on general assumptions about properties of the item response functions and on assumptions about the available item universe. Section 2 deals with the measurement of individual change or “modifiability” based on a Rasch test. A conditional likelihood approach is presented that yields (a) an ML estimator of modifiability for given item parameters, (b) allows one to test hypotheses about change by means of a Clopper-Pearson confidence interval for the modifiability parameter, or (c) to estimate modifiability jointly with the item parameters. Uniqueness results for all three methods are also presented. In section 3, the Mantel-Haenszel method for detecting DIF is discussed under a novel perspective: What is the most general framework within which the Mantel-Haenszel method correctly detects DIF of a studied item? The answer is that this is a 2PL model where, however, all discrimination parameters are known and the studied item has the same discrimination in both populations. Since these requirements would hardly be satisfied in practical applications, the case of constant discrimination parameters, that is, the Rasch model, is the only realistic framework. A simple Pearson x2 test for DIF of one studied item is proposed as an alternative to the Mantel-Haenszel test; moreover, this test is generalized to the case of two items simultaneously studied for DIF.
A procedure for ordering object (stimulus) pairs based on individual preference ratings is described. The basic assumption is that individual responses are consistent with a nonmetric multidimensional unfolding model. The method requires data where a numerical response is independently generated for each individual-object pair. In conjunction with a nonmetric multidimensional scaling procedure, it provides a vehicle for recovering meaningful object configurations.
A method is developed to investigate the additive structure of data that (a) may be measured at the nominal, ordinal or cardinal levels, (b) may be obtained from either a discrete or continuous source, (c) may have known degrees of imprecision, or (d) may be obtained in unbalanced designs. The method also permits experimental variables to be measured at the ordinal level. It is shown that the method is convergent, and includes several previously proposed methods as special cases. Both Monte Carlo and empirical evaluations indicate that the method is robust.
A new procedure is discussed which fits either the weighted or simple Euclidian model to data that may (a) be defined at either the nominal, ordinal, interval or ratio levels of measurement; (b) have missing observations; (c) be symmetric or asymmetric; (d) be conditional or unconditional; (e) be replicated or unreplicated; and (f) be continuous or discrete. Various special cases of the procedure include the most commonly used individual differences multidimensional scaling models, the familiar nonmetric multidimensional scaling model, and several other previously undiscussed variants.
The procedure optimizes the fit of the model directly to the data (not to scalar products determined from the data) by an alternating least squares procedure which is convergent, very quick, and relatively free from local minimum problems.
The procedure is evaluated via both Monte Carlo and empirical data. It is found to be robust in the face of measurement error, capable of recovering the true underlying configuration in the Monte Carlo situation, and capable of obtaining structures equivalent to those obtained by other less general procedures in the empirical situation.
Various areas in psychology are interested in whether specific processes underlying judgments and behavior operate in an automatic or nonautomatic fashion. In social psychology, valuable insights can be gained from evidence on whether and how judgments and behavior under suboptimal processing conditions differ from judgments and behavior under optimal processing conditions. In personality psychology, valuable insights can be gained from individual differences in behavioral tendencies under optimal and suboptimal processing conditions. The current chapter provides a method-focused overview of different features of automaticity (e.g., unintentionality, efficiency, uncontrollability, unconsciousness), how these features can be studied empirically, and pragmatic issues in research on automaticity. Expanding on this overview, the chapter describes the procedures of extant implicit measures and the value of implicit measures for studying automatic processes in judgments and behavior. The chapter concludes with a discussion of pragmatic issues in research using implicit measures.
In this chapter we review advanced psychometric methods for examining the validity of self-report measures of attitudes, beliefs, personality style, and other social psychological and personality constructs that rely on introspection. The methods include confirmatory-factor analysis to examine whether measurements can be interpreted as meaningful continua, and measurement invariance analysis to examine whether items are answered the same way in different groups of people. We illustrate the methods using a measure of individual differences in openness to political pluralism, which includes four conceptual facets. To understand how the facets relate to the overall dimension of openness to political pluralism, we compare a second-order factor model and a bifactor model. We also check to see whether the psychometric patterns of item responses are the same for males and females. These psychometric methods can both document the quality of obtained measurements and inform theorists about nuances of their constructs.
This chapter highlights the utility of cultural imagination, the ability to see human behaviors not just as the result of their dispositions or immediate situations but also as the result of larger cultural contexts. Our cultural imagination, as researchers, evolves as we are increasingly exposed to ideas from different parts of the world, either through collaboration with other researchers or interacting with individuals outside our immediate cultural context. While cross-cultural research has become simpler with the rise of the Internet, there still remain many challenges. This current chapter delineates concrete steps one can take to conduct an informative cross-cultural study, increasing the diversity of databases for generalizable theories of personality and social behaviors.
Since India had been controlled by the British, it regressed to a lower stage. Poverty had been a lived reality for Indians, including for some of the Indian economists, since the late seventeenth century. International trade networks were disrupted by economic crisis and wars. Meanwhile, the Indian subcontinent was experiencing some of the most severe famines in its history. The Indian economists felt these crises sweeping their cities and villages. In particular, Dadabhai Naoroji and Romesh Chunder Dutt would spend most of their adult lives examining the regress that they saw in India. They would explore how it could be measured, how it varied from region to region, and its causes.
This chapter reconstructs the ethical ambiguities and popular anxieties that emerged during a spectacular period of coffee smuggling in the 1970s, centered in Chepkube village near the border of Kenya and Uganda. The criminalized trade provided residents with newfound wealth and consumptive possibility; magendo, as it was known, also was a stark challenge to the Ugandan state’s ability to monopolize the valuation of its most important export. However, participants’ unease did not reflect the illegality of magendo. Rather, the excessive and rapid riches acquired through coffee smuggling challenged prevailing ideas of propriety, respectability, and morality. In other words, existing ideas about how proper value should be morally produced—through laborious effort and familial networks—were undermined by the sudden revaluation of coffee. Smuggling is a form of arbitrage, a style of economic action premised on the capitalization on disjunctures of jurisdiction, of measurement, and of appearance. Magendo participants actively worked to produce such differences in order to acquire wealth; yet arbitrage generated an ambiguous mix of desire and disdain. Based on oral histories and fieldwork on both sides of the border, this chapter reveals how the careful orchestration of social relations and material goods is at the heart of valuation, and it emphasizes how popular valuation practices change and conflict with state projects of governing value and defining citizenship.
The accumulation of empirical evidence that has been collected in multiple contexts, places, and times requires a more comprehensive understanding of empirical research than is typically required for interpreting the findings from individual studies. We advance a novel conceptual framework where causal mechanisms are central to characterizing social phenomena that transcend context, place, or time. We distinguish various concepts of external validity, all of which characterize the relationship between the effects produced by mechanisms in different settings. Approaches to evidence accumulation require careful consideration of cross-study features, including theoretical considerations that link constituent studies and measurement considerations about how phenomena are quantifed. Our main theoretical contribution is developing uniting principles that constitute the qualitative and quantitative assumptions that form the basis for a quantitative relationship between constituent studies. We then apply our framework to three approaches to studying general social phenomena: meta-analysis, replication, and extrapolation.
This is a reprinting of Wolfe’s response to the EPR paper. Wolfe insists upon an epistemic reading of the wavefunction, arguing that, under such an interpretation, the EPR paradox dissolves.
This is a reprinting of Ruark’s response to the EPR paper. Ruark puts the EPR debate down to disagreement over the criterion of reality. Ruark states that the majority of physicists will, pace EPR, consider this criterion satisfied even when the elements of a theory correspond only to indirectly measured features of reality.
This Element examines how climate scientists have arrived at answers to three key questions about climate change: How much is earth's climate warming? What is causing this warming? What will climate be like in the future? Resources from philosophy of science are employed to analyse the methods that climate scientists use to address these questions and the inferences that they make from the evidence collected. Along the way, the analysis contributes to broader philosophical discussions of data modelling and measurement, robustness analysis, explanation, and model evaluation.