To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
This textbook reflects the changing landscape of water management by combining the fields of satellite remote sensing and water management. Divided into three major sections, it begins by discussing the information that satellite remote sensing can provide about water, and then moves on to examine how it can address real-world management challenges, focusing on precipitation, surface water, irrigation management, reservoir monitoring, and water temperature tracking. The final part analyses governance and social issues that have recently been given more attention as the world reckons with social justice and equity aspects of engineering solutions. This book uses case studies from around the globe to demonstrate how satellite remote sensing can improve traditional water practices and includes end-of-chapter exercises to facilitate student learning. It is intended for advanced undergraduate and graduate students in water resource management, and as reference textbook for researchers and professionals.
Retrofitting aircraft cabins is characterized by a large number of documents created and required, most of which are currently processed manually. Engineers need to identify which documents include the information that is required for a specific task. This paper proposes an approach that builds upon a digital knowledge base and moves towards automatically processing the quantity-on-hand documents to reduce the work required to identify the required documents without the labour-intensive creation of the knowledge base in beforehand. After describing the scenario this work faces, comparable approaches and promising techniques are discussed. A process-chain that builds upon these fundamentals is presented, including a selection of feasible techniques and algorithms. Finally, the steps towards an implementation as part of the transformation towards a data-driven value chain are presented.
This work develops a method to integrate operational data into system models following MBSE principles. Empirical analysis reveals significant obstacles to data-driven development, including heterogeneous and non-transparent data structures, poor metadata documentation, insufficient data quality, lack of references, and limited data-driven mindset. A method based on the RFLP chain links operating data structures to logical-level elements. Data analyses are aligned with specific requirements or functional/physical elements, enabling systematic data-driven modeling. This method improves efficiency, fosters system knowledge development, and connects technical systems with operational data.
Gracia de Luna conducted experiments with an HMD virtual environment in which human subjects were presented with surprise distractions. His collected data for head, dominant hand, and non-dominant hand included 6 DOF human subject trajectories. This paper examines this data from 57 human subject responses to those surprise virtual environment distractions using statistical trajectory clustering algorithms. The data is organized and processed with a Dynamic Time Warping (DTW) algorithm and then analyzed using the Density Based Spatial Clustering (DBSCAN) algorithm. The K-means method was used to determine the appropriate number of clusters. Chi Squared goodness of fit was used to determine statistical significance. For five of the data sets, a p value of less than 0.05 was found. These five data sets were found to have a limited relationship to the measured variables.
No study has evaluated the relationship between heavy rain disasters and influenza by comparing victims and non-victims, and we investigated the association between the 2018 western Japan heavy rain disaster and influenza.
Methods
All patients registered in the National Health Insurance Claims Database and treated in the Hiroshima, Okayama, and Ehime prefectures were included in this retrospective cohort study conducted 1-year post-disaster. A multivariate mixed-effects logistic regression analysis was used to assess the association between the disaster and anti-influenza drug prescribing. A difference-in-differences analysis was conducted to assess anti-influenza drug use for the 4-month period immediately before and every 4 months for a year post-disaster.
Results
This study included 6 176 300 individuals (victims: 36 076 [0.60%]); 2573 (7.1%) and 458 157 (7.4%) in the victim and non-victim groups, respectively, used anti-influenza drugs in the year following the flood. The victims were significantly more likely than non-victims to use anti-influenza drug (risk ratio 1.18; 95% confidence interval [CI] 1.12-1.42). The victims used significantly more anti-influenza drugs in the 4 months immediately post-disaster compared with just before the disaster (odds ratio 3.62; 95% CI 1.77-7.41).
Conclusions
Anti-influenza drug use was higher among victims of the 2018 Western Japan heavy rain disaster than among non-victims.
Adults with mood and/or anxiety disorders have increased risks of comorbidities, chronic treatments and polypharmacy, increasing the risk of drug–drug interactions (DDIs) with antidepressants.
Aims
To use primary care records from the UK Biobank to assess DDIs with citalopram, the most widely prescribed antidepressant in UK primary care.
Method
We classified drugs with pharmacokinetic or pharmacodynamic DDIs with citalopram, then identified prescription windows for these drugs that overlapped with citalopram prescriptions in UK Biobank participants with primary care records. We tested for associations of DDI status (yes/no) with sociodemographic and clinical characteristics and with cytochrome 2C19 activity, using univariate tests, then fitted multivariable models for variables that reached Bonferroni-corrected significance.
Results
In UK Biobank primary care data, 25 508 participants received citalopram prescription(s), among which 11 941 (46.8%) had at least one DDI, with an average of 1.96 interacting drugs. The drugs most commonly involved were proton pump inhibitors (40% of co-prescription instances). Individuals with DDIs were more often female and older, had more severe and less treatment-responsive depression, and had higher rates of psychiatric and physical disorders. In the multivariable models, treatment resistance and markers of severity (e.g. history of suicidal and self-harm behaviours) were strongly associated with DDIs, as well as comorbidity with cardiovascular disorders. Cytochrome 2C19 activity was not associated with the occurrence of DDIs.
Conclusions
The high frequency of DDIs with citalopram in fragile groups confirms the need for careful consideration before prescribing and periodic re-evaluation.
The secrecy of intelligence institutions might give the impression that intelligence is an ethics-free zone, but this is not the case. In The Ethics of National Security Intelligence Institutions, Adam Henschke, Seumas Miller, Andrew Alexandra, Patrick Walsh, and Roger Bradbury examine the ways that liberal democracies have come to rely on intelligence institutions for effective decision-making and look at the best ways to limit these institutions’ power and constrain the abuses they have the potential to cause. In contrast, the value of Amy Zegart’s and Miah Hammond-Errey’s research, in their respective books, Spies, Lies, and Algorithms: The History and Future of American Intelligence and Big Data, Emerging Technologies and Intelligence: National Security Disrupted, is the access each of them provides to the thoughts and opinions of the intelligence practitioners working in these secretive institutions. What emerges is a consensus that the fundamental moral purpose of intelligence institutions should be truth telling. In other words, intelligence should be a rigorous epistemic activity that seeks to improve decision-makers’ understanding of a rapidly changing world. Moreover, a key ethical challenge for intelligence practitioners in liberal democracies is how to do their jobs effectively in a way that does not undermine public trust. Measures recommended include better oversight and accountability mechanisms, adoption of a ‘risk of transparency’ principle, and greater understanding of and respect for privacy rights.
This chapter addresses how one could quantify and explore the impact of geopolitics on global businesses. Computational geopolitics is an attempt to integrate quantitative methods and geopolitical analysis to understand and predict trends. The explosive growth of data, improvements in computational power, and access to cloud computing have led to a proliferation of computational methods in analyzing geopolitics and its impact on companies. The chapter explores some tools and techniques used in computational geopolitics, including events-based approaches to measuring geopolitical tensions, textual approaches, and empirical approaches. In addition, it provides examples of ways in which analysts can quantify the impact of geopolitics on trade and foreign direct investment. It also introduces experimental methods to assess the effectiveness of companies’ strategic responses to geopolitical tensions. Large language models (LLMs) can be used for sentiment analysis, spotting trends, scenario building, risk assessment, and strategic recommendations. While they methods offer advances in quantifying the impact of geopolitics on global businesses, analysts should also be cautious about data quality and availability as well as the complexity of the phenomenon and the geopolitics of AI. The chapter concludes by pointing the reader to some widely used data sources for computational geopolitics.
Nowadays, both researchers and clinicians alike have to deal with increasingly larger datasets, specifically also in the context of mental health data. Sophisticated tools for dataset visualization of information from various item-based instruments, such as questionnaire data or data from digital applications or clinical documentations, are still lacking, specifically for an integration at multiple levels and for use in both data organization and appropriate construction for its valid use in subsequent analyses.
Methods
Here, we introduce ItemComplex, a Python-based framework for ex-post visualization of large datasets. The method exploits the comprehensive recognition of instrument alignments and the identification of new content networks and graphs based on item similarities and shared versus differential conceptual bases within and across data and studies.
Results
The ItemComplex framework was evaluated using four existing large datasets from four different cohort studies and demonstrated successful data visualization across multi-item instruments within and across studies. ItemComplex enables researchers and clinicians to navigate through big datasets reliably, informatively, and quickly. Moreover, it facilitates the extraction of new insights into construct representations and concept identifications within the data.
Conclusions
The ItemComplex app is an efficient tool in the field of big data management and analysis addressing the growing complexity of modern datasets to harness the potential hidden within these extensive collections of information. It is also easily adjustable for individual datasets and user preferences, both in the research and clinical field.
A distinction between types of methods (understanding and explanation) that generate different kinds of evidence relevant to the psychiatric assessment is characterised. The distinction is animated with both non-clinical and clinical examples and exercises. Scepticism about the distinction is addressed, and three influential systems of psychiatric knowledge which collapse understanding and explanation in different ways are discussed. The argument is made that the distinction (analogous to the romantic/classic distinction) resurfaces and is compelling. However, another challenge becomes important – holism in psychiatric assessment – which the understanding/explanation distinction leaves in an unsatisfactory state.
While previous chapters discussed deep learning recommender systems from a theoretical and algorithmic perspective, this chapter shifts focus to the engineering platform that supports their implementation. Recommender systems are divided into two key components: data and model. The data aspect involves the engineering of the data pipeline, while the model aspect is split between offline training and online serving. This chapter is structured into three parts: (1) the data pipeline framework and big data platform technologies; (2) popular platforms for offline training of recommendation models like Spark MLlib, TensorFlow, and PyTorch; and (3) online deployment and serving of deep learning recommendation models. Additionally, the chapter covers the trade-offs between engineering execution and theoretical considerations, offering insights into how algorithm engineers can balance these aspects in practice.
Current approaches to identifying individuals at risk for psychosis capture only a small proportion of future psychotic disorders. Recent Finnish research suggests a substantial proportion of individuals at risk of psychosis attend child and adolescent mental health services (CAMHS) earlier in life, creating important opportunities for prediction and prevention. To what extent this is true outside Finland is unknown.
Aims
To establish the proportion of psychotic and bipolar disorder diagnoses that occurred in individuals who had attended CAMHS in Wales, UK, and whether, within CAMHS, certain factors were associated with increased psychosis risk.
Method
We examined healthcare contacts for individuals born between 1991 and 1998 (N = 348 226), followed to age 25–32. Using linked administrative healthcare records, we identified all psychotic and bipolar disorder diagnoses in the population, then determined the proportion of cases where the individual had attended CAMHS. Regression analyses examined associations between sociodemographic and clinical risk markers with psychotic and bipolar disorder outcomes.
Results
Among individuals diagnosed with a psychotic or bipolar disorder, 44.78% had attended CAMHS (hazard ratio = 6.28, 95% CI = 5.92–6.65). Low birth weight (odds ratio = 1.33, 95% CI = 1.15–1.53), out-of-home care experience (odds ratio = 2.05, 95% CI = 1.77–2.38), in-patient CAMHS admission (odds ratio = 1.49, 95% CI = 1.29–1.72) and attending CAMHS in childhood (in addition to adolescence; odds ratio = 1.16, 95% CI = 1.02–1.30) were all within-CAMHS risk markers for psychotic and bipolar disorders.
Conclusions
A substantial proportion (45%) of future psychotic and bipolar disorder cases emerge in individuals who had attended CAMHS, demonstrating large-scale opportunities for early intervention and prevention within CAMHS.
The integration of big data into criminal investigations is advancing significantly. Big data fundamentally involves the utilization of artificial intelligence technologies to analyse vast quantities of electronic information. The inherent features of big data contribute to minimizing subjectivity in investigative procedures and facilitate the evolution of criminal investigation methodologies and incident identification. However, challenges persist regarding the protection of rights and potential biases in data collection, as well as issues of subjectivity and the “black box effect” in data processing, alongside security concerns related to data storage. To address these challenges, it is essential to implement strategies such as enhancing the quality of big data, restricting the transparency of data processing methods and establishing a tiered protection framework for personal information.
The digital age, characterized by the rapid development and ubiquitous nature of data analytics and machine learning algorithms, has ushered in new opportunities and challenges for businesses. As the digital evolution continues to reshape commerce, it has empowered firms with unparalleled access to in-depth consumer data, thereby enhancing the implementation of a variety of personalization strategies. These strategies utilize sophisticated machine learning algorithms capable of attaining personal preferences, which can better tailor products and services to individual consumers. Among these personalization strategies, the practice of personalized pricing, which hinges on leveraging customer-specific data, is coming to the forefront.
The criteria for evaluating research studies often include large sample size. It is assumed that studies with large sample sizes are more meaningful than those that include a fewer number of participants. This chapter explores biases associated with the traditional application of null hypothesis testing. Statisticians now challenge the idea that retention of the null hypothesis signifies that a treatment is not effective. A finding associated with an exact probability value of p = 0.049 is not meaningfully different from one in which p = 0.051. Yet the interpretation of these two studies can be dramatically different, including the likelihood of publication. Large studies are not necessarily more accurate or less biased. In fact, biases in sampling strategy are amplified in studies with large sample sizes. These problems are of increasing concern in the era of big data and the analysis of electronic health records. Studies that are overpowered (because of very large sample sizes) are capable of identifying statistically significant differences that are of no clinical importance.
Despite enormous efforts at healthcare improvement, major challenges remain in achieving optimal outcomes, safety, cost, and value. This Element introduces the concept of learning health systems, which have been proposed as a possible solution. Though many different variants of the concept exist, they share a learning cycle of capturing data from practice, turning it into knowledge, and putting knowledge back into practice. How learning systems are implemented is highly variable. This Element emphasises that they are sociotechnical systems and offers a structured framework to consider their design and operation. It offers a critique of the learning health system approach, recognising that more has been said about the aspiration than perhaps has been delivered. This title is also available as open access on Cambridge Core.
Physical activities are widely implemented for non-pharmacological intervention to alleviate depressive symptoms. However, there is little evidence supporting their genotype-specific effectiveness in reducing the risk of self-harm in patients with depression.
Aims
To assess the associations between physical activity and self-harm behaviour and determine the recommended level of physical activity across the genotypes.
Method
We developed the bidirectional analytical model to investigate the genotype-specific effectiveness on UK Biobank. After the genetic stratification of the depression phenotype cohort using hierarchical clustering, multivariable logistic regression models and Cox proportional hazards models were built to investigate the associations between physical activity and the risk of self-harm behaviour.
Results
A total of 28 923 subjects with depression phenotypes were included in the study. In retrospective cohort analysis, the moderate and highly active groups were at lower risk of self-harm behaviour. In the followed prospective cohort analysis, light-intensity physical activity was associated with a lower risk of hospitalisations due to self-harm behaviour in one genetic cluster (adjusted hazard ratio, 0.28 [95% CI, 0.08–0.96]), which was distinguished by three genetic variants: rs1432639, rs4543289 and rs11209948. Compliance with the guideline-level moderate-to-vigorous physical activities was not significantly related to the risk of self-harm behaviour.
Conclusions
A genotype-specific dose of light-intensity physical activity reduces the risk of self-harm by around a fourth in depressive patients.
Analysts often seek to compare representations in high-dimensional space, e.g., embedding vectors of the same word across groups. We show that the distance measures calculated in such cases can exhibit considerable statistical bias, that stems from uncertainty in the estimation of the elements of those vectors. This problem applies to Euclidean distance, cosine similarity, and other similar measures. After illustrating the severity of this problem for text-as-data applications, we provide and validate a bias correction for the squared Euclidean distance. This same correction also substantially reduces bias in ordinary Euclidean distance and cosine similarity estimates, but corrections for these measures are not quite unbiased and are (non-intuitively) bimodal when distances are close to zero. The estimators require obtaining the variance of the latent positions. We (will) implement the estimator in free software, and we offer recommendations for related work.
Recent advances in natural language processing (NLP), particularly in language processing methods, have opened new avenues in semantic data analysis. A promising application of NLP is data harmonization in questionnaire-based cohort studies, where it can be used as an additional method, specifically when only different instruments are available for one construct as well as for the evaluation of potentially new construct-constellations. The present article therefore explores embedding models’ potential to detect opportunities for semantic harmonization.
Methods
Using models like SBERT and OpenAI’s ADA, we developed a prototype application (“Semantic Search Helper”) to facilitate the harmonization process of detecting semantically similar items within extensive health-related datasets. The approach’s feasibility and applicability were evaluated through a use case analysis involving data from four large cohort studies with heterogeneous data obtained with a different set of instruments for common constructs.
Results
With the prototype, we effectively identified potential harmonization pairs, which significantly reduced manual evaluation efforts. Expert ratings of semantic similarity candidates showed high agreement with model-generated pairs, confirming the validity of our approach.
Conclusions
This study demonstrates the potential of embeddings in matching semantic similarity as a promising add-on tool to assist harmonization processes of multiplex data sets and instruments but with similar content, within and across studies.
Exposure to maternal mental illness during foetal development may lead to altered development, resulting in permanent changes in offspring functioning.
Aims
To assess whether there is an association between prenatal maternal psychiatric disorders and offspring behavioural problems in early childhood, using linked health administrative data and the Australian Early Development Census from New South Wales, Australia.
Method
The sample included all mother–child pairs of children who commenced full-time school in 2009 in New South Wales, and met the inclusion criteria (N = 69 165). Univariable logistic regression analysis assessed unadjusted associations between categories of maternal prenatal psychiatric disorders with indicators of offspring behavioural problems. Multivariable logistic regression adjusted the associations of interest for psychiatric categories and a priori selected covariates. Sensitivity analyses included adjusting the final model for primary psychiatric diagnoses and assessing association of interest for effect modification by child's biological gender.
Results
Children exposed in the prenatal period to maternal psychiatric disorders had greater odds of being developmentally vulnerable in their first year of school. Children exposed to maternal anxiety disorders prenatally had the greatest odds for behavioural problems (adjusted odds ratio 1.98; 95% CI 1.43–2.69). A statistically significant interaction was found between child biological gender and prenatal hospital admissions for substance use disorders, for emotional subdomains, aggression and hyperactivity/inattention.
Conclusions
Children exposed to prenatal maternal mental illness had greater odds for behavioural problems, independent of postnatal exposure. Those exposed to prenatal maternal anxiety were at greatest risk, highlighting the need for targeted interventions for, and support of, families with mental illness.