Hostname: page-component-68c7f8b79f-pksg9 Total loading time: 0 Render date: 2026-01-15T11:53:20.769Z Has data issue: false hasContentIssue false

Transforming evidence synthesis: A systematic review of the evolution of automated meta-analysis in the age of AI

Published online by Cambridge University Press:  09 January 2026

Lingbo Li*
Affiliation:
School of Mathematical and Computational Sciences, Massey University , New Zealand
Anuradha Mathrani
Affiliation:
School of Mathematical and Computational Sciences, Massey University , New Zealand
Teo Susnjak
Affiliation:
School of Mathematical and Computational Sciences, Massey University , New Zealand
*
Corresponding author: Lingbo Li; Email: l.li5@massey.ac.nz
Rights & Permissions [Opens in a new window]

Abstract

Exponential growth in scientific literature has heightened the demand for efficient evidence-based synthesis, driving the rise of the field of automated meta-analysis (AMA) powered by natural language processing and machine learning. This PRISMA systematic review introduces a structured framework for assessing the current state of AMA, based on screening 13,216 papers (2006–2024) and analyzing 61 studies across diverse domains. Findings reveal a predominant focus on automating data processing (52.5%), such as extraction and statistical modeling, while only 16.4% address advanced synthesis stages. Just one study (approximately 2%) explored preliminary full-process automation, highlighting a critical gap that limits AMA’s capacity for comprehensive synthesis. Despite recent breakthroughs in large language models and advanced AI, their integration into statistical modeling and higher-order synthesis, such as heterogeneity assessment and bias evaluation, remains underdeveloped. This has constrained AMA’s potential for fully autonomous meta-analysis (MA). From our dataset spanning medical (67.2%) and non-medical (32.8%) applications, we found that AMA has exhibited distinct implementation patterns and varying degrees of effectiveness in actually improving efficiency, scalability, and reproducibility. While automation has enhanced specific meta-analytic tasks, achieving seamless, end-to-end automation remains an open challenge. As AI systems advance in reasoning and contextual understanding, addressing these gaps is now imperative. Future efforts must focus on bridging automation across all MA stages, refining interpretability, and ensuring methodological robustness to fully realize AMA’s potential for scalable, domain-agnostic synthesis.

Information

Type
Review
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of The Society for Research Synthesis Methodology

Highlights

What is already known?

  • Traditional meta-analysis (MA) is resource-intensive, struggles with scalability, and suffers from reproducibility issues, limiting its efficiency and reliability.

  • Automated MA (AMA) has advanced with machine learning and specialized tools, improving data extraction, statistical synthesis, and expanding beyond medical research.

  • Integration challenges remain, including workflow fragmentation, analytical limitations, and interoperability barriers, hindering full automation.

What is new?

  • This study presents a timely systematic and comparative analysis of AMA research progress and applications across medical and non-medical domains to reveal distinct patterns in implementation challenges and opportunities.

  • This study introduces a structured analytical framework to systematically evaluate the alignment between technological solutions and specific meta-analytical tasks, ensuring more effective automation implementation.

  • This study identifies gaps in current AMA capabilities and presents a roadmap for advancement, taking recent progress in AI, and specifically breakthroughs in large language models, into account.

Potential impact for RSM readers

  • For researchers, it maps AI-enhanced AMA tools, boosting efficiency in MA and inspiring cross-disciplinary innovation through AI’s transformative power.

  • For tool developers, it highlights gaps in AI-driven heterogeneity and bias assessment, urging advanced AI integration for scalable, interdisciplinary tools.

  • For policymakers, it emphasizes the vitality of AI-powered AMA for evidence-based decisions and the need for standardization and investment in comprehensive systems.

1 Introduction

Automation has become integral to modern life; it is driving efficiencies across industries and is now transforming knowledge-intensive domains such as academia.Reference Schwab 1 While businesses have long leveraged automation for operational gains,Reference Brynjolfsson and Mitchell 2 scholarly research is now accelerating the adoption of AI-driven tools to enhance both efficiency and scalability in evidence synthesis.Reference Marshall and Wallace 3 Systematic literature reviews (SLRs) are a cornerstone of academic research and are also among the most resource-intensive academic endeavors, whose workflows stand to be revolutionized by recent advancements in natural language processing (NLP, a field of AI focused on enabling computers to understand and generate human language),Reference Kwabena, Wiafe, John, Bernard and Boateng 4 machine learning (ML, algorithms that learn patterns from data to make predictions or decisions),Reference Tsafnat, Glasziou, Choong, Dunn, Galgani and Coiera 5 , Reference Nedelcu, Oerther and Engel 6 and large language models (LLMs, a type of ML model trained on massive text corpora to perform advanced language tasks).Reference Khraisha, Put, Kappenberg, Warraitch and Hadfield 7 These technologies are accelerating automation in literature curation, data extraction, and synthesis, and thereby addressing the growing challenge of processing vast and rapidly expanding volume of scientific outputs.Reference Van Dinter, Tekinerdogan and Catal 8

Meta-analyses (MAs) represent a key methodology within the context of SLRs for aggregating quantitative findingsReference Cooper 9 , Reference Deeks, Higgins and Altman 10 for which the current technological advancements present both opportunities and challenges for further automation.Reference Van Dinter, Tekinerdogan and Catal 8 Conducting MAs is resource-intensive, often spanning months or years. With the explosion in the number of papers being published in academic databases, researchers have estimated that the average time to complete and publish a systematic review requiring five researchers is 67 weeks, with an approximate cost of US$140,000.Reference Borah, Brown, Capers and Kaiser 11 Moreover, robust MA reviews tend to require an engagement with 3–5 domain experts to ensure its thoroughness, reliability, and accuracy.Reference Higgins, Thomas and Chandler 12 Such heavy demands on time, human resource, and financial investment pose barriers toward getting timely evidence-based synthesis, particularly in disciplines where rapid and accurate decision-making is essential. Consequently, automation has gained traction across various MA stages to mitigate these constraints. Studies have applied AI-driven techniques to enhance efficiency in literature screening,Reference Xiong, Liu and Tse 13 Reference Feng, Liang and Zhang 15 data extraction,Reference Michelson 16 Reference Schmidt, Hair and Graziosi 19 risk-of-bias assessment,Reference Cheng, Katz-Rogozhnikov, Varshney and Baldini 20 and heterogeneity reduction.Reference Rodriguez-Hernandez, Gorski, Tellez-Plaza, Schlosser and Wuttke 21 Despite these gains, automation efforts remain fragmented, with uneven progress across stages, particularly in those requiring complex reasoning and synthesis tasks.

While these advancements contributed toward progress in streamlining various stages of MAs in isolation, no comprehensive undertaking has been made recently to assess the current state of research on the automation of MAs and to situate the existing gaps within the significant and evolving breakthroughs in AI and LLMs, which are increasingly capable of performing complex reasoning.Reference Ofori-Boateng, Aceves-Martins, Wiratunga and Moreno-Garcia 22 The only dedicated reviewReference Christopoulou 23 synthesizing automated MA (AMA) focused narrowly on clinical trials and identified 38 approaches across 39 articles that applied ML techniques to various stages of MA. However, it concluded that automation remains “far from significantly supporting and facilitating the work of researchers.Reference Christopoulou 23 ” Clinical trials generally involve standardized procedures and well-defined outcomes that enable automation, while other domains, such as education and social sciences, exhibit more heterogeneous study designs and data formats, making automation more complex. While informative, this clinical trial-focused review has limited relevance for broader AMA research, as it overlooks recent methodological developments and predates advances in LLMs that could transform automation opportunities. Aside from this work, semi-AMA (SAMA) has also emerged as an interim solution, shortening MA timelines while maintaining rigor through expert.Reference Ajiji, Cottin and Picot 24 However, SAMA depends on human intervention in key steps, such as study selection and results interpretation, which limits its scalability. Given these shortcomings, a comprehensive review of AMA progress across domains is urgently needed to harness AI’s full potential and address persistent limitations in evidence synthesis automation.

Therefore, this study critically examines the current state of automation in MA research, identifying existing approaches and challenges in preparation for the next wave of AI-driven breakthroughs that are poised further transform the field. It addresses a critical gap by providing the first comprehensive and systematic synthesis of AMA applications across both medical and non-medical domains using a structured analytical framework. Through this analysis, we highlight key challenges and opportunities in AMA and offer insights into its evolving role in quantitative evidence synthesis. Our study therefore makes three meaningful contributions to AMA:

  • First, it presents a timely systematic and comparative analysis of AMA research progress and applications across medical and non-medical domains to reveal distinct patterns in implementation challenges and opportunities.

  • Second, it introduces a structured analytical framework to systematically evaluate the alignment between technological solutions and specific meta-analytical tasks, ensuring more effective automation implementation.

  • Third, it identifies critical gaps in current AMA capabilities—such as the need for deeper analytical integration and enhanced evidence synthesis—and presents a roadmap for advancement taking the recent AI, and specifically LLM breakthroughs into account.

2 Background

The following section explores the history and evolution of MA, tracing its development from a relatively nascent statistical technique to its current prominence in evidence synthesis across disciplines. The second section discusses our analytical framework that informs on technology adoption and task characteristics for conducting the AMA research process. These have helped lay out the research questions for this study.

2.1 History and evolution of meta-analysis

MA originated from the pioneering work by GlassReference Glass and Smith 25 in the late 1970s, who developed a statistical framework for synthesizing research findings across educational and psychological studies, formally coining the term “MA.”Reference Glass and Smith 25 The methodology expanded significantly into medicine and other scientific domains during the 1980s–1990s, particularly for analyzing randomized controlled trials (RCTs). This expansion was driven by the growing demand for evidence-based decision-making, enabling researchers to address contradictory results and overcome limitations of small sample sizes.Reference Egger, Smith and Altman 26 One landmark application in cardiovascular medicine evaluated statin use in reducing cholesterol levels by pooling data from numerous clinical trials to demonstrate clear benefits in lowering heart disease risks, which ultimately provided compelling evidence. 27 The field advanced further through more sophisticated statistical models and refined effect size estimation techniques,Reference Hedges and Olkin 28 enhancing the robustness of quantitative synthesis. The development of Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines,Reference Takkouche and Norman 29 Reference Page, McKenzie and Bossuyt 32 with its most recent 2020 update, established rigorous reporting standards that minimize bias and improve finding reliability. MA has become instrumental in healthcare research, and as the highest level of evidence synthesis,Reference Sackett 33 MA provides critical insights for clinical guidelines and public health policies.

However, the exponential growth in published research—exemplified by ScienceDirect (https://www.sciencedirect.com/) with 16 million papers from 2,500 journals serving 25 million monthly researchers—has challenged traditional MA approaches. The growing volume of literature has necessitated the development of automation tools to streamline and expedite the review process. Scholars anticipate that automated systematic reviews will revolutionize evidence-based medicine through real-time analysis capabilities and optimized workflows.Reference Tsafnat, Glasziou, Choong, Dunn, Galgani and Coiera 5 , Reference Van Dinter, Tekinerdogan and Catal 8 While various software packages (RevMan,Footnote 1 Comprehensive Meta-Analysis,Footnote 2 Stata,Footnote 3 and SPSSFootnote 4 ) support MA through features like effect size calculation and heterogeneity assessment, they are better characterized as “computer-assisted” rather than truly automated. For instance, RevMan, despite its user-friendly interface for MA, still requires substantial manual data extraction. Similarly, while Comprehensive Meta-Analysis offers advanced statistical modeling, and Stata and SPSS provide flexible analysis capabilities, they all demand significant user intervention and statistical expertise. In addition, the commercial nature and high licensing costs of them limit accessibility for researchers with limited funding.

Recent advancements in AI-driven techniques, including NLP, ML, and LLMs, have markedly improved the efficiency of MA. Automated processes now condense tasks—once requiring months and multiple authors—into days, leveraging enhanced computational methods. For instance, LLMs have demonstrated sensitivity approaching human performance in the initial screening of systematic reviews,Reference Matsui, Utsumi, Aoki, Maruki, Takeshima and Yoshikazu 34 illustrating their potential to streamline early AMA stages. Despite these improvements, full deployment of AMA remains in development, with current efforts limited in scope. Research has primarily focused on clinical trialsReference Christopoulou 23 and SAMA,Reference Ajiji, Cottin and Picot 24 where SAMA reducing timelines while relying on human oversight for rigor. This narrow emphasis restricts AMA’s broader deployment across diverse domains, highlighting an ongoing challenge in achieving comprehensive automation. Emerging AI “thinking models,”Reference Brown, Mann, Ryder, Larochelle, Ranzato, Hadsell, Balcan and Lin 35 Reference Huang and Chang 37 capable of performing complex reasoning, offer a promising avenue to bridge this gap. These models could automate sophisticated synthesis tasks, such as heterogeneity assessment and statistical integration, thereby enhancing AMA’s precision and scalability. For instance, their ability to adapt reasoning to varied research contexts holds potential for wider application, which points to a critical opportunity to advance evidence synthesis automation.Reference Ofori-Boateng, Aceves-Martins, Wiratunga and Moreno-Garcia 22

2.2 Analytical frameworks for technology evaluation

Having a robust analytical framework is crucial to synthesizing data drawn from pertinent studies and in drawing meaningful conclusions. The choice of framework for technology evaluation depends on the relevancy of data collected and the research questions that have been posed to analyse this data for understanding aspects related to technology’s usage and its overall performance. Two analytical frameworks that can provide us with a comprehensive assessment on the usage of various information systems (ISs)/information technology (IT) systems are unified theory of acceptance and use of technology (UTAUT), proposed by Venkatesh et al.Reference Venkatesh, Morris, Davis and Davis 38 , Reference Venkatesh, Thong and Xu 39 (for explaining user attitudes, their behavioral intentions, and overall acceptance of the technology in use) and task–technology fit (TTF), proposed by Goodhue and ThompsonReference Goodhue and Thompson 40 (for interpreting technology alignment with the proposed tasks that can lead to high-performance impacts). However, a key element missing in UTAUT is the disposition of the users, namely, users’ computer self-efficacy or their innovativeness in making best use of the technology; therefore, user expectancy (performance expectancy and effort expectancy) and contextual factors (facilitating conditions and social influences) have been proposed as extensions to UTAUT.Reference Dwivedi, Rana, Jeyaraj, Clement and Williams 41 TTF, on the other hand, posits that the effectiveness of technology adoption and its usage depends on how well the technology supports the specific needs of a given task. It emphasizes on characteristics of both the task and the technology to make a statement on task–technology fitness. If there is a good fit between task and technology, it increases the likelihood of technology utilization and leads to increased performance impact.

TTF explains technology adoption (e.g., data locatability, data quality, data accessibility, timeliness, technology reliability, and ease of use) by focusing on the actual usage of the technology which can in turn offer valuable insights on technology design and improvement strategies. We can delve deeper into the functional match between technology and tasks, which is particularly relevant for complex, multistage process like AMA, where task demands vary significantly at different stages. It lays a strong foundation to understand how technology characteristics would influence task behaviors and consequently the utilization of that technology for the given purpose, and finally to provide a measure of the performance impacts. These impacts could lead to further development of more tools and services already in the marketplace or lead to redesigning of tasks to take better advantage of the technology or to further embark on training programs to better engage users in using the technologies.Reference Goodhue and Thompson 40

In this study, TTF is applied at the task level, focusing on concrete AMA tasks at each phase rather than abstract task attributes.Reference D’Ambra, Wilson and Akter 42 , Reference Ali, Romero, Morrison, Hafeez and Ancker 43 By applying TTF to AMA, we can evaluate the suitability of automation tools across various stages of AMA and assess how well available technologies support the specific tasks and user needs at each stage.

  • Task demands and technology support across AMA stages: The distinct stages of AMA involve different types of tasks with varying demands and levels of complexity. For instance, in the data extraction phase, automation tools must handle unstructured data from diverse sources, ensuring accuracy in identifying relevant studies and variables. In the synthesis stages, the technology needs to support complex statistical computations while ensuring methodological rigor. In the reporting phase, automation tools must generate clear, interpretable results that comply with reporting standards. At each stage, TTF is applied to examine how effectively technologies meet these task requirements.

  • TTF across AMA stages: TTF emphasizes the alignment of technology with the specific tasks to be performed. In AMA, this approach ensures that automation technologies remain well-suited to the requirements of each task, leading to improved performance and greater user satisfaction.

The application of TTF in AMA allows for a richer understanding of how automation technologies can improve efficiency, accuracy, and user satisfaction. For instance, while UTAUT may evaluate whether researchers will intend to use automated tools in MA, TTF assesses whether these tools will improve accuracy and reduce time and labor requirements. Specifically, TTF in AMA helps enable evaluation across three critical dimensions: data quality assessment (reliability of automated extraction), system effectiveness (enhancement of the MA process), and user satisfaction (accessibility across expertise levels). This analytical framework has therefore been used in this study to share insights for optimizing AMA tools, advancing task–technology research, and improving user experience and performance outcomes.

2.3 Research questions

Having laid out the background of history and evolution of MA, this article applies TTF constructs to better inform on aspects related to AMA deployment, such as the current approaches in use, challenges being faced, future trends, and the overall impact on evidence synthesis. We provide a comprehensive review of the development of AMA over the past decade. Our review highlights how various tools are being applied across different disciplines and how they have developed over time to provide a comprehensive understanding of AMA in evidence synthesis. Accordingly, we have posed four research questions that will be addressed in this review. These are:

  • RQ1 (Descriptive): What are the current landscape and key characteristics of AMA approaches?

  • RQ2 (Analytical): How does our analytical framework illuminate the strengths and limitations of current AMA approaches within each information processing stage?

  • RQ3 (Comparative): What are the distinct patterns in AMA implementation, effectiveness, and challenges observed across medical and non-medical domains?

  • RQ4 (Future-oriented): What are the critical gaps and future directions for AMA development, and what obstacles need to be addressed to realize its full potential for evidence synthesis?

3 Methodology

This section details the methodologies that provide a prelude to the review process and for the presentation of our results. The review follows the PRISMA criteria in providing answers to the four research questions. Next, we introduce our information processing-centric model to evaluate the alignment between AMA tools and the specific tasks they are designed to support.

3.1 PRISMA process

Our investigation of automation in MA followed PRISMA guidelines, employing a systematic search strategy. The database search was restricted to “MA” terms to enhance precision. Broader terms, such as “systematic review,” were not included, as pilot testing indicated they mainly retrieved irrelevant records. Accordingly, the initial search employed the string (“meta-analysis” OR “meta analysis”) AND (automation OR automated OR “machine learning” OR “artificial intelligence” OR AI OR “natural language processing” OR “large language model” OR LLM), with title, abstract, and keyword field filters applied where available (e.g., TITLE-ABS-KEY in Scopus; [Title/Abstract] in PubMed), across PubMed, Scopus, Google Scholar, IEEE Xplore, and ACM Digital Library databases. This presented a preliminary overview of research activities within the stated field of interest, offering a broad yet comprehensive summary of the general characteristics of MA prevalent in existing literature. Next, we established clear inclusion criteria and practical constraints: (1) published from January 2014 to August 2024; (2) focus on explicitly MA-specific automation tools; (3) full-text availability with sufficient methodological detail (e.g., at least four pages with technical descriptions); and (4) shorter records (e.g., abstracts and posters) were excluded due to insufficient information and empirical or quantitative evaluations in automation techniques of MA. Table 1 details the eligibility criteria. Furthermore, to enhance coverage and overcome potential oversights from database-centric searches, we conducted bidirectional citation chaining “snowball” methods.Reference Goodman 44 , Reference Parker, Scott and Geddes 45 This involved both backward tracing (reviewing references cited in the retrieved studies) and forward tracing (identifying later works citing the selected papers) through Scopus and Google Scholar, expanding our temporal scope to 2006–2024. This approach not only expanded coverage by incorporating relevant gray literature and emerging frameworks but also established thematic linkages between foundational methodologies and their contemporary implementations.

Table 1 Criteria for study selection using PRISMA

The systematic review process, managed through Zotero 7, began with duplicate removal followed by a two-phase screening. One reviewer (L.L.) conducted the initial screening of titles and abstracts to exclude irrelevant records, while two additional reviewers (A.M. and T.S.) independently checked and confirmed the results. The same procedure was applied for the full-text review and any discrepancies in selection were resolved through consensus discussions among the reviewers. The PRISMA flowchart (shown in Figure 1) details the selection process. Data were analyzed and narratively summarized, with descriptive statistics presented in tables or graphs based on each study’s aim. This process identified 13,145 initial studies (2,200 PubMed, 6,204 Scopus, 4,574 Google Scholar, 132 IEEE Xplore, 35 ACM Digital Library, and 71 snowball), which were refined to 61 studies (see the Supplementary Material) after removing 3,363 duplicates and excluding 9,708 studies through screening. The whole visual illustration of our systematic review (shown in Figure 2) outlines the key steps, addresses four research questions, highlighting key contributions, current challenges, and future trends in AMA.

Figure 1 PRISMA workflow.

Note: “Too short” = records with fewer than four pages, excluded for lacking methodological detail.

Figure 2 Holistic framework for this review.

3.2 Progressive phase structure in TTF

AMA streamlines the traditional resource-intensive process of MA by integrating automation for data extraction, analysis, and synthesis, enhancing efficiency while reducing human error and statistical expertise requirements. MAs can be categorized along multiple dimensions, including network structure, data type, statistical framework, update approach, and purpose. Among these, this review focuses on widely used and methodologically distinct approaches: conventional MA (CMA) for direct pairwise comparisons and network MA (NMA) for integrating both direct and indirect evidence across multiple interventions.

While MA methodologies (e.g., the Cochrane Handbook) provide a comprehensive set of methodological stages, these frameworks were not originally designed with automation in mind. Existing automation tools target isolated stages of the process; however, there remains no overarching framework to systematically organize and align automation efforts across the full MA workflow. To address RQ1 (What are the current landscape and key characteristics of AMA approaches?), we introduce the progressive phase structure (PPS), an automation-oriented framework that complements existing methodologies by structurally organizing automation tasks across the entire process. Figure 3 illustrates the PPS framework, which categorizes automation processes into three distinct phases across both CMA and NMA:

  • Pre-processing stage: Encompasses problem definition, query design, and literature retrieval. NLP, ML, and LLMs can significantly reduce time spent on these labor-intensive tasks.

  • Processing stage: Involves information extraction (IE) and statistical modeling in CMA or network model construction and refinement in NMA. Automated tools leveraging NLP, ML, and LLMs help extract required datasets and other relevant information and achieve high efficiency.

  • Post-processing stage: Focuses on database establishment, diagnostics, and reporting in CMA, and robustness enhancement and visualization in NMA. Different automation tools can enhance reproducibility through standardized reports and dynamic visualizations, thereby improving transparency.

Figure 3 Progressive phase structure with TTF model.

Note: “PreQ”, “ProcQ”, and “PostQ” denote analytical questions from the pre-processing, processing, and post-processing stages, respectively. Assessment ratings (H = High, M = Moderate, L = Low) are defined above.

PPS does not aim to replace existing methodological taxonomies but to provide a complementary, automation-oriented structure. It organizes the workflow into three high-level phases that correspond to established MA tasks outlined in the Cochrane Handbook (see Table 2), ensuring methodological completeness while offering a streamlined and automation-compatible perspective on the review process. To assess automation effectiveness, we also integrate the TTF model with PPS, providing a structured approach to evaluating alignment between specific MA tasks and available automation tools. This approach systematically deconstructs the automation process into granular components and assesses technological fit at each stage, which operationalizes this alignment by defining:

  • Tasks: Fundamental, well-defined tasks performed at each PPS phase (e.g., problem definition, query design, and literature retrieval). These tasks represent the concrete units of analysis for evaluating technology alignment.

  • Technology characteristics: Functional capabilities (e.g., IE, document classification, and text generation) enabled by current automation tools, such as NLP models, ML algorithms, and LLMs.

  • TTF assessment: Structured evaluation questions designed to systematically assess the degree to which available automation tools support the defined tasks, using a three-level qualitative scale (high/moderate/low).

Table 2 Task-level mapping between PPS automation phases and Cochrane methodological stages

The three-level TTF assessment questions were developed through iterative pilot coding and team discussions to enhance conceptual clarity and ensure coverage across all AMA stages. For the actual application of these assessments, one reviewer (L.L.) initially extracted data from each included study (e.g., tool name, application stage, functionality, methodological approach, reported performance, and limitations). Two additional reviewers (A.M. and T.S.) independently checked and confirmed the extracted information. The extracted data were then mapped to the PPS framework. TTF ratings were assigned using the same process: L.L. provided initial judgments based on the predefined scale (H-high, M-moderate, and L-low), and A.M. and T.S. independently reviewed the ratings. Any discrepancies in data extraction or assessment were resolved through structured consensus discussions. By structuring AMA through PPS and rigorously applying the TTF model (all details were fully provided in Figure 3), this framework provides a robust methodological foundation for evaluating automation effectiveness in AMA.

4 Results

Our review identified AMA publications primarily from journals (72%), conferences (25%), and preprints (3%). Figure 4(a) illustrates the temporal trends showing growth from a single publication in 2006–2009 to seven in 2024. This acceleration, particularly from 2018 onward, coincides with broader AI advancements and increased availability of computational resources. Despite the growth, the relatively low publication volume indicates AMA remains an emerging field with substantial exploration potential. Besides, analysis of PPS implementation revealed that 89% of studies focused on automating a specific MA step, while only 11% addressed multiple stages. Notably, just one study (2%) attempted full integration across all MA stages, highlighting a significant methodological gap. This indicates that while isolated automation tools have advanced considerably, creating seamless multi-stage workflows remains challenging. Figure 4(b) shows that the processing stage dominates AMA research efforts. This concentration likely stems from the technical feasibility and maturity of NLP and ML tools for IE. As IE represents a fundamental prerequisite for all MAs, automation in this area yields substantial efficiency gains. In contrast, the later MA stage involves complex, context-dependent synthesis, which raises further automation challenges, limiting the broader adoption of the system throughout the process.

Figure 4 Temporal patterns in AMA publications (a) and proportional discrepancies across different stages (b).

To address RQ2 (How does our analytical framework illuminate the strengths and limitations of current AMA approaches within each information processing stage?), we provide a comprehensive task breakdown aligned with our analytical framework (refer Figure 3). Automation requirements and success rates vary significantly due to differences in data structure, synthesis models, and computational complexity. Our analysis examines automation strategies and tools employed in both CMA and NMA, identifying distinctive characteristics in each approach. The following sections detail automation processes across PPS stages within the TTF model.

4.1 Automation of data pre-processing

Pre-processing in MA comprises problem definition, query design, and literature retrieval, which are all critical for refining datasets for subsequent analysis. The quality of research questions and query design directly influences the relevance and comprehensiveness of retrieved literature. With increasing data volumes, automation has become essential for managing meta-analytic datasets while minimizing bias, as MA performance fundamentally depends on the retrieved literature. In this study, automation is defined for each stage of the meta-analytic workflow. In pre-processing, it refers to algorithmic or rule-based methods that minimize human involvement in problem formulation, query design, and literature retrieval. Our review shows that AMA studies on pre-processing have been developed and tested only within pairwise (CMA) frameworks. Although the procedures are conceptually identical for CMA and NMA, no studies have explicitly adapted automation to accommodate NMA-specific practical considerations, such as capturing all relevant interventions and comparators so that studies form a connected treatment network. Accordingly, current automation efforts in pre-processing remain limited to the CMA context. Table 3 presents a structured evaluation of studies focused on automating this phase, applying TTF to assess alignment between technologies and specific pre-processing tasks.

Table 3 Task–technology fit assessment for pre-processing in AMA

Note: PreQ = pre-processing question (1: alignment with domain-specific query requirements; 2: balance of comprehensiveness and precision; and 3: reproducibility of automated retrieval and screening).

Traditional MA requires researchers to craft search strategies that balance breadth and specificity, an inherently complex process that is dependent on researcher expertise.Reference Higgins, Thomas and Chandler 12 Automation tools have progressively reshaped this stage. Early systems, such as MetaBUS and MeSH-based expansion frameworks, standardized query formulation within domains, capturing up to 99% of eligible studies while reducing screening workloads by over 80%.Reference Bosco, Uggerslev and Steel 51 , Reference Deng, Yin and Bao 53 ML approaches, exemplified by K-means classification, achieved comparable precision to manual screening while drastically cutting dataset size.Reference Xiong, Liu and Tse 13 LLM-based methods further advanced retrieval efficiency and comprehensiveness, reporting sensitivities up to 95% and workload reductions of 40%–83%.Reference Wei, Zhang and Zhang 54 Reference Luo, Sastimoglu, Faisal and Deen 56 Despite these gains, reproducibility remains a limiting factor, whereas ML and LLM pipelines, though reproducible within fixed configurations, are sensitive to domain shifts and prompt variation. Overall, automation in pre-processing in CMA has evolved from static keyword expansion to semantically enriched retrieval, improving coverage and efficiency but still lacking in cross-domain generalizability and standardized reproducibility.

4.2 Automation of data processing

Our review highlights the critical role of automation in the data processing phase for both CMA and NMA methodologies. While both approaches aim to enhance meta-analytic efficiency and reliability, they involve distinct automation requirements. CMA prioritizes IE and statistical modeling for synthesizing individual study data, whereas NMA focuses on network model construction and refinement, addressing challenges in inconsistency detection and network connectivity assessment. The following sections provide an in-depth examination of these tasks and their automation potential.

4.2.1 Automated data processing in CMA

Following the pre-processing stage, the next critical task in CMA is IE and statistical modeling. IE techniques transform unstructured text into analyzable, structured data—a fundamental prerequisite for CMA. Key subtasks include named entity recognition (NER) for identifying critical variables and relation extraction (RE) for determining relationships between entities across research articles. Automated IE substantially reduces manual data entry, although performance often varies by domain complexity and terminology. For statistical modeling, automation increasingly relies on algorithmic and workflow-level standardization, enabling reproducible computation of effect sizes and model estimation across meta-analytic datasets. However, model selection must align with specific data types and research objectives: some statistical frameworks offer specialized capabilities for particular data structures (e.g., binary or time-to-event outcomes), while others provide broader applicability across diverse datasets and contexts. Table 4 summarizes studies on automating the data processing phase in CMA using the TTF model, assessing alignment between technologies and tasks requirements while providing insights into their effectiveness and limitations.

Table 4 Task–technology fit assessment for automated data processing in CMA

Note: ProcQ = processing question (1: extraction of complex patterns and contextual relationships; 2: enhancement of structure and interpretability of unstructured data; and 3: reliability of automated standardization and statistical consistency).

Table 5 Task–technology fit assessment for automated data processing in NMA

Note: ProcQ = processing question (1: extraction of complex patterns and contextual relationships; 2: enhancement of structure and interpretability of unstructured data; and 3: reliability of automated standardization and statistical consistency).

Information extraction: Across reviewed studies, automation in IE has evolved from rule-based NER toward deep learning and LLM-driven RE. Early NLP pipelines, such as EXACT,Reference Pradhan, Hoaglin, Cornell, Liu, Wang and Yu 61 achieved 100% data accuracy and a 60% reduction in extraction time, demonstrating high data precision and complete structural mapping. However, reliance on domain-specific vocabularies constrained generalization, limiting interpretability beyond predefined contexts. Hybrid and deep learning methods, such as BERT-based PICO extraction, improved both structural accuracy and semantic clarity, though they still required carefully curated training data to ensure consistent interpretation across studies.Reference Mutinda, Yada, Wakamiya and Aramaki 17 , Reference Mutinda, Liew, Yada, Wakamiya and Aramaki 65 More recent LLM-based frameworks, such as MetaMateReference Wang and Luo 70 and GPT-3.5-powered systems,Reference Kartchner, Ramalingam, Al-Hussaini, Kronick, Mitchell, Demner-fushman, Ananiadou and Cohen 67 , Reference Shah-Mohammadi and Large 68 further enhanced contextual linking and conceptual coherence. However, they show only moderate reliability due to sensitivity to prompts and inconsistent XML parsing. Overall, IE automation has advanced in structural precision and interpretability, but continues to face challenges in achieving stable, cross-domain reproducibility.

Statistical modeling: Automation in statistical modeling primarily enhances computational reproducibility and standardization. Early probabilistic frameworksReference Choi, Shen, Chinnaiyan and Ghosh 71 , Reference Marot, Foulley, Mayer and Jaffrézic 72 automated effect-size estimation and variance moderation, achieving high precision and structural completeness and strong interpretability in small-sample analyses. However, assumptions of study homogeneity and computational intensity limited robustness in large-scale applications. R-based automation tools like metafor,Reference Viechtbauer 73 METAL,Reference Willer, Li and Abecasis 74 and MetaOmics Reference Wang, Kang and Shen 75 established computational pipelines, incorporating parameter and varied heterogeneity estimators, resulting in high reproducibility. These systems ensured reliable model fitting across datasets but still required domain expertise to ensure valid inferences. Subsequent tools, including metamisc,Reference Debray, Damen and Riley 77 AMANIDA,Reference Llambrich, Correig, Gumà, Brezmes and Cumeras 80 and NeuroQuery,Reference Dockès, Poldrack and Primet 78 extended automation to predictive and cross-domain contexts. These frameworks integrated heterogeneous data types and automated statistical evaluations, achieving high data precision and semantic interpretability while maintaining reproducibility through transparent algorithms and standardized output. Overall, statistical modeling automation now provides stable and fully reproducible computational workflows, supported by R-based and algorithmic frameworks. Nonetheless, model selection, sensitivity analysis, and interpretation of complex heterogeneous data remain reliant on expert judgment.

4.2.2 Automated data processing in NMA

In NMA, constructing and refining network models represents a critical challenge distinct from CMA. While CMA primarily extracts information and develops statistical models, NMA must integrate both direct and indirect evidence across multiple interventions through complex network structures. Automation in this domain enhances model consistency, computational efficiency, and reduces manual intervention. Table 5 provides an overview of the included studies in NMA through the TTF model. Van Valkenhoef et al.Reference Van Valkenhoef, Lu, De Brock, Hillege, Ades and Welton 82 pioneered this progress by developing a Bayesian consistency model generation framework that transformed what was previously a manual process requiring subjective parameter decisions. Extending this work, they introduced an automated node-splitting procedure for systematic inconsistency detection, further strengthening analytical transparency and comparability across treatment networks, though reproducibility was moderately limited by residual ambiguity between heterogeneity and inconsistency.Reference Van Valkenhoef, Dias, Ades and Welton 83 More recently, Thom et al.Reference Thom, White, Welton and Lu 84 applied graph-theoretical metrics to automate network connectivity analysis, reducing manual workload and reinforcing confidence in indirect comparisons across complex networks. Collectively, these advances reducing researcher subjectivity and enabling reproducible, scalable evaluation of increasingly intricate treatment networks.

4.3 Automation of data post-processing

Having examined automation in data pre-processing and processing for both CMA and NMA, we now focus on data post-processing, a critical phase involving result refinement and synthesis to ensure reporting accuracy and clarity. This increasingly important area of AMA research enhances analytical precision and advances evidence synthesis. The following sections will provide a detailed characteristics of CMA and NMA post-processing automation.

4.3.1 Automated data post-processing in CMA

Our examination categorizes CMA automated post-processing into three domains: (1) database establishment for structured data organization; (2) diagnostics and extensions for bias or heterogeneity assessment; and (3) reporting synthesis and result interpretation for standardized findings presentation. In this context, automation refers to the use of computational systems (e.g., AI-assisted reporting frameworks and visualization tools) to perform data consolidation, interpretive synthesis, and result presentation with minimal human intervention. Table 6 presents the TTF alignment for CMA post-processing studies.

Table 6 Task–technology fit assessment for automated data post-processing in CMA

Note: PostQ = post-processing question (1: accuracy and completeness of database construction and bias correction; 2: reliability and transparency of AI-assisted interpretation; and 3: clarity and robustness of visualization for decision-making).

Database establishment: Automated database establishment in CMA has evolved from structured text-mining to dynamic web-integrated systems that enhance data accessibility and error correction. Early tools, such as Neurosynth,Reference Yarkoni, Poldrack, Nichols, Van Essen and Wager 85 achieved strong domain accuracy by constructing a MySQL-based database of 3,489 studies with 84% sensitivity and 97% specificity, but lexical coding inconsistencies limited interpretive precision and reproducibility. CancerMAReference Feichtinger, McFarlane and Larcombe 86 and CancerESTReference Feichtinger, McFarlane and Larcombe 87 integrating up to 80 curated datasets in oncology from multiple experiments and standardizing normalization pipelines, improving structural reliability but remaining constrained to specific platforms and outdated repositories. Subsequent frameworks, such as ShinyMDE,Reference Shashirekha and Wani 88 enhanced accessibility through R-based integration of GEO studies, while RetroBioCatReference Finnigan, Lubberink and Hepworth 96 further advanced reproducibility via real-time data updating and open-access deployment. Across these systems, automation demonstrated strong error control and data standardization but still required domain-specific validation and frequent manual curation to maintain interpretive consistency and update coverage.

Diagnostics and extensions: Automation in diagnostics and extensions has significantly improved bias detection, heterogeneity assessment, and interpretive transparency in CMA. Craig et al.Reference Craig, Bae and Taswell 89 employed web crawlers and NLP-based inference engines to extract effect sizes and detect cross-study inconsistencies, achieving high domain precision and interpretability, though reproducibility was limited by system complexity and lack of visual diagnostics. Cheng et al.Reference Cheng, Katz-Rogozhnikov, Varshney and Baldini 20 introduced causal inference diagnostics for bias detection in RCTs, enabling adjustment for hidden confounders but requiring manual causal graph interpretation. Recent frameworks demonstrate marked advances in reproducibility: metaGWASmanagerReference Rodriguez-Hernandez, Gorski, Tellez-Plaza, Schlosser and Wuttke 21 standardized over 130 genome-wide association study (GWAS) workflows, achieving high stability across analyses, while BiNDiscoverReference Bremer, Wohlgemuth and Fiehn 97 and MetaExplorerReference Kale, Lee, Goan, Tipton and Hullman 98 further integrated bias quantification and visualization for metabolomics data, reaching high consistency across 350 datasets. Overall, diagnostic automation now exhibits uniformly high domain precision and interpretive clarity, with reproducibility advancing from low to high through modular, transparent pipeline architectures.

Result synthesis and interpretation: Automated result synthesis and interpretation have progressed from conventional random-effects modeling to fully integrated AI- and LLM-driven systems. Michelson et al.Reference Michelson 16 and Yang et al.Reference Yang, Tang, Dongye and Chen 52 laid the groundwork with random-effects and R-Meta-based synthesis, achieving moderate to high interpretability but limited reproducibility due to scalability constraints. Subsequent frameworks, such as MetaCyto,Reference Hu, Jujjavarapu and Hughey 90 MetaMSD,Reference Ryu and Wendt 91 and RICOPILI,Reference Lam, Awasthi and Watson 92 improved integration accuracy and automation depth, reaching high reproducibility and interpretive precision across large datasets. In parallel, web-based platforms, such as CogTaleReference Sabates, Belleville and Castellani 94 and PsychOpen,Reference Burgard, Bosnjak and Studtrucker 95 introduced semi-automated reporting pipelines that produced standardized, dynamically updated outputs. Recent advancements, including Text2BrainReference Ngo, Nguyen, Chen and Sabuncu 93 Chat2Brain,Reference Wei, Zhang and Zhang 54 have integrated LLMs and visualization modules to enhance semantic synthesis. Overall, CMA post-processing remains dominated by conventional synthesis pipelines. LLM-based synthesis is still in its early exploratory stage, offering conceptual promise but limited practical adoption so far.

4.3.2 Automated data post-processing in NMA

In contrast to CMA, automated post-processing in NMA focuses on two primary areas: (1) robustness enhancement to ensure network model validity and stability and (2) graph-based visualization and reporting to facilitate complex treatment network interpretation. Table 7 presents studies relevant to NMA post-processing.

Table 7 Task–technology fit assessment for automated data post-processing in NMA

PostQ = post-processing question (1: accuracy and completeness of database construction and bias correction; 2: reliability and transparency of AI-assisted interpretation; 3: clarity and robustness of visualization for decision-making).

Robustness enhancement: Automation for robustness enhancement in NMA has evolved from traditional statistical comparisons toward integrated credibility frameworks. Neupane et al.Reference Neupane, Richer, Bonner, Kibret and Beyene 100 compared three major R packages—gemtc, pcnetmeta, and netmeta—evaluating their efficiency and flexibility across Bayesian and frequentist models. This comparison achieved high domain fit by clarifying methodological options but required statistical expertise for practical use. Nikolakopoulou et al.Reference Nikolakopoulou, Higgins and Papakonstantinou 102 developed CINeMA, a simulation-based framework providing quantitative credibility assessments for complex treatment networks, addressing uncertainty, bias, and heterogeneity. ROB-MENReference Chiocchia, Holloway and Salanti 103 automated the evaluation of missing data bias, further strengthening analysis reliability. Most recently, Reaon et al.Reference Reason, Benbow, Langham, Gimblett, Klijn and Malcolm 105 explored the integration of LLMs (GPT-4) with Bayesian modeling to achieve near end-to-end automation. However, as in CMA, LLM use in NMA remains preliminary, showing clear potential but lacking solid support at this point.

Graph-based visualization and reporting: Graph-based visualization is essential for interpreting complex treatment networks in NMA. Van Valkenhoef et al.Reference Van Valkenhoef, Tervonen, Zwinkels, De Brock and Hillege 99 pioneered this direction with ADDIS, a Bayesian graph-based platform that visualized evidence networks and treatment effects through standardized analysis and transparent output. Subsequently, MetaInsightReference Owen, Bradbury, Xin, Cooper and Sutton 101 combined R netmeta with a Shiny interface, enabling real-time network visualization, automatic inconsistency diagnostics, and simplified workflows for non-technical users, though limited to frequentist models. Extending accessibility to new fields, Liu et al.Reference Liu, Béliveau, Wei and Chen 104 developed BUGSnet, a Bayesian R package designed for psychology and social sciences. Nevertheless, current visualization and reporting systems remain dominated by rule-based and R-based architectures, AI- or LLM-assisted visualization has not yet been applied in NMA practice representing a promising direction for future development.

4.4 Patterns of AMA across domains

The PPS with TTF model provides a comprehensive framework for AMA by automating each process step. To answer RQ3 (What are the distinct patterns in AMA implementation, effectiveness, and challenges observed across medical and non-medical domains?), our analysis revealed significant domain-specific variations.

A primary distinction lies in the predominant data characteristics and their implications for automation adoption. Medical domains more frequently utilize standardized, structured data from clinical trials, healthcare records, and standardized literature (e.g., CONSORT-compliant reports and structured abstracts). This prevalence of standardization creates a stronger TTF for automated tools, such as NLP, ML, and LLMs, which can efficiently process consistent terminology with minimal human intervention. While medical research certainly includes unstructured elements (such as clinical notes and narrative case reports), the presence of substantial standardized components has enabled earlier and more widespread adoption of AMA. Conversely, non-medical fields (social sciences, management, education, and STEM) predominantly present heterogeneous, less structured data with varied reporting styles and terminologies. Although pockets of standardization exist (e.g., large-scale survey data in the social sciences or structured experimental datasets in STEM), the relative lack of universal standardization frameworks creates greater challenges for current automated tools. This helps explain the lower adoption rates observed in our review (Figure 5). Methodological traditions further differentiate these domains. In the medical domain, established protocols, statistical methods, and data collection guidelines support automation through more predictable, uniform data formats. The greater methodological diversity in non-medical fields, while valuable for exploratory research, complicates the application of automated tools.

Figure 5 Interdisciplinary applications of AMA across various domains.

Figure 6 Cross-domain mapping of AMA applications, data types, and methodological approaches.

Our systematic review provides empirical evidence for these domain-specific distinctions. Figure 5 illustrates the disparity: medical applications account for 67.2% of reviewed studies ( $n=41$ ), compared to 32.8% in non-medical domains ( $n=20$ ), with clinical trials alone comprising 21.3% of all AMA implementations. This marked imbalance highlights differences in the maturity of TTF across domains. To complement this proportional view, Figure 6 offers a quantitative mapping of cross-domain linkages among domain applications, data types, and computational methods, with line thickness reflecting application frequency. The visualization reveals a clear pattern: medical AMA primarily draws on structured or semi-structured data sources (e.g., public databases and abstracts) that align closely with NLP- and ML-based automation pipelines. In contrast, non-medical AMA remains centered on statistical and methodological tool development, with social science and STEM studies still at early stages of practical adoption. Taken together, these findings reveal a maturity gradient across domains: medical fields exhibit dense, stable connections between standardized data and well-established computational methods, whereas non-medical domains display more fragmented, exploratory linkages that are still evolving toward systematic automation. Notably, the emergence of LLMs marks a new trend across domains, but their practical applications remain limited in both medical and non-medical fields. Current studies are largely exploratory, focusing on assessing feasibility rather than achieving full automation. Detailed analysis in the following sections illustrates how these differences translate into distinct task–technology alignment dynamics, tool specialization trajectories, and varying levels of automation maturity across medical and non-medical subfields. Importantly, our dual-perspective analysis highlights reciprocal learning opportunities that bridge traditional disciplinary divides. These opportunities for cross-disciplinary collaboration emphasize that advancing AMA requires not only innovation within each domain but also purposeful knowledge exchange across the medical-non-medical divide.

4.4.1 Medical field

AMA in the medical field is rapidly evolving across clinical trials, molecular and cellular omics, neuroscience, and specialized domains.

Clinical trials: AMA in clinical trials has progressed from abstract-based data extraction to complex full-text analysis, propelled by NLP, ML, LLMs, and other enhanced computational tools. This development has enabled automation across critical tasks, including literature selection, data extraction, publication bias evaluation, and results synthesis. Early efforts established foundational synthesis capabilities,Reference Michelson 16 while subsequent research improved literature selectionReference Xiong, Liu and Tse 13 and significantly enhanced data extraction efficiency.Reference Mutinda, Yada, Wakamiya and Aramaki 17 , Reference Yang, Tang, Dongye and Chen 52 , Reference Pradhan, Hoaglin, Cornell, Liu, Wang and Yu 61 , Reference Mutinda, Liew, Yada, Wakamiya and Aramaki 65 , Reference Kartchner, Ramalingam, Al-Hussaini, Kronick, Mitchell, Demner-fushman, Ananiadou and Cohen 67 Reference Yun, Pogrebitskiy, Marshall, Wallace, Deshpande, Fiterau, Joshi, Lipton, Ranganath and Urteaga 69 Progress also extends to mitigating publication bias.Reference Cheng, Katz-Rogozhnikov, Varshney and Baldini 20 However, achieving fully AMA remains elusive due to persistent challenges, including data inconsistency, incomplete datasets, and limitations in processing complex full-text content. LLMs also require refinement to interpret intricate analytical demands effectively. Domain-specific applications illustrate both potential and constraints: ChatGPT has been adapted for screening radiology abstracts,Reference Issaiy, Ghanaati and Kolahi 55 and general LLMs have improved NMA for binary and time-to-event outcomes.Reference Reason, Benbow, Langham, Gimblett, Klijn and Malcolm 105 This trajectory highlights AMA’s promise while underscoring the need to address technical barriers for broader applicability. This trajectory illustrates that in structured and protocol-driven clinical environments, automation advances in depth and complexity, yet remains bounded by the limits of text understanding and data completeness.

Molecular and cellular omics: This subdomain exemplifies the structured-data and integration-oriented pattern of AMA, where automation can also build on large-scale, standardized repositories and statistical synthesis frameworks. Unlike literature-based clinical trials, omics AMA leverages structured datasets from public repositories like Genevestigator,Reference Zimmermann, Hirsch-Hoffmann, Hennig and Gruissem 106 GEO, and ArrayExpress,Reference Parkinson 107 emphasizing statistical analysis and integration. Key tasks include data processing, multi-omics integration, and differential expression analysis, a range of specialized tools supports these efforts: RankProd,Reference Hong, Breitling, McEntee, Wittner, Nemhauser and Chory 108 alongside frameworks by Boyko et al.Reference Boyko, Kaidina and Kim 57 and Devyatkin et al.,Reference Devyatkin, Molodchenkov and Lukin 60 facilitates gene expression dataset processing; METALReference Willer, Li and Abecasis 74 enables GWAS MA; ShinyMDEReference Shashirekha and Wani 88 aids in detecting differentially expressed genes; MetaGWASManagerReference Rodriguez-Hernandez, Gorski, Tellez-Plaza, Schlosser and Wuttke 21 handles large-scale GWAS data; MetaCytoReference Hu, Jujjavarapu and Hughey 90 analyzes high-dimensional cytometry data; and AmanidaReference Llambrich, Correig, Gumà, Brezmes and Cumeras 80 detects study discrepancies. These tools have enhanced data processing and integration, with MetaCyto notably improving efficiency in high-dimensional analysis,Reference Hu, Jujjavarapu and Hughey 90 yet challenges like dataset heterogeneity, platform variability, and incomplete data persist. Domain-specific adaptations address some of these issues, such as Amanida’s focus on metabolomics data gapsReference Llambrich, Correig, Gumà, Brezmes and Cumeras 80 and MetaGWASManager’s automation of GWAS analysis,Reference Rodriguez-Hernandez, Gorski, Tellez-Plaza, Schlosser and Wuttke 21 but broader application remains limited by these constraints. Overall, the trajectory in molecular and cellular omics shows that AMA is still constrained by inter-platform heterogeneity.

Neuroscience: The coexistence of quantitative brain-imaging outputs and narrative research reports illustrates the challenge of aligning heterogeneous data modalities within a single automation framework. Neuroscience AMA synthesizes brain-related data using NLP, ML, LLMs, and predictive modeling to identify patterns in cognitive and neural states. This approach targets tasks, such as brain activation mapping, cognitive intervention analysis, and event-related potential (ERP) analysis. Progress in brain mapping tools has evolved from NeuroSynthReference Yarkoni, Poldrack, Nichols, Van Essen and Wager 85 to NeuroQuery,Reference Dockès, Poldrack and Primet 78 Text2Brain,Reference Ngo, Nguyen, Chen and Sabuncu 93 and Chat2Brain,Reference Wei, Zhang and Zhang 54 while predictive modeling has advanced through NPDS 0.9Reference Craig, Bae and Taswell 89 and cognitive intervention repository via CogTale.Reference Sabates, Belleville and Castellani 94 Despite these developments, challenges persist, including limited data availability, variability in experimental design, and difficulties in processing large volumes of unstructured text. Domain-specific adaptations, such as probabilistic ERPs literature analysisReference Donoghue and Voytek 64 and neural networks linking text queries to brain activation in Chat2Brain,Reference Wei, Zhang and Zhang 54 demonstrate potential but highlight the need to address these constraints for broader application. Therefore, this mixed-structure in AMA demonstrates that the coexistence of structured and unstructured data both drives methodological innovation and constrains full automation.

Specialized domains: AMA applications extend to specialized domains, including traditional Chinese medicine (TCM), epidemiology, and biochemistry, demonstrating adaptability across diverse research contexts. In these fields, AMA focuses on data processing and evidence synthesis, employing tools, such as logistic regression, ML, and NLP. Notable successes include automated logistic regression for epidemiological individual participant data (IPD) MAs, which reduces processing time and errors.Reference Lorenz, Abdi and Scheckenbach 59 However, the specialized nature of these systems restricts their generalizability. Domain-specific adaptations, such as TCM literature synthesis for splenogastric diseasesReference Zhang, Wang and Yao 66 and the RetroBioCat Database for biocatalysis data exploration,Reference Finnigan, Lubberink and Hepworth 96 reveal a pattern of constrained generalization. Automation performs exceptionally well in targeted, well-defined contexts but faces challenges when extending beyond these specialized frameworks.

4.4.2 Non-medical field

AMA applications remain limited outside medicine, with only nascent adoption in three key domains: statistical tools, social sciences, and STEM. This scarcity reflects both challenges and opportunities for expanding AMA beyond medical contexts.

Statistical tools: This category illustrates a cross-domain methodological pattern of AMA, where automation focuses on enhancing statistical modeling, consistency checking, and computational reproducibility. Although inherently applicable across both medical and non-medical contexts, these tools are typically introduced in the literature as methodological contributions rather than domain-specific applications. For this reason, we present them here at the beginning of the non-medical section. These tools encompass Bayesian random-effects models, graph theory, web-based platforms, and decision rules. Statistical packages, such as metafor,Reference Viechtbauer 73 Meta-Essentials,Reference Suurmond, Van Rhee and Hak 76 and metamisc,Reference Debray, Damen and Riley 77 have improved analysis accessibility, while NMA tools, including gemtc, pcnetmeta, and netmeta,Reference Neupane, Richer, Bonner, Kibret and Beyene 100 support complex modeling. Semi-automated systems ADDISReference Van Valkenhoef, Tervonen, Zwinkels, De Brock and Hillege 99 and analytical frameworks for consistency checks and bias assessment such as Bayesian random-effects models,Reference Van Valkenhoef, Lu, De Brock, Hillege, Ades and Welton 82 Reference Thom, White, Welton and Lu 84 CINeMA,Reference Nikolakopoulou, Higgins and Papakonstantinou 102 and ROB-MENReference Chiocchia, Holloway and Salanti 103 further refine precision. Web platforms MetaInsightReference Owen, Bradbury, Xin, Cooper and Sutton 101 enhance usability for researchers without extensive statistical expertise. Nonetheless, challenges remain, including limited multi-modal data processing and the growing complexity of modern meta-analytical frameworks. Domain-specific adaptations, such as ADDIS, CINeMA, ROB-MEN, and MetaInsight, address specialized NMA needs but reflect the persistent tension between tool sophistication and broad applicability. Therefore, these developments show a methodological pattern emphasizing statistical rigor and reproducibility.

Social science: Social sciences have begun adopting AMA tools for synthesizing diverse data types across disciplines, such as human resource management, psychology, and education, focusing on tasks like data synthesis and predictive modeling. Tools, such as Bayesian methods and LLMs, underpin these efforts. Notable advances include MetaBUS, which streamlines MA across extensive literature volumesReference Bosco, Uggerslev and Steel 51 ; Bayesian NMA opens new possibilities for quantitative analysisReference Liu, Béliveau, Wei and Chen 104 ; and MetaMate leverages few-shot prompting for data extraction in education.Reference Wang and Luo 70 However, the diversity of data types, particularly qualitative data and complex models, poses significant challenges to automation, highlight the ongoing difficulty of achieving broad applicability across heterogeneous datasets. AMA in the social sciences remains in an exploratory stage, these early attempts mark an important foundation for future integration of LLM-based synthesis, suggesting a gradual but steady shift toward more systematic automation in social research.

STEM: AMA in STEM shows progress in literature retrieval and data extraction, leveraging ML-based tools and deep transfer learning. Tools like MetaSeer.STEMReference Neppalli, Caragea, Mayes, Nimon and Oswald 58 streamlines data extraction from research articles, enhancing literature analysis efficiency, while deep transfer learning systems improve retrieval processes.Reference Anisienia, Mueller, Kupfer, Staake and Bui 62 AMA adoption in STEM remains in its early stages, with automation primarily targeting specific tasks like information retrieval rather than comprehensive evidence synthesis. The lack of consistent data standards and the wide-ranging diversity of STEM research hinder scalability. However, the presence of structured experimental data makes STEM a promising area for future advancements as methodological and integration frameworks evolve.

5 Challenges and future potential for AMA

Despite increasing adoption of AMA techniques, significant challenges remain that must be addressed to realize its full potential for evidence synthesis. To answer RQ4 (What are the critical gaps and future directions for AMA development, and what obstacles need to be addressed to realize its full potential for evidence synthesis?), this section examines key barriers and future directions to enhance AMA’s credibility and utility. These challenges span multiple dimensions: enhancing analytical capabilities while mitigating automation biases; maintaining methodological rigor and transparency; adapting to evolving research technology developments; gaining broader acceptance among stakeholders; and ensuring reliability of synthesized evidence. Prior to presenting our proposed roadmap for AMA development, we assigned values to the “Difficulty” and “Priority” based on a structured methodological framework, grounded in a qualitative assessment of technical, methodological, organizational, ethical, and data-related factors, as well as their anticipated impact on AMA’s development. “Difficulty” reflects the complexity of implementation, considering factors, such as technical barriers (e.g., algorithm complexity and data availability), methodological challenges (e.g., validation rigor), organizational constraints (e.g., interdisciplinary collaboration), and ethical considerations (e.g., transparency). Ratings range from “Low” to “High,” with “Medium” indicating moderate challenges requiring moderate effort or expertise. “Priority” evaluates the urgency and impact of each research direction, integrating immediate practical needs (e.g., addressing current gaps), high-impact potential (e.g., improving validity or scalability), and long-term benefits (e.g., credibility and broad applicability). Ratings are categorized as “Immediate,” “Medium,” or “Long-term,” often combined with qualitative descriptors (e.g., “High Impact and” “Trust”) to reflect multifaceted outcomes. This approach ensures a balanced, evidence-based assessment, and Table 8 presents a prioritized roadmap for AMA development according to this assessment framework. This analysis aims to reveal that advancing AMA requires not only technical innovation, but also methodological refinement and strategic implementation approaches to improve its credibility and utility in diverse research contexts.

Table 8 Future research directions for AMA

5.1 Advancing analytical depth and balancing efficiency in AMA

A critical and persistent limitation in AMA remains the automation of advanced analytical methodologies, including sensitivity analyses, heterogeneity assessments, publication bias evaluations, and stratified subgroup analyses. While preliminary data processing has advanced significantly, sophisticated analytical automation remains underexplored, compromising the reproducibility and scientific validity of AMA findings. Future research should prioritize three critical areas: (1) Algorithm advancement. Developing frameworks that execute complex analytical functions with minimal human intervention while maintaining methodological rigor, including automated sensitivity analysis and bias detection tools. (2) Methodological balance. Creating frameworks that enhance efficiency without compromising analytical depth and integrity, with strategic human oversight at critical analytical stages. (3) Multi-modal data integration. Incorporating heterogeneous data types (numerical data, medical images, tables, and raw data) through adaptable extraction techniques for comprehensive, statistically sound evidence synthesis. These advancements would elevate AMA beyond basic automation to deliver both sophisticated analytical capabilities and enhanced efficiency, strengthening its credibility in high-impact research domains.

5.2 LLMs with fine-tuning and complex reasoning in AMA

LLMs, including those with advanced “thinking” capabilities capable of complex reasoning, offer transformative potential for AMA by efficiently processing unstructured text and extracting critical variables (effect sizes and confidence intervals) from research articles. These “thinking models” in LLMs extend beyond basic data extraction, enabling genuine knowledge synthesis and methodological reasoning, potentially revolutionizing evidence synthesis by integrating statistical and semantic understanding. They function as methodological thought partners, assessing heterogeneity between studies, and adapting analytical strategies to the specific characteristics of included studies, thereby enhancing AMA’s precision, scalability, and adaptability across diverse research contents. However, several challenges hinder their full-scale deployment of LLMs in AMA. These include hallucinations that fabricate results, which is unacceptable in high-stakes applications like healthcare; propagation of implicit biases from training corpora into synthesized outputs; and limitations with extensive context windows when processing journal articles, dissertations, and complex figures/tables.Reference Liu, Lin and Hewitt 109 To maximize this potential opportunity, future research should prioritize developing specialized “thinking LLMs” for analyzing long-form academic content with multi-modal capabilities; enhancing transparency through explainable AI (XAI) techniques to facilitate expert validation of automated extractions; and designing benchmarks and protocols to ensure the accuracy, reliability, and reproducibility of LLM-generated results. These advancements will significantly enhance the reliability and interpretability of LLM-assisted AMA workflows, potentially reshaping the foundations of evidence synthesis methodology and positioning “thinking LLMs” as a cornerstone of future AMA innovation.

5.3 Living AMA

Current AMA implementations primarily automate discrete stages of MA but lack mechanisms for continuous, real-time evidence updates. This limitation is particularly evident in Cochrane MAs, which require periodic updates to maintain clinical relevance. A “living AMA” addresses this gap by envisioning a system that can automatically and continuously scan databases for new studies, extract relevant data, and integrate fresh evidence into existing analyses. Realizing this vision should focus on three key aspects. First, designing robust AI-driven mechanisms to identify and validate new studies as they emerge. Second, developing algorithms to make a version control and reconcile conflicting data across studies while preserving analytical transparency. Third, creating efficient alert mechanisms that update researchers without overwhelming them with excessive information. Living AMA approaches have already emerged in related domains, such as “living literature review,”Reference Wijkstra, Lek, Kuhn, Welbers and Steijaert 110 COVID-19 living MAs,Reference Boutron, Chaimani and Devane 111 MetaCOVID project,Reference Evrenoglou, Boutron, Seitidis, Ghosn and Chaimani 112 and SOLES system.Reference Hair, Wilson, Wong, Tsang, Macleod and Bannach-Brown 113 Building on these foundations, future work must refine the methodological framework for Living AMA to ensure delivery of up-to-date, high-quality evidence synthesis.

5.4 Fostering multidisciplinary collaborations

The success of AMA depends on requiring seamless collaborations between statisticians, computer scientists, domain experts, and policymakers. However, interdisciplinary cooperation remains a bottleneck due to differences in methodologies, terminology, and research priorities. Addressing this challenge requires three strategic approaches: (1) interdisciplinary training programs to familiarize researchers with AMA methodologies and computational techniques; (2) joint funding initiatives to support large-scale, collaborative AMA projects; and (3) shared platforms and community to promote cross-disciplinary integration. These approaches leverage complementary expertise: statisticians ensure methodological rigor, computer scientists develop the technical framework, and domain experts provide contextual knowledge to interpret findings meaningfully. Through effective communication and trust-building, AMA can evolve into a widely adopted tool bridging computational power with domain-specific expertise.

5.5 Interpretability and transparency in AMA

As AMA tools become more sophisticated, transparency in their decision-making processes becomes increasingly paramount, particularly in high-stakes domains such as medical research where evidence synthesis directly influences clinical decisions. The integration of XAI methods into AMA represents a critical frontier in ensuring credibility and adoption. The challenge of interpretability in AMA extends beyond mere technical performance. While automated systems can significantly reduce the time and effort required for MA, their value diminishes if end-users cannot understand or trust their outputs. This is particularly crucial during the evidence synthesis phase, where complex algorithms process and integrate diverse evidence sources. Recent researchReference Luo, Sastimoglu, Faisal and Deen 56 highlights the delicate balance required between efficient automation and maintaining the depth and accuracy of evidence synthesis. Future research should prioritize the standardization of XAI integration within AMA workflows, ensuring automated processes remain transparent, reproducible, and trustworthy. Various XAI techniques, such as rule-based explanations, visual explanations, and sensitivity analysis, may integrate into AMA findings with more accessible and easier adjustments. Through these approaches, AMA can evolve into robust and widely accepted tools that enhance the quality of evidence synthesis.

6 Discussion

AMA has emerged as a transformative innovation in quantitative evidence synthesis, driven by exponential growth in literature that demands efficient, scalable, and reproducible quantitative research methods. Advanced AI, particularly “thinking models” with the capable of complex reasoning, has become a cornerstone of this evolution. This review has provided an evaluation of AMA via a descriptive lens (RQ1), analytical lens (RQ2), comparative lens (RQ3), and a future-oriented lens (RQ4). Despite AMA offering significant benefits compared to traditional MA, full automation remains aspirational rather than becoming a standard. This gap underscores the urgent imperative to harness “thinking models,” bridging technical and methodological barriers to position AMA as a critical frontier for future evidence synthesis innovation.

6.1 Methodological disparities between CMA and NMA

Quantitative analysis reveals a clear research imbalance, with 81% of AMA studies focusing on CMA versus only 19% addressing NMA. This disparity arises from the inherent complexity of NMA, which requires integrating both direct and indirect comparisons across multiple interventions while accounting for data heterogeneity. CMA, involving primarily pairwise comparisons, is more amenable to automation through established statistical frameworks and increasingly supported by emerging technologies, such as NLP and LLMs. Table 9 highlights key distinctions between CMA and NMA automation. While both CMA and NMA can employ fixed-effects, random-effects, or Bayesian models, CMA analyses are typically simpler in structure and supported by packages, such as metaMA Reference Marot, Foulley, Mayer and Jaffrézic 72 and metafor,Reference Viechtbauer 73 enabling more streamlined automation. In contrast, NMA often involves additional modeling layers to account for indirect comparisons and network consistency, requiring more specialized tools. Visualization in CMA often centers around forest and funnel plots. NMA, while also employing these, additionally requires complex network graphs, inconsistency plots, rankograms, and SUCRA plots, many of which demand manual adjustments or customization. Addressing the NMA automation gap necessitates algorithms capable of handling multi-dimensional, network-structured data while ensuring model transparency and consistency.

Table 9 Comparative analysis of CMA and NMA automation

6.2 Complexity of full automation

Despite notable advancements, full AMA workflow remains elusive. Our review reveals that the majority of existing studies employed semi-automated approaches, with automation largely confined to data extraction and preliminary synthesis. Only one studyReference Yang, Tang, Dongye and Chen 52 has explored full automation across all AMA stages. This highlights a critical gap between the automation of individual components, highlighting the gap between automating individual components and developing integrated, end-to-end systems. Key barriers include: (1) Technical challenges. Data heterogeneity across formats (structured databases, unstructured literature, and high-dimensional biomedical data); computational complexity of advanced meta-analytic methods; and LLM limitations in interpreting context-sensitive statistical details. (2) Methodological barriers. Difficulty automating qualitative judgments (risk-of-bias assessment, confounding adjustment, and evidence grading). (3) Organizational and infrastructural hinders. Limited cross-disciplinary adaptability and absence of universal standards for seamless data integration. Addressing these challenges calls for advancements in AI, carefully designed methodological frameworks, and a clearer, more detailed grasp of how automation demands differ across the AMA workflow. These demands are not uniform but shift significantly across stages. For instance, in Stage 1 (literature retrieval) or Stage 2 (IE), recall is typically prioritized over precision, as omitting relevant studies or variables undermines the comprehensiveness and validity of MA. In contrast, later stages, such as statistical calculation, synthesis, and interpretation, place greater emphasis on accuracy, transparency, and methodological consistency, where overly inclusive results may introduce ambiguity, bias, and analytical distortions. These tradeoffs highlight that what constitutes “optimal” automation is inherently stage-specific and context-dependent. Therefore, we believe that the development of stage-aware benchmarks is an important direction for future research. Importantly, the goal of AMA is not to achieve perfect end-to-end automation but to pursue context-sensitive optimization that enhances analytical utility while preserving scientific integrity.

6.3 Ethical and practical considerations

AMA adoption also raises critical ethical questions. One of the most pressing ethical concerns is the risk of bias amplification. AMA systems typically rely on existing datasets, including published literature, trained on existing literature may reinforce systematic publication biases, particularly if they rely on biased data sources. Additionally, the increasing reliance on automation in evidence synthesis introduces concerns about deskilling of researchers, which means as automation takes over certain tasks, researchers may become less proficient in critical appraisal and statistical analysis, potentially reducing the quality of evidence synthesis. Furthermore, the development and adoption of AMA tools are disproportionate, creating risks of global inequity in access to advanced evidence synthesis technologies. If AMA tools remain proprietary, cost-prohibitive, or require specialized technical expertise, low-resource settings may struggle to leverage these innovations, potentially widening disparities in research capacity. Another ethical use of AMA is transparency, without transparency, stakeholder trust in automated evidence synthesis may be undermined, raising concerns about reproducibility and accountability in decision-making. Therefore, researchers looking to adopt AMA should consider that: (1) not all AMA tools are equally effective across disciplines. Choosing the right tool requires an understanding of its strengths, limitations, and adaptability. (2) Rather than seeking full automation, researchers should integrate AMA as an assistive tool while maintaining expert oversight in critical analytical processes. Through these approaches, researchers can balance ethical responsibility, methodological rigor, and transparency without overshadow AMA potential benefits.

6.4 Implications for evidence synthesis

The ability of AMA to streamline quantitative evidence synthesis has been widely recognized across biomedicine, neuroscience, epidemiology, and omics research through automating data extraction, statistical modeling, and synthesis processes. Its evolution could significantly enhance efficiency, reproducibility, and scalability. For example, the continued advancement of AMA has the potential to reshape the landscape of evidence synthesis, which enabling more dynamic and responsive updates to existing evidence bases. This could be particularly valuable in rapidly evolving research domains, such as pandemic response or emerging medical technologies. Besides, automation approaches can facilitate data extraction and statistical analysis, thereby minimizing inconsistencies introduced by subjective human interpretation. Large-scale and complex analysis AMA from heterogeneous datasets will extend beyond traditional systematic reviews based solely on published clinical trials.

However, over-reliance on automation without addressing limitations risks undermining synthesized evidence reliability. One of the concerns is the diminished role of expert judgment in study selection, data interpretation, and result contextualization. Besides, many AMA tools operate as black-box systems, making it challenging to trace how decisions—such as study inclusion/exclusion criteria—are made. Moreover, if automation is trained on biased or incomplete datasets, it may affect the accuracy of evidence synthesis, thereby affecting clinical and policy-related decision-making. Furthermore, most existing AMA systems primarily rely on published studies indexed in databases, such as PubMed or Scopus, this focus may reinforce existing publication biases by systematically underrepresenting negative or inconclusive findings, particularly those available only in gray literature, preprints, or non-English sources. Addressing this issue will require developing more inclusive and adaptive search strategies within AMA frameworks.

The emergence of “thinking models” with complex reasoning in advanced AI and LLMs presents a transformative opportunity to revolutionize AMA by bridging computational power with sensitive analytical capabilities. They enable adaptive analytical strategies that can dynamically handle multi-modal datasets, reducing human intervention while maintaining methodological precision. To maximize the benefits of AMA, a balanced and methodologically rigorous approach is therefore essential, integrating “thinking LLMs” while mitigating its inherent challenges. Domain experts should remain actively involved in tasks, such as study selection, risk of bias assessment, sensitivity analyses, and interpretation, ensuring AMA outputs align with established research methodologies. Standardized reporting frameworks should be established to enhance the transparency of AMA methodologies, allowing researchers to audit and validate automated results. More sophisticated statistical modeling techniques and advanced AI techniques should be developed based on data complexity. Finally, as AMA tools become more widely adopted, policymakers should establish clear guidelines in evidence synthesis such as ethical considerations in AI-driven MAs and equitable access to AMA technologies. By integrating these principles and embracing AI-driven breakthroughs, AMA can evolve into a more robust and ethically responsible tool for evidence synthesis, bridging the gap between automation-driven efficiency and the need for methodological rigor and interpretability, and fundamentally transform evidence synthesis across disciplines.

6.5 Study limitations

This review is constrained by several interconnected challenges. First, this study’s predominant focus on well-documented tools from literature databases, potentially overlooks other innovative methodologies. Moreover, given the rapid evolution of automation technologies, particularly in artificial intelligence and LLMs, the review’s findings may quickly become outdated without regular updates. The dynamic nature of this field necessitates continuous revision to maintain relevance and usefulness. Second, while the proposed PPS with the TTF model offers a framework to understand AMA development, the criteria for assessing the level of automation remain subjective and qualitative, making it difficult to quantitatively compare the automation capabilities of different tools. Developing more standardized criteria for evaluating these tools would enhance the objectivity and reliability of future reviews. Third, this review is primarily based on the summary and classification of existing literature, without conducting empirical validation or performance evaluation (e.g., comparative experiments of different tools). As a result, some conclusions rely on self-reported findings in the reviewed studies, lacking independent external verification. These limitations highlight the need for ongoing research and development to refine AMA tools, address integration challenges, and ensure that they remain reliable and applicable in diverse research contexts.

7 Conclusion

MAs are critical to advancing science. The prospect of automating MAs opens opportunities for transforming quantitative research synthesis and redefining scientific progress. The automation of MAs is desperately needed right now to manage the expanding volume of academic research. We currently stand at the threshold of a significant AI revolution which holds potential to provide solutions to many remaining limitations and unsolved questions in the field of MA automation. To proceed effectively and maximize this opportunity, this study fills the gap in the literature by comprehensively investigating the current landscape of these automation efforts for MAs using a robust framework. This research has assessed existing methodological approaches, compared implementation patterns across various domains, and synthesized key challenges as well as future directions. Our emphasis has been on the potential that increasingly sophisticated LLMs with enhanced reasoning capabilities offer to accelerate progress further. Our research has found that automated tools have excelled at streamlining data extraction and statistical modeling, yet they still remain limited in achieving full-process automation, particularly in advanced synthesis and bias evaluation. Our work finds that future research efforts must prioritize the development of integrated frameworks that not only enhance individual meta-analytic stages but also bridge gaps between them. Efforts need to focus on also refining AI-driven models to improve interpretability and robustness, ensuring that heterogeneous data sources and complex synthesis tasks are effectively managed. Furthermore, standardizing methodologies across disciplines will be essential to unlock the full transformative potential of AMA.

Therefore, as the volume and complexity of academic research continue to escalate, the evolution of AMA represents a pivotal innovation for evidence synthesis. By harnessing advanced AI capabilities and addressing current methodological shortcomings, the research community can significantly enhance the efficiency, accuracy, and reproducibility of meta-analytic practices—ultimately revolutionizing the way we synthesize scientific knowledge.

Author contributions

Conceptualization: L.L. and T.S.; Data curation, formal analysis, methodology, investigation, visualization, and writing—original draft: L.L.; Project administration: A.M. and T.S.; Supervision, validation, and writing—review and editing: A.M. and T.S.

Competing interest statement

The authors declare that no competing interests exist.

Data availability statement

There is no dataset.

Funding statement

The authors declare that no specific funding has been received for this article.

Ethical standards

Research meets ethical standards—no ethics approval was required.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/rsm.2025.10065.

References

Schwab, K. The Fourth Industrial Revolution. Crown Publishing Group; 2017.Google Scholar
Brynjolfsson, E, Mitchell, T. What can machine learning do? Workforce implications. Science. 2017;358:15301534.10.1126/science.aap8062CrossRefGoogle ScholarPubMed
Marshall, IJ, Wallace, BC. Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. Syst Rev. 2019;8:110.10.1186/s13643-019-1074-9CrossRefGoogle ScholarPubMed
Kwabena, AE, Wiafe, OB, John, BD, Bernard, A, Boateng, FAF. An automated method for developing search strategies for systematic review using natural language processing (NLP). MethodsX. 2023;10:101935.10.1016/j.mex.2022.101935CrossRefGoogle ScholarPubMed
Tsafnat, G, Glasziou, P, Choong, MK, Dunn, A, Galgani, F, Coiera, E. Systematic review automation technologies. Syst Rev. 2014;3:74.10.1186/2046-4053-3-74CrossRefGoogle ScholarPubMed
Nedelcu, A, Oerther, B, Engel, H, et al. A machine learning framework reduces the manual workload for systematic reviews of the diagnostic performance of prostate magnetic resonance imaging. Eur Urol Open Sci. 2023;56:1114.10.1016/j.euros.2023.07.005CrossRefGoogle ScholarPubMed
Khraisha, Q, Put, S, Kappenberg, J, Warraitch, A, Hadfield, K. Can large language models replace humans in systematic reviews? Evaluating GPT-4’s efficacy in screening and extracting data from peer-reviewed and grey literature in multiple languages. Res Synth Methods. 2024;15:616626.10.1002/jrsm.1715CrossRefGoogle ScholarPubMed
Van Dinter, R, Tekinerdogan, B, Catal, C. Automation of systematic literature reviews: a systematic literature review. Inf Softw Technol. 2021;136:106589.10.1016/j.infsof.2021.106589CrossRefGoogle Scholar
Cooper, H. Research Synthesis and Meta-Analysis: A Step-by-Step Approach. 5th ed. SAGE Publications, Inc; 2017. https://doi.org/10.4135/9781071878644.CrossRefGoogle Scholar
Deeks, JJ, Higgins, JP, Altman, DG. Cochrane statistical methods group on behalf of the. Analysing data and undertaking meta-analyses. In: Cochrane Handbook for Systematic Reviews of Interventions. John Wiley & Sons, Ltd; 2019:241284, Chap. 10. https://doi.org/10.1002/9781119536604.ch10.CrossRefGoogle Scholar
Borah, R, Brown, AW, Capers, PL, Kaiser, KA. Analysis of the time and workers needed to conduct systematic reviews of medical interventions using data from the PROSPERO registry. BMJ Open. 2017;7:e012545.10.1136/bmjopen-2016-012545CrossRefGoogle ScholarPubMed
Higgins, J, Thomas, J, Chandler, J, et al. Cochrane handbook for systematic reviews of interventions. In: The Cochrane Collaboration. Wiley; 2019. https://doi.org/10.1002/9781119536604.Google Scholar
Xiong, Z, Liu, T, Tse, G, et al. A machine learning aided systematic review and meta-analysis of the relative risk of atrial fibrillation in patients with diabetes mellitus. Front Physiol. 2018;9:835.10.3389/fphys.2018.00835CrossRefGoogle ScholarPubMed
Chai, KEK, Lines, RLJ, Gucciardi, DF, Ng, L. Research screener: a machine learning tool to semi-automate abstract screening for systematic reviews. Syst Rev. 2021;10:93.10.1186/s13643-021-01635-3CrossRefGoogle ScholarPubMed
Feng, Y, Liang, S, Zhang, Y, et al. Automated medical literature screening using artificial intelligence: a systematic review and meta-analysis. J Am Med Inform Assoc. 2022;29:14251432.10.1093/jamia/ocac066CrossRefGoogle ScholarPubMed
Michelson, M. Automating meta-analyses of randomized clinical trials: a first look. In: Michalowski M, O’Sullivan D, Tenenbaum JM, Wilk S, eds. 2014 AAAI Fall Symposium Series. AAAI Press; 2014.Google Scholar
Mutinda, FW, Yada, S, Wakamiya, S, Aramaki, E. AUTOMETA: automatic meta-analysis system employing natural language processing. Stud Health Technol Inform. 2022;290:612616.Google ScholarPubMed
Büchter, RB, Rombey, T, Mathes, T, et al. Systematic reviewers used various approaches to data extraction and expressed several research needs: a survey. J Clin Epidemiol. 2023;159:214224.10.1016/j.jclinepi.2023.05.027CrossRefGoogle ScholarPubMed
Schmidt, L, Hair, K, Graziosi, S, et al. Exploring the use of a large language model for data extraction in systematic reviews: a rapid feasibility study. In: Proceedings of the 3rd Workshop on Augmented Intelligence for Technology-Assisted Reviews Systems (ALTARS 2024). 2024;3832.Google Scholar
Cheng, L, Katz-Rogozhnikov, DA, Varshney, KR, Baldini, I. Automated meta-analysis in medical research: a causal learning perspective. In: Ghassemi M, Naumann T, Pierson E, eds. ACM Conference on Health, Inference, and Learning. Association for Computational Linguistics; 2021.Google Scholar
Rodriguez-Hernandez, Z, Gorski, M, Tellez-Plaza, M, Schlosser, P, Wuttke, M. metaGWASmanager: a toolbox for an automated workflow from phenotypes to meta-analysis in GWAS consortia. Bioinformatics. 2024;40:btae294.10.1093/bioinformatics/btae294CrossRefGoogle ScholarPubMed
Ofori-Boateng, R, Aceves-Martins, M, Wiratunga, N, Moreno-Garcia, CF. Towards the automation of systematic reviews using natural language processing, machine learning, and deep learning: a comprehensive review. Artif Intell Rev. 2024;57:200.10.1007/s10462-024-10844-wCrossRefGoogle Scholar
Christopoulou, SC. Towards automated meta-analysis of clinical trials: an overview. BioMedInformatics. 2023;3:115140.10.3390/biomedinformatics3010009CrossRefGoogle Scholar
Ajiji, P, Cottin, J, Picot, C, et al. Feasibility study and evaluation of expert opinion on the semi-automated meta-analysis and the conventional meta-analysis. Eur J Clin Pharmacol. 2022;78:11771184.10.1007/s00228-022-03329-8CrossRefGoogle ScholarPubMed
Glass, GV, Smith, ML. Meta-analysis of research on class size and achievement. Educ Eval Policy Anal. 1979;1:216.10.3102/01623737001001002CrossRefGoogle Scholar
Egger, M, Smith, GD, Altman, D. Systematic Reviews in Health Care: Meta-Analysis in Context. John Wiley & Sons; 2008.Google Scholar
Collaboration CTT. Efficacy and safety of more intensive lowering of LDL cholesterol: a meta-analysis of data from 170,000 participants in 26 randomised trials. Lancet. 2010;376:16701681.10.1016/S0140-6736(10)61350-5CrossRefGoogle Scholar
Hedges, LV, Olkin, I. Statistical Methods for Meta-Analysis. Academic press; 2014.Google Scholar
Takkouche, B, Norman, G. PRISMA statement. Epidemiology. 2011;22:128.10.1097/EDE.0b013e3181fe7999CrossRefGoogle Scholar
Page, MJ, Moher, D, Bossuyt, PM, et al. PRISMA 2020 explanation and elaboration: updated guidance and exemplars for reporting systematic reviews. BMJ. 2021;372:n71.10.1136/bmj.n71CrossRefGoogle ScholarPubMed
Page, MJ, McKenzie, JE, Bossuyt, PM, et al. Updating guidance for reporting systematic reviews: development of the PRISMA 2020 statement. J Clin Epidemiol. 2021;134:103112.10.1016/j.jclinepi.2021.02.003CrossRefGoogle ScholarPubMed
Page, MJ, McKenzie, JE, Bossuyt, PM, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71.10.1136/bmj.n71CrossRefGoogle ScholarPubMed
Sackett, DL. Evidence-based medicine. Semin Perinatol. 1997;21:35. Fatal and Neonatal Hematology for the 21st Century.10.1016/S0146-0005(97)80013-4CrossRefGoogle ScholarPubMed
Matsui, K, Utsumi, T, Aoki, Y, Maruki, T, Takeshima, M, Yoshikazu, T. Large language model demonstrates human-comparable sensitivity in initial screening of systematic reviews: a semi-automated strategy using GPT-3.5. 2023.10.2139/ssrn.4520426CrossRefGoogle Scholar
Brown, T, Mann, B, Ryder, N, et al. Language models are few-shot learners. In: Larochelle, H, Ranzato, M, Hadsell, R, Balcan, M, Lin, H, eds. Advances in Neural Information Processing Systems. Vol. 33. Curran Associates, Inc.; 2020:18771901.Google Scholar
Wei, J, Wang, X, Schuurmans, D, et al. Chain-of-thought prompting elicits reasoning in Large language models. In: Koyejo, S, Mohamed, S, Agarwal, A, Belgrave, D, Cho, K, Oh, A, eds. Advances in Neural Information Processing Systems. Vol. 35. Curran Associates, Inc.; 2022:2482424837.Google Scholar
Huang, J, Chang, KCC. Towards reasoning in large language models: a survey. Findings of the Association for Computational Linguistics: ACL 2023. Toronto: Association for Computational Linguistics; 2023:10491065. https://arxiv.org/abs/2212.10403.10.18653/v1/2023.findings-acl.67CrossRefGoogle Scholar
Venkatesh, V, Morris, MG, Davis, GB, Davis, FD. User acceptance of information technology: toward a unified view. MIS Q. 2003;27:425.10.2307/30036540CrossRefGoogle Scholar
Venkatesh, V, Thong, JYL, Xu, X. Consumer acceptance and use of information technology: extending the unified theory of acceptance and use of technology. MIS Q. 2012;36:157178.10.2307/41410412CrossRefGoogle Scholar
Goodhue, DL, Thompson, RL. Task-technology fit and individual performance. MIS Q. 1995;19:213236.10.2307/249689CrossRefGoogle Scholar
Dwivedi, YK, Rana, NP, Jeyaraj, A, Clement, M, Williams, MD. Re-examining the unified theory of acceptance and use of technology (UTAUT): towards a revised theoretical model. Inf Syst Front. 2019;21:719734.10.1007/s10796-017-9774-yCrossRefGoogle Scholar
D’Ambra, J, Wilson, CS, Akter, S. Application of the task-technology fit model to structure and evaluate the adoption of E-books by academics. J Am Soc Inf Sci Technol. 2013;64:4864.10.1002/asi.22757CrossRefGoogle Scholar
Ali, SB, Romero, J, Morrison, K, Hafeez, B, Ancker, JS. Focus section health IT usability: applying a task-technology fit model to adapt an electronic patient portal for patient work. Appl Clin Inform. 2018;9:174184.Google ScholarPubMed
Goodman, LA. Snowball sampling. Ann Math Stat. 1961;32:148170.10.1214/aoms/1177705148CrossRefGoogle Scholar
Parker, C, Scott, S, Geddes, A. Snowball sampling. SAGE Research Methods Foundations, 2019.Google Scholar
Aalami, AH, Aalami, F, Sahebkar, A. Apolipoprotein A-1 as a potential biomarker for solid tumors: a systematic review and meta-analysis. Curr Med Chem. 2023;30:33563367.10.2174/0929867330666230210112700CrossRefGoogle ScholarPubMed
Luo, X, Chen, F, Zhu, D, et al. Potential roles of large language models in the production of systematic reviews and meta-analyses. J Med Internet Res. 2024;26:e56780.10.2196/56780CrossRefGoogle ScholarPubMed
Quan, Y, Tytko, T, Hui, B. Utilizing ASReview in screening primary studies for meta-research in SLA: a step-by-step tutorial. Res Methods Appl Linguist. 2024;3:100101.10.1016/j.rmal.2024.100101CrossRefGoogle Scholar
Data, S, Lee, K, Paek, H, et al. MSR126 AutoCriteria: advancing clinical trial study with AI-powered eligibility criteria extraction. Value Health. 2023;26:S417.10.1016/j.jval.2023.09.2185CrossRefGoogle Scholar
Siberchicot, A, Bessy, A, Guéguen, L, Marais, GAB. MareyMap online: a user-friendly web application and database service for estimating recombination rates using physical and genetic maps. Genome Biol Evol. 2017;9:25062509.10.1093/gbe/evx178CrossRefGoogle ScholarPubMed
Bosco, FA, Uggerslev, KL, Steel, P. MetaBUS as a vehicle for facilitating meta-analysis. Hum Resour Manag Rev. 2017;27:237254.Google Scholar
Yang, X, Tang, H, Dongye, X, Chen, G. Exploration of meta analysis automation. In: 2018 International Conference on Network Infrastructure and Digital Content (IC-NIDC). 2018:218222. https://doi.org/10.1109/ICNIDC.2018.8525710.CrossRefGoogle Scholar
Deng, Z, Yin, K, Bao, Y, et al. Validation of a semiautomated natural language processing-based procedure for meta-analysis of cancer susceptibility gene penetrance. JCO Clin Cancer Inform. 2019;3:19.10.1200/CCI.19.00043CrossRefGoogle ScholarPubMed
Wei, Y, Zhang, T, Zhang, H, et al. Chat2Brain: a method for mapping open-ended semantic queries to brain activation maps. In: 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). 2023:15231530. https://doi.org/10.1109/BIBM58861.2023.10385933.CrossRefGoogle Scholar
Issaiy, M, Ghanaati, H, Kolahi, S, et al. Methodological insights into ChatGPT’s screening performance in systematic reviews. BMC Med Res Methodol. 2024;24:78.10.1186/s12874-024-02203-8CrossRefGoogle ScholarPubMed
Luo, R, Sastimoglu, Z, Faisal, AI, Deen, MJ. Evaluating the efficacy of large language models for systematic review and meta-analysis screening. 2024. https://doi.org/10.1101/2024.06.03.24308405.CrossRefGoogle Scholar
Boyko, AA, Kaidina, AM, Kim, YC, et al. A framework for automated meta-analysis: dendritic cell therapy case study. In: 2016 IEEE 8th International Conference on Intelligent Systems (IS). 2016:160166. https://doi.org/10.1109/IS.2016.7737416.CrossRefGoogle Scholar
Neppalli, K, Caragea, C, Mayes, R, Nimon, K, Oswald, F. MetaSeer.STEM: towards automating meta-analyses. In: Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 30. 2016:40354040. https://doi.org/10.1609/aaai.v30i2.19081. (visited on 09/30/2024).CrossRefGoogle Scholar
Lorenz, MW, Abdi, NA, Scheckenbach, F, et al. Automatic identification of variables in epidemiological datasets using logic regression. BMC Med Inform Decis Mak. 2017;17:40.10.1186/s12911-017-0429-1CrossRefGoogle ScholarPubMed
Devyatkin, D, Molodchenkov, A, Lukin, A, et al. Towards automated meta-analysis of biomedical texts in the field of cell-based immunotherapy. Biomedical Chemistry: Research and Methods. 2019;2:e00109.Google Scholar
Pradhan, R, Hoaglin, DC, Cornell, M, Liu, W, Wang, V, Yu, H. Automatic extraction of quantitative data from ClinicalTrials.Gov to conduct meta-analyses. J Clin Epidemiol. 2019;105:92100.10.1016/j.jclinepi.2018.08.023CrossRefGoogle ScholarPubMed
Anisienia, A, Mueller, RM, Kupfer, A, Staake, T. In: Bui, TX, ed. Research Method Classification with Deep Transfer Learning for Semi-Automatic Meta-Analysis of Information Systems Papers. Vol. 2020. IEEE Computer Society; 2021:60996108.Google Scholar
Alisa, G, Dmitry, D, Anton, L, Alexey, L, Alexey, M, Irina, K. Method for biomedical information extraction of immunosuppressive cell properties. In: 2021 IEEE Ural-Siberian Conference on Computational Technologies in Cognitive Science, Genomics and Biomedicine (CSGB). 2021:210213. https://doi.org/10.1109/CSGB53040.2021.9496030.CrossRefGoogle Scholar
Donoghue, T, Voytek, B. Automated meta-analysis of the event-related potential (ERP) literature. Sci Rep. 2022;12:1867.10.1038/s41598-022-05939-9CrossRefGoogle ScholarPubMed
Mutinda, FW, Liew, K, Yada, S, Wakamiya, S, Aramaki, E. Automatic data extraction to support meta-analysis statistical analysis: a case study on breast cancer. BMC Med Inform Decis Mak. 2022;22:158.10.1186/s12911-022-01897-4CrossRefGoogle Scholar
Zhang, X, Wang, C, Yao, Y, et al. Construction of a meta-evidence prototype database of traditional Chinese medicine splenogastric diseases and its application in an automatic meta-analysis system. Evid Based Complement Alternat Med. 2022;2022:6933523.Google Scholar
Kartchner, D, Ramalingam, S, Al-Hussaini, I, Kronick, O, Mitchell, C. Zero-shot information extraction for clinical meta-analysis using Large language models. In: Demner-fushman, D, Ananiadou, S, Cohen, K, eds. The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks. Association for Computational Linguistics; 2023:396405. https://doi.org/10.18653/v1/2023.bionlp-1.37.CrossRefGoogle Scholar
Shah-Mohammadi, F, Large, FJ. Language model-based architecture for automatic outcome data extraction to support meta-analysis. In: 2024 IEEE 14th Annual Computing and Communication Workshop and Conference (CCWC). 2024:7985. https://doi.org/10.1109/CCWC60891.2024.10427829.CrossRefGoogle Scholar
Yun, HS, Pogrebitskiy, D, Marshall, IJ, Wallace, BC. Automatically extracting numerical results from randomized controlled trials with large language models. In: Deshpande, K, Fiterau, M, Joshi, S, Lipton, Z, Ranganath, R, Urteaga, I, eds. Proceedings of the 9th Machine Learning for Healthcare Conference. Vol. 252. Proceedings of Machine Learning Research; 2024:124.Google Scholar
Wang, X, Luo, G. MetaMate: Large language model to the rescue of automated data extraction for educational systematic reviews and meta-analyses. Preprint, Washington, DC; 2024. https://doi.org/10.35542/osf.io/wn3cd.CrossRefGoogle Scholar
Choi, H, Shen, R, Chinnaiyan, AM, Ghosh, D. A latent variable approach for meta-analysis of gene expression Data from multiple microarray experiments. BMC Bioinformatics. 2007;8:364.10.1186/1471-2105-8-364CrossRefGoogle ScholarPubMed
Marot, G, Foulley, JL, Mayer, CD, Jaffrézic, F. Moderated effect size and P-value combinations for microarray meta-analyses. Bioinformatics. 2009;25:26922699.10.1093/bioinformatics/btp444CrossRefGoogle ScholarPubMed
Viechtbauer, W. Conducting meta-analyses in R with the metafor package. J Stat Softw. 2010;36:148.10.18637/jss.v036.i03CrossRefGoogle Scholar
Willer, CJ, Li, Y, Abecasis, GR. METAL: fast and efficient meta-analysis of genomewide association scans. Bioinformatics. 2010;26:21902191.10.1093/bioinformatics/btq340CrossRefGoogle ScholarPubMed
Wang, X, Kang, DD, Shen, K, et al. An R package suite for microarray meta-analysis in quality control, differentially expressed gene analysis and pathway enrichment detection. Bioinformatics. 2012;28:25342536.10.1093/bioinformatics/bts485CrossRefGoogle Scholar
Suurmond, R, Van Rhee, H, Hak, T. Introduction, comparison, and validation of meta-essentials: a free and simple tool for meta-analysis. Res Synth Methods. 2017;8:537553.10.1002/jrsm.1260CrossRefGoogle ScholarPubMed
Debray, TP, Damen, JA, Riley, RD, et al. A framework for meta-analysis of prediction model studies with binary and time-to-event outcomes. Stat Methods Med Res. 2019;28:27682786.10.1177/0962280218785504CrossRefGoogle ScholarPubMed
Dockès, J, Poldrack, RA, Primet, R, et al. NeuroQuery, comprehensive meta-analysis of human brain mapping. elife. 2020;9:e53385.10.7554/eLife.53385CrossRefGoogle ScholarPubMed
Peñaloza, R. Towards a logic of meta-analysis. In: Proceedings of the Seventeenth International Conference on Principles of Knowledge Representation and Reasoning. International Joint Conferences on Artificial Intelligence Organization; 2020:672676. https://doi.org/10.24963/kr.2020/68.CrossRefGoogle Scholar
Llambrich, M, Correig, E, Gumà, J, Brezmes, J, Cumeras, R. Amanida: an R package for meta-analysis of metabolomics non-integral data. Bioinformatics 2022;38:583585. Ed. by Wren J.10.1093/bioinformatics/btab591CrossRefGoogle Scholar
Lu, Y, Zhou, G, Ewald, J, Pang, Z, Shiri, T, Xia, J. MicrobiomeAnalyst 2.0: comprehensive statistical, functional and integrative analysis of microbiome data. Nucleic Acids Res. 2023;51:W310W318.10.1093/nar/gkad407CrossRefGoogle ScholarPubMed
Van Valkenhoef, G, Lu, G, De Brock, B, Hillege, H, Ades, AE, Welton, NJ. Automating network meta-analysis. Res Synth Methods. 2012;3:285299.10.1002/jrsm.1054CrossRefGoogle ScholarPubMed
Van Valkenhoef, G, Dias, S, Ades, AE, Welton, NJ. Automated generation of node-splitting models for assessment of inconsistency in network meta-analysis. Res Synth Methods. 2016;7:8093.10.1002/jrsm.1167CrossRefGoogle ScholarPubMed
Thom, H, White, IR, Welton, NJ, Lu, G. Automated methods to test connectedness and quantify indirectness of evidence in network meta-analysis. Res Synth Methods. 2019;10:113124.10.1002/jrsm.1329CrossRefGoogle ScholarPubMed
Yarkoni, T, Poldrack, RA, Nichols, TE, Van Essen, DC, Wager, TD. Large-scale automated synthesis of human functional neuroimaging data. Nat Methods. 2011;8:665670.10.1038/nmeth.1635CrossRefGoogle ScholarPubMed
Feichtinger, J, McFarlane, RJ, Larcombe, LD. CancerMA: a web-based tool for automatic meta-analysis of public cancer microarray data. Database. 2012;2012:91.10.1093/database/bas055CrossRefGoogle ScholarPubMed
Feichtinger, J, McFarlane, RJ, Larcombe, LD. CancerEST: a web-based tool for automatic meta-analysis of public EST data. Database. 2014;2014:bau024.10.1093/database/bau024CrossRefGoogle ScholarPubMed
Shashirekha, HL, Wani, AH. ShinyMDE: shiny tool for microarray meta-analysis for differentially expressed gene detection. In: 2016 International Conference on Bioinformatics and Systems Biology (BSB). 2016:15. https://doi.org/10.1109/BSB.2016.7552152.CrossRefGoogle Scholar
Craig, A, Bae, SH, Taswell, C. Bridging the semantic and lexical webs: Concept-validating and hypothesis-exploring ontologies for the Nexus-PORTAL-DOORS System. Journal of Systemics, Cybernetics and Informatics. 2017;15:813.Google Scholar
Hu, Z, Jujjavarapu, C, Hughey, JJ, et al. MetaCyto: a tool for automated meta-analysis of mass and flow cytometry Data. Cell Rep. 2018;24:13771388.10.1016/j.celrep.2018.07.003CrossRefGoogle ScholarPubMed
Ryu, SY, Wendt, GA. MetaMSD: meta analysis for mass spectrometry data. PeerJ. 2019;7:e6699.10.7717/peerj.6699CrossRefGoogle ScholarPubMed
Lam, M, Awasthi, S, Watson, HJ, et al. RICOPILI: rapid imputation for COnsortias PIpeLIne. Bioinformatics. 2020;36:930933.10.1093/bioinformatics/btz633CrossRefGoogle ScholarPubMed
Ngo, GH, Nguyen, M, Chen, NF, Sabuncu, MR. Text2Brain: synthesis of brain activation maps from free-form text query. In: de Bruijne M, Cattin PC, Cotin S, Padoy N, Speidel S, Zheng Y, Essert C, eds.Medical Image Computing and Computer Assisted Intervention—MICCAI. Vol. 12907. 2021:605614.10.1007/978-3-030-87234-2_57CrossRefGoogle Scholar
Sabates, J, Belleville, S, Castellani, M, et al. CogTale: an online platform for the evaluation, synthesis, and dissemination of evidence from cognitive interventions studies. Syst Rev. 2021;10:236.10.1186/s13643-021-01787-2CrossRefGoogle ScholarPubMed
Burgard, T, Bosnjak, M, Studtrucker, R. PsychOpen CAMA: publication of community-augmented meta-analyses in psychology. Res Synth Methods. 2022;13:134143.10.1002/jrsm.1536CrossRefGoogle ScholarPubMed
Finnigan, W, Lubberink, M, Hepworth, LJ, et al. RetroBioCat database: a platform for collaborative curation and automated meta-analysis of biocatalysis data. ACS Catal. 2023;13:1177111780.10.1021/acscatal.3c01418CrossRefGoogle ScholarPubMed
Bremer, PL, Wohlgemuth, G, Fiehn, O. The BinDiscover database: a biology-focused meta-analysis tool for 156,000 GC–TOF MS metabolome samples. J Chem. 2023;15:66.Google ScholarPubMed
Kale, A, Lee, S, Goan, T, Tipton, E, Hullman, J. Metaexplorer: Facilitating Reasoning with Epistemic Uncertainty in Meta-Analysis. Association for Computational Linguistics; 2023. https://dl.acm.org/doi/abs/10.1145/3544548.3580869.CrossRefGoogle Scholar
Van Valkenhoef, G, Tervonen, T, Zwinkels, T, De Brock, B, Hillege, H. ADDIS: a decision support system for evidence-based medicine. Decis Support Syst. 2013;55:459475.10.1016/j.dss.2012.10.005CrossRefGoogle Scholar
Neupane, B, Richer, D, Bonner, AJ, Kibret, T, Beyene, J. Network meta-analysis using R: a review of currently available automated packages. PLoS One. 2014;9:e115065.10.1371/journal.pone.0115065CrossRefGoogle Scholar
Owen, RK, Bradbury, N, Xin, Y, Cooper, N, Sutton, A. MetaInsight: an interactive web-based tool for analyzing, interrogating, and visualizing network meta-analyses using R-shiny and netmeta. Res Synth Methods. 2019;10:569581.10.1002/jrsm.1373CrossRefGoogle ScholarPubMed
Nikolakopoulou, A, Higgins, JP, Papakonstantinou, T, et al. CINeMA: an approach for assessing confidence in the results of a network meta-analysis. PLoS Med. 2020;17:e1003082.10.1371/journal.pmed.1003082CrossRefGoogle ScholarPubMed
Chiocchia, V, Holloway, A, Salanti, G. Semi-automated assessment of the risk of bias due to missing evidence in network meta-analysis: a guidance paper for the ROB-MEN web-application. BMC Med Res Methodol. 2023;23:223.10.1186/s12874-023-02038-9CrossRefGoogle ScholarPubMed
Liu, Y, Béliveau, A, Wei, Y, Chen, M. A gentle introduction to Bayesian network meta-analysis using an automated R package. Multivar Behav Res. 2023;58:706722.10.1080/00273171.2022.2115965CrossRefGoogle ScholarPubMed
Reason, T, Benbow, E, Langham, J, Gimblett, A, Klijn, SL, Malcolm, B. Artificial intelligence to automate network meta-analyses: four case studies to evaluate the potential application of Large language models. PharmacoEconomics Open. 2024;8:205220.10.1007/s41669-024-00476-9CrossRefGoogle ScholarPubMed
Zimmermann, P, Hirsch-Hoffmann, M, Hennig, L, Gruissem, W. GENEVESTIGATOR. Arabidopsis microarray database and analysis toolbox. Plant Physiol. 2004;136:26212632.10.1104/pp.104.046367CrossRefGoogle ScholarPubMed
Parkinson, H. ArrayExpress–a public repository for microarray gene expression data at the EBI. Nucleic Acids Res. 2004;33:D553D555.10.1093/nar/gki056CrossRefGoogle Scholar
Hong, F, Breitling, R, McEntee, CW, Wittner, BS, Nemhauser, JL, Chory, J. RankProd: a bioconductor package for detecting differentially expressed genes in meta-analysis. Bioinformatics. 2006;22:28252827.10.1093/bioinformatics/btl476CrossRefGoogle ScholarPubMed
Liu, NF, Lin, K, Hewitt, J, et al. Lost in the middle: how language models use long contexts. Transactions of the Association for Computational Linguistics. 2024;12:157173.10.1162/tacl_a_00638CrossRefGoogle Scholar
Wijkstra, M, Lek, T, Kuhn, T, Welbers, K, Steijaert, M. Living literature reviews. In: Gentile AL, Gonçalves R, eds. Proceedings of the 11th Knowledge Capture Conference. ACM; 2021:241248. https://dl.acm.org/doi/10.1145/3460210.3493567.CrossRefGoogle Scholar
Boutron, I, Chaimani, A, Devane, D, et al. Interventions for the prevention and treatment of COVID-19: a living mapping of research and living network meta-analysis. Cochrane Database Syst Rev. 2020;11:CD013769.Google Scholar
Evrenoglou, T, Boutron, I, Seitidis, G, Ghosn, L, Chaimani, A. metaCOVID: a web-application for living meta-analyses of COVID-19 trials. Res Synth Methods. 2023;14:479488.10.1002/jrsm.1627CrossRefGoogle ScholarPubMed
Hair, K, Wilson, E, Wong, C, Tsang, A, Macleod, M, Bannach-Brown, A. Systematic online living evidence summaries: emerging tools to accelerate evidence synthesis. Clin Sci. 2023;137:773784.10.1042/CS20220494CrossRefGoogle ScholarPubMed
Figure 0

Table 1 Criteria for study selection using PRISMA

Figure 1

Figure 1 PRISMA workflow.Note: “Too short” = records with fewer than four pages, excluded for lacking methodological detail.

Figure 2

Figure 2 Holistic framework for this review.

Figure 3

Figure 3 Progressive phase structure with TTF model.Note: “PreQ”, “ProcQ”, and “PostQ” denote analytical questions from the pre-processing, processing, and post-processing stages, respectively. Assessment ratings (H = High, M = Moderate, L = Low) are defined above.

Figure 4

Table 2 Task-level mapping between PPS automation phases and Cochrane methodological stages

Figure 5

Figure 4 Temporal patterns in AMA publications (a) and proportional discrepancies across different stages (b).

Figure 6

Table 3 Task–technology fit assessment for pre-processing in AMA

Figure 7

Table 4 Task–technology fit assessment for automated data processing in CMA

Figure 8

Table 5 Task–technology fit assessment for automated data processing in NMA

Figure 9

Table 6 Task–technology fit assessment for automated data post-processing in CMA

Figure 10

Table 7 Task–technology fit assessment for automated data post-processing in NMA

Figure 11

Figure 5 Interdisciplinary applications of AMA across various domains.

Figure 12

Figure 6 Cross-domain mapping of AMA applications, data types, and methodological approaches.

Figure 13

Table 8 Future research directions for AMA

Figure 14

Table 9 Comparative analysis of CMA and NMA automation

Supplementary material: File

Li et al. supplementary material

Li et al. supplementary material
Download Li et al. supplementary material(File)
File 187.2 KB