Introduction
As scientific challenges become more complex, solutions increasingly require interdisciplinary approaches, necessitating team science as a critical strategy for advancing discovery and translating research into practice. The need for effective collaboration is particularly pronounced in clinical and translational research (CTR), where bridging the gap between basic science and real-world healthcare applications requires high-performing research teams. An extensive body of literature documents that ineffective collaboration causes poor performance, dissatisfaction, and high turnover of team members [Reference Lin and Huang1,Reference Park and Johnson2]. Despite the importance of effective teaming in science, training in team science competencies is only slowly being integrated into post-baccalaureate educational programs [Reference Mendell, Fritter and Helm3–Reference Fiore, Gabelica, Wiltshire, Stokols, Hall, Vogel and Croyle6]. The Science of Team Science field has expanded over the past two decades to study and improve scientific collaboration [Reference Hall, Vogel and Croyle7,Reference Hall, Vogel and Huang8]. The field equally strives to research and document factors associated with success in science teams, as well as build a body of literature on impactful training and team interventions to improve team effectiveness and the capacity of integrative teams to solve complex problems [Reference Hall, Vogel and Huang8].
Scientists increasingly recognize the requirements, challenges, and opportunities to enhance collaborative capacity for interdisciplinary projects [Reference Fiore, Gabelica, Wiltshire, Stokols, Hall, Vogel and Croyle6,Reference Love, Fosdick and Cross9]. Core competencies of collaborative capacity include (1) the planning necessary to build and grow a team, (2) the management required to monitor progress and stay on track, and (3) the fostering of effective Interpersonal Relations among team members needed for mutual learning and social support [Reference Bisbey, Wooten, Campo, Lant and Salas5,Reference Lotrecchiano, DiazGranados and Sprecher12]. Many Clinical and Translational Science Award (CTSA) Hubs and interdisciplinary research centers have introduced Team Science training programs to strengthen these competencies, particularly in biomedical and health research [Reference Mayowski, Norman, Schenker, Proulx and Kapoor4,Reference Rolland, Cross, Hohl, Johnson, Wooten and Brasier13,Reference Read, O’Rourke and Hong14]. Yet, despite the proliferation of training programs, there remains little empirical evidence demonstrating their impact on actual teaming behaviors, scientific outputs, or research impact [Reference Mayowski, Norman, Schenker, Proulx and Kapoor4,Reference Bisbey, Wooten, Campo, Lant and Salas5,Reference Abu-Rish Blakeney, Kang and Henrikson15].This paper introduces and validates an assessment tool for Team Science competencies and reports findings from its application in evaluating a structured training program. Unlike research on traditional educational domains, such as reading and mathematics, where learning mechanisms are well studied, the effects of Team Science training on real-world research teams remain poorly understood. For example, regarding reading ability in elementary school, we know that children with higher baseline reading skills tend to improve their reading skills more quickly, a phenomenon known as the Matthew effect. The explanation is that children who excel at reading receive praise for their abilities, enjoy reading, and consequently read more and learn faster [Reference Duff, Tomblin and Catts16,Reference Petersen, Jung, Yang and Stanley17]. On the other hand, there is also the Dunning–Kruger effect, which is well-documented among students in medicine and psychology, where novices tend to overestimate their competence because they lack the skills to accurately recognize deficient performance [Reference Bradley, Dreifuerst, Johnson and Loomis18,Reference Fitzmaurice19]. For Team Science, it remains unclear whether a similar Matthew Effect exists, where individuals with stronger teaming skills thrive and continue to develop, or whether a lack of awareness about ineffective teaming practices prevents researchers from recognizing their deficits until they are exposed to structured training. On the one hand, good team players may enjoy the collaborative process, be successful in science teams, learn from their experiences, and feel encouraged to apply their collaborative skills to more team-based research. On the other hand, researchers may be unaware of their lack of teaming competencies. They may not recognize this deficiency or its impact until they are exposed to team science frameworks and experience better teaming practices.
This study evaluated the effectiveness of the Colorado Clinical and Translational Science Institute’s (CCTSI) Team Science training program using pre- and post-program surveys. The primary objectives were twofold: (1) to validate the grouping of Team Science competencies – namely, Team Planning, Managing a Team, and Interpersonal Relations – within the newly developed Teaming Evaluation and Assessment for Multidisciplinary Science (TEAMS) instrument, and (2) to assess, through self-reports, the program’s impact on enhancing teaming capacity in clinical and translational science teams. The TEAMS instrument was designed to measure key competencies in team science. By analyzing pre- and post-program survey data, the study aimed to determine whether the training effectively improved participants’ self-reported competencies in these three domains. The findings offer valuable insights into participants’ experiences with the Team Science training program in Colorado and highlight areas for further research to enhance the effectiveness of such training initiatives.
Methods
Team science program
The CCTSI’s Team Science training program was provided three times a year. The program began with a self-paced online Teaming 101 web module, comprising six modules featuring 20-minute lecture videos, scenario videos, and a short assessment to evaluate participants’ understanding of core concepts. Completion of the Teaming 101 web module was mandatory before attending any of the six virtual (Zoom) interactive workshops. This prerequisite ensured that all participants had a foundational understanding of team science concepts before engaging in interactive discussions. After completing the Teaming 101 web module, participants could register for any workshop in any order. The six workshops covered key domains of Team Science, focusing on building relationships and trust (Workshop 1), setting expectations (Workshop 2), shared language and vision (Workshop 3), collaborative knowledge creation (Workshop 4), change management and conflict resolution (Workshop 5), and leadership (Workshop 6). Each workshop addressed specific competencies, such as psychological safety, authorship agreements, reflexivity, power dynamics, and conflict management (see Table 1 for a detailed breakdown of workshop topics).
Table 1. Description of the topics covered in the six interactive virtual workshops facilitated on zoom

All workshops employed evidence-based active learning strategies, including small-group discussions, polls, role-playing, skill practice, problem-solving, group ideation, and reflection [Reference Arikan, Dochy and Segers20]. After each workshop, participants received a written summary of key content and takeaways for future reference.
Target audience
The program was developed for and open to all early-career investigators and clinical research professionals. Specifically, early-career investigators were defined as pre- and post-doctoral students, instructors, and assistant professors. The CCTSI community includes three University of Colorado (CU) campuses (CU Boulder, CU Denver, CU Anschutz Medical Campus), Colorado State University, and healthcare partners (Children’s Hospital of Colorado, Denver Health, National Jewish Health, University of Colorado Health, and the Rocky Mountain Regional VA Medical Center). Potential participants were informed of the program through CCTSI newsletters, email blasts, and the CCTSI website.
Pre- and post-program assessments
The University of Colorado Institutional Review Board deemed the study exempt and advised that these activities are program evaluation rather than human subject research. As part of the registration process, participants were required to complete a registration form, which collected their demographic characteristics. Participants then received the URL link through Qualtrics to complete the Teaming 101 web module. A pre-post survey assessed and measured changes in familiarity and frequency of use in applying teaming concepts. The pre-program survey was administered as part of the Teaming 101 web module. Post-program surveys were sent to participants one month after attending all six workshops.
To systematically assess the effectiveness of the CCTSI Team Science training program, we developed the Team Evaluation and Assessment for Multidisciplinary Science (TEAMS) instrument. This tool evaluated key competencies for effective team science, structured around three core domains: Team Planning, Managing a Team, and Interpersonal Skills. The development process involved identifying specific behaviors and strategies pertinent to each domain, ensuring that the instrument accurately captures the multifaceted nature of team-based scientific collaboration. By aligning the TEAMS instrument with the objectives of the training program, we aimed to provide a reliable measure of participants’ competencies and the program’s impact on enhancing team science skills. In the pre- and post-program TEAMS surveys, participants were asked to self-report their familiarity with Team Science competencies for Team Planning, including building a shared language, establishing a shared vision, setting ground rules, and creating team charters. Additionally, participants were asked to report the frequency of using team science strategies when working on research projects for two other components: Managing a Team, which involves creating meeting agendas, monitoring progress, establishing authorship agreements, and incorporating others’ perspectives, and Interpersonal Skills, which includes perspective-taking, facilitating awareness and exchange, acknowledging and including diversity, and promoting equal turn-taking. Familiarity with the Team Planning and Managing a Team concepts was assessed using a five-point Likert scale, with responses ranging from 1 (“Strongly Disagree”) to 5 (“Strongly Agree”). The frequency of using team science strategies for Interpersonal Relations Skills was measured on a separate five-point scale: 1 (“Never”), 2 (“Rarely”), 3 (“Sometimes”), 4 (“Often”), and 5 (“Always”). Team Planning and Managing a Team were assessed using a familiarity scale because these areas introduced new concepts to participants, necessitating an evaluation of their understanding. In contrast, Interpersonal Relations competencies were measured using a frequency-of-use scale, as participants were presumed to be already acquainted with these concepts; the focus was thus on how often they applied these skills to enhance collaboration.
Data analysis
We applied principal component analysis (PCA) to confirm underlying factors in participants’ responses across three Team Science skill domains. PCA is a statistical technique that reduces the dimensionality of complex datasets by transforming correlated variables into a smaller set of uncorrelated variables, called principal components, which retain most of the original information. Using factor scores based on the three components, the study examined whether self-reported teaming skills improved significantly, as determined by paired t-tests. Pre-program scores were subtracted from post-program scores to determine the impact of the CCTSI team science training program. While the dataset was nearly complete, pairwise deletion was applied to the few missing data points. All surveys were analyzed using STATA 18.
Results
Participant characteristics
The response rate for the pre-program survey was 53% (117 of 221 registered participants). All participants completed the online prerequisite Teaming 101 web module, and 46 out of 81 participants who completed all six workshops responded to the post-program survey, yielding a response rate of 57%. Only individuals who completed all six workshops are included in this analysis. The post-program survey had fewer respondents (n = 46) than the pre-program survey (n = 117), resulting in a notable drop-off in participation (see Table 2 for comparison).
Table 2. Demographics of team science training participants in the sample for pre-and post-program assessments

The sample comprised three cohorts from 2020 to 2022. The Team Science training program participants included graduate students, postdocs, instructors, assistant professors, clinical research professionals, research coordinators, and research assistants. The career stages of the respondents varied, with the categories being faculty (31%), graduate students (30%), postdoctoral trainees (26%), and clinical research professional staff (14%). Most participants held a doctoral degree (65%) and were female (74%). Most participants were White (78%) and of non-Hispanic descent (92%), which is comparable to the overall demographics of Colorado [21]. Five percent of participants reported living with a disability, 24% were first-generation students, and 9% were from a low-income family.
Performance characteristics of the tool
To assess how well the TEAMS assessment tool fulfills its function – measuring Team Science competencies and organizing related skills into meaningful categories – we conducted PCA. This statistical method helps identify patterns by grouping related skills into clusters. Using orthogonal varimax rotation; we found that all items load on the respective components of Team Planning, Managing a Team, or Interpersonal Relationships with eigenvalues of 2.97, 2.77, and 2.12, respectively, well over the commonly used cut-off score of 1 [Reference Kline22]. Each component demonstrated strong internal consistency, indicating that the skills within each category are highly related and measure the same underlying competency. These three components explain 23%, 21%, and 16% of the variance. All the items loaded above the commonly used cut-off score of 0.3 on their respective component, and none loaded on more than one component, indicating a clear and distinct factor structure, where each item is strongly associated with a single underlying competency domain. Such a structure enhances the interpretability and validity of the assessment tool, confirming that it effectively measures the intended Team Science competencies. Higher loadings suggest that an item aligns more closely with that specific domain, meaning it contributes significantly to measuring that competency. Furthermore, Cronbach’s’ alpha – which assesses how reliably the set of items within each category measures the same underlying construct – suggests excellent internal consistency for all three components with scale reliability coefficients of 0.84 for the Team Planning component, 0.74 for the Team Managing component, and 0.81 for the Interpersonal Relations component (see Table 3). These scores are above the commonly recommended minimum value of 0.7 for scale reliability [Reference Bujang, Omar and Baharum23].
Table 3. Team planning, managing a team, and interpersonal relations components with corresponding loadings (orthogonal varimax rotation) and cronbach’s alpha (α) scale reliability coefficient

Note: Values greater than 0.30 are bold and indicate an association with one of the three components.
Mean pre- and post-program familiarity scores and frequency of Team Science strategy use values were plotted on a four-quadrant graph to demonstrate the effectiveness of the training on perceived familiarity and frequency of use regarding the core competencies. These visualizations are organized with the pre-scores on the X-axis and the post-scores on the Y-axis. Low pre- and post-program scores would indicate that more Team Science training is needed, while low pre- and high post-program scores suggest that the workshops substantially improved team science capability. High pre- and post-program scores suggest that information and skill development may have been overkill, as participants were already proficient, as indicated by their self-reports. Finally, if the pre-program scores were high, but the post-program scores were low, it could suggest that the training program was ineffective, or participants initially overestimated their ability before the program (see Figure 1).

Figure 1. Mean pre- and post-program 5-item likert scores for variables related to team planning (Familiarity), managing a team (Familiarity), and interpersonal relationship (Frequency) skills.
Impact of team science training by competency domains
The four-quadrant graph (see Figure 1) shows that the training achieved “big wins” for the Team Planning and Managing Team competency domains. Familiarity with Team Planning skills showed the greatest gains from pre- to post-program, with the mean shifting from 1 to 2 total points (see Table 4). Four items in the Team Planning component (shared language, shared vision, creating ground rules, and navigating change) scored below 3 in the pre-program survey and over 4 in the post-program survey (see Table 4 for mean scores). The fifth item, creating team charters, showed the greatest change, where participants initially indicated little familiarity (1.9) and reported substantially stronger familiarity (3.9) after the workshop series.
Table 4. Mean pre- and post-program scores with standard deviations

Note: all values are rated on a five-point Likert scale ranging from strongly disagree (1) to strongly agree (5) to measure familiarity (Team Planning, Managing a Team) or frequency of using strategies when working on a research project (1) never to (5) always (and Interpersonal Relations) pre- and post-program.
All four items in the Managing a Team competency (creating meeting agendas, monitoring progress, establishing authorship agreements, incorporating others’ perspectives) were also in the “big wins” quadrant. The items associated with this component had higher pre-program scores than the Team Planning component. Overall, these items showed moderate pre- and high post-program scores as a cluster and suggested that skills associated with Managing a Team can be effectively taught in a short virtual workshop setting. These results show “big wins” as a cluster for both the Team Planning and Managing a Team components, suggesting that the team science training effectively increased collaborative capacity across these two core competency domains.
The items in the Interpersonal Relations component showed more moderate increases, ranging from 0.64 to 1.1, compared to the more extensive growth or final means above 4.0 in the Team Planning and Managing a Team components. Participants’ pre-program results in this competency range from 2.66 to 3.23, indicating an opportunity for skill development through targeted training. While scores on all items increased in the post-program assessment (p values <0.05), the average score for all items remained slightly below the optimal outcome (see Table 4).
Conclusions
Results indicate that the CCTSI Team Science training program significantly (p =< 0.05) increased the self-reported teaming competency of participants. The three distinct, empirically confirmed components – Team Planning, Managing a Team, and Interpersonal Relations – from the TEAMS survey instrument presents a novel contribution to team science training, evaluation, and team science literature. While prior research has emphasized the importance of communication, coordination, and trust within teams, these specific groupings and their operationalization through validated survey items have not been formalized. Using PCA with strong reliability coefficients (α = 0.84, 0.74, and 0.81, respectively), this study provides initial evidence of construct validity for evaluating team science competencies. Furthermore, pre- and post-program results (Table 4) show substantial gains across all three domains, particularly in Team Planning, where familiarity with practices such as creating team charters and establishing ground rules nearly doubled. These findings suggest an opportunity for other CTSA hubs and interdisciplinary research training programs to adopt this structure as a framework for evaluating team science initiatives. By incorporating clear domains and validated metrics, programs can more systematically assess where participants are starting from, how their familiarity with and use of practices evolve, and where additional support is needed. The approach also fills a gap in the current evaluation landscape by offering a flexible, evidence-based tool for measuring familiarity and frequency of use for essential team science strategies.
In the assessment of the Team Science Training program, we found that the Team Planning and Managing a Team competencies may have achieved the most substantial self-reported gains because they involve structured, explicit knowledge and tangible activities, such as authorship agreements and meeting agendas, that are easier to implement in practice. Participants also reported statistically significant improvements in the development of Interpersonal Relationship competency. However, the mean scores did not reach the desired application levels, as measured by the frequency of “most of the time” application on research projects. This finding suggests that more and/or different training may be needed to enhance Interpersonal Relationship skills, such as increased role-playing activities or exercises to practice these skills in real-life situations. Considering these findings, we recommend that future training programs for teaming assess explicit and tacit collaboration skills, which are often overlooked in traditional evaluations. These programs should consider developing context- and team-specific interventions and training methods tailored to the unique dynamics of each team [Reference Linhardt, Bisbey and Salas10]. By doing so, organizations can better equip teams with the visible and underlying skills necessary for effective collaboration.
This study highlights that participant significantly increased their self-assessed Interpersonal Relationship skills, there is still more room for growth in these competencies. Further development may require greater emphasis and integration of learning tools that deepen interpersonal skills and promote self-reflection. Future training efforts may benefit from involving the entire team in practicing interpersonal skills in a more context-specific setting or incorporating more relationship-building activities, whether virtual or in-person [Reference Lacerenza, Marlow, Tannenbaum and Salas24]. Team Science training should actively engage with cross-cultural differences and multi-layered identities. It could benefit from considering established practices in intercultural communication training to enhance collaborative capacity and foster an inclusive environment [Reference Feitosa, Hagenbuch, Patel and Davis25,Reference Perry, Mutignani and Gette26]. To increase innovation in research teams by broadening participation, Team Science program instructors must acknowledge and create a safe space in workshops to discuss personal experiences, identify targeted strategies to foster communication across interdisciplinary differences, and explore cultural differences [Reference Feitosa, Hagenbuch, Patel and Davis25,Reference O’Rourke, Rinkus, Cardenas, McLeskey and Gosselin27]. Sharing experiences can be validating for those who have faced similar encounters and raise awareness for those who haven’t. Furthermore, Team Science program instructors should collect and share the experiences and strategies of individuals with similar demographics who have overcome uncollaborative settings to empower self-efficacy in participants. In addition, scientists on teams should familiarize themselves with both on-campus resources, such as academic advisors, university counseling services, and research offices, as well as off-campus resources, including professional organizations, mental health professionals, or specialized consultants, in case a teaming challenge arises that could benefit from professional assistance. Finally, the faculty of Team Science training programs should consider tailoring the program with examples relevant to specific demographics to acknowledge persistent inequalities, create a safer space, foster social support, and promote solidarity.
Limitations and directions for future research
This study has several important limitations that should be considered when interpreting and applying the findings. First, the post-program survey responses were lower (n = 47) than the pre-program survey response rate (n = 117), which could introduce response bias. The high attrition rate underscores the need for future studies to explore strategies for improving participant retention. Additionally, using self-reported measures, while helpful in capturing participant perceptions, may be influenced by cognitive biases, where individuals either overestimate or underestimate their familiarity and frequency of use. Despite these limitations, the results suggest that participants found value in Team Science training. Self-reported improvements were observed in key teaming aspects, such as Team Planning, Managing a Team, and Interpersonal Relationship skills. The increase in participants’ perceived teaming competencies suggests that these general skills can be reinforced through active learning workshops focused on skills identified in the literature [Reference Bisbey, Wooten, Campo, Lant and Salas5,Reference Lotrecchiano, DiazGranados and Sprecher12].
Further research is needed to explore how phenomena like the Matthew Effect – where individuals or groups with existing advantages, such as prior leadership experience or access to resources, tend to gain even more recognition and opportunities – might influence self-reported outcomes in team science training programs. This is particularly important for participants from historically underrepresented groups, who may face unique structural challenges, such as limited access to resources, which can affect both their experiences in the program and their perception of progress. Future studies might consider how cognitive biases, such as the Dunning-Kruger Effect, where individuals may overestimate or underestimate their skills, might impact participants’ self-assessments and inform the development of new measures for Team Science competencies that don’t rely exclusively on self-assessment. To ensure that future findings are more robust and generalizable, additional research should prioritize recruitment and retention strategies that enhance the diversity and representation of participants throughout the study. Larger and more diverse cohorts of early-career researchers and clinical research professionals, combined with behavioral observations, are crucial for validating these findings and further investigating how these biases and structural inequities impact outcomes.
Acknowledgments
We would like to thank Lauren Smith for her assistance with various aspects of the CCTSI Team Science Training Program and for proofreading this manuscript.
Author contributions
Anne Mook: Conceptualization, Formal analysis, Methodology, Project administration, Supervision, Visualization, Writing-original draft, Writing-review and editing; Verena Knerich: Conceptualization, Visualization, Writing-original draft, Writing-review and editing; Goldie Komaie: Investigation, Writing-review and editing; Lisa Cicutto: Conceptualization, Funding acquisition, Writing-review and editing; Jeni Cross: Conceptualization, Funding acquisition, Methodology, Project administration, Supervision, Writing-review and editing.
Funding statement
This study is supported by NIH/NCATS Colorado CTSA Grant Number UM1 TR004399 and UL1 TR002535. Contents are the authors’ sole responsibility and do not necessarily represent official NIH views.
Competing interests
There are no conflicts of interest.