To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Medicare claims are frequently used to study Clostridioides difficile infection (CDI) epidemiology. However, they lack specimen collection and diagnosis dates to assign location of onset. Algorithms to classify CDI onset location using claims data have been published, but the degree of misclassification is unknown.
Methods:
We linked patients with laboratory-confirmed CDI reported to four Emerging Infections Program (EIP) sites from 2016–2021 to Medicare beneficiaries with fee-for-service Part A/B coverage. We calculated sensitivity of ICD-10-CM codes in claims within ±28 days of EIP specimen collection. CDI was categorized as hospital, long-term care facility, or community-onset using three different Medicare claims-based algorithms based on claim type, ICD-10-CM code position, duration of hospitalization, and ICD-10-CM diagnosis code presence-on-admission indicators. We assessed concordance of EIP case classifications, based on chart review and specimen collection date, with claims case classifications using Cohen’s kappa statistic.
Results:
Of 12,671 CDI cases eligible for linkage, 9,032 (71%) were linked to a single, unique Medicare beneficiary. Compared to EIP, sensitivity of CDI ICD-10-CM codes was 81%; codes were more likely to be present for hospitalized patients (93.0%) than those who were not (56.2%). Concordance between EIP and Medicare claims algorithms ranged from 68% to 75%, depending on the algorithm used (κ = 0.56–0.66).
Conclusion:
ICD-10-CM codes in Medicare claims data had high sensitivity compared to laboratory-confirmed CDI reported to EIP. Claims-based epidemiologic classification algorithms had moderate concordance with EIP classification of onset location. Misclassification of CDI onset location using Medicare algorithms may bias findings of claims-based CDI studies.
The incubation period for Clostridioides difficile infection (CDI) is generally considered to be less than 1 week, but some recent studies suggest that prolonged carriage prior to disease onset may be common.
Objective:
To estimate the incubation period for patients developing CDI after initial negative cultures.
Methods:
In 3 tertiary care medical centers, we conducted a cohort study to identify hospitalized patients and long-term care facility residents with negative initial cultures for C. difficile followed by a diagnosis of CDI with or without prior detection of carriage. Cases were classified as healthcare facility-onset, community-onset, healthcare facility-associated, or community-associated and were further classified as probable, possible, or unlikely CDI. A parametric accelerated failure time model was used to estimate the distribution of the incubation period.
Results:
Of 4,179 patients with negative enrollment cultures and no prior CDI diagnosis within 56 days, 107 (2.6%) were diagnosed as having CDI, including 19 (17.8%) with and 88 (82.2%) without prior detection of carriage. When the data were censored to only include participants with negative cultures collected within 14 days, the estimated median incubation period was 6 days with 25% and 75% of estimated incubation periods occurring within 3 and 12 days, respectively. The observed estimated incubation period did not differ significantly for patients classified as probable, possible, or unlikely CDI.
Conclusion:
Our findings are consistent with the previous studies that suggested the incubation period for CDI is typically less than 1 week and is less than 2 weeks in most cases.
Why is parenting in adolescence predictive of maladaptive personality in adulthood? This study sets out to investigate environmental and genetic factors underlying the association between parenting and maladaptive personality longitudinally in a large sample of twins. The present study addressed this question via a longitudinal study focused on two cohorts of twins assessed on aspects of perceived parenting (parent- and adolescent-reported) at age 14 years (n =1,094 pairs). Participants were followed to adulthood, and maladaptive personality traits were self-reported using the Personality Inventory for DSM-5 (PID-5) at age 24 or 34 years. We then modeled these data using a bivariate biometric model, decomposing parenting-maladaptive personality associations into additive genetic, shared environmental, and nonshared environmental factors. Numerous domains of adolescent-reported parenting predicted adult maladaptive personality. Further, we found evidence for substantial additive genetic (ra ranging from 0.22 to 0.55) and (to a lesser extent) nonshared environmental factors (re ranging from 0.10 to 0.15) that accounted for the association between perceived parenting reported in adolescence and adult personality. Perceived parenting in adolescence and maladaptive personality in adulthood may be related due to some of the same genetic factors contributing to both phenotypes at different developmental periods.
Background: Medicare claims are frequently used to study Clostridioides difficile infection (CDI) epidemiology. Categorizing CDI based on location of onset and potential exposure is critical in understanding transmission patterns and prevention strategies. While claims data are well-suited for identifying prior healthcare utilization exposures, they lack specimen collection and diagnosis dates to assign likely location of onset. Algorithms to classify CDI onset and healthcare association using claims data have been published, but the degree of misclassification is unknown. Methods: We linked patients with laboratory-confirmed CDI reported to four Emerging Infections Program (EIP) sites from 2016-2020 to Medicare beneficiaries using residence, birth date, sex, and hospitalization and/or healthcare exposure dates. Uniquely linked patients with fee-for-service Medicare A/B coverage and complete EIP case report forms were included. Patients with a claims CDI diagnosis code within ±28 days of a positive CDI test reported to EIP were categorized as hospital-onset (HO), long-term care facility onset (LTCFO), or community-onset (CO, either healthcare facility-associated [COHCFA] or community-associated [CA]) using a previously published algorithm based on claim type, ICD-10-CM code position, and duration of hospitalization (if applicable). EIP classifies CDI into these categories using positive specimen collection date and other information from chart review (e.g. admit/discharge dates). We assessed concordance of EIP and claims case classifications using Cohen’s kappa. Results: Of 10,002 eligible EIP-identified CDI cases, 7,064 were linked to a unique beneficiary; 3,451 met Medicare A/B fee-for-service coverage inclusion criteria. Of these, 650 (19%) did not have a claims diagnosis code ±28 days of the EIP specimen collection date (Table); 48% (313/650) of those without a claims diagnosis code were categorized by EIP as CA CDI. Among those with a CDI diagnosis code, concurrence of claims-based and EIP CDI classification was 68% (κ=0.56). Concurrence was highest for HO and lowest for COHCFA CDI. A substantial number of EIP-classified CO CDIs (30%, Figure) were misclassified as HO using the claims-based algorithm; half of these had a primary ICD-10 diagnosis code of sepsis (226/454; 50%). Conclusions: Evidence of CDI in claims data was found for 81% of EIP-reported CDI cases. Medicare classification algorithms concurred with the EIP classification in 68% of cases. Discordance was most common for community-onset CDI patients, many of whom were hospitalized with a primary diagnosis of sepsis. Misclassification of CO-CDI as HO may bias findings of claims-based CDI studies.
Understanding characteristics of healthcare personnel (HCP) with SARS-CoV-2 infection supports the development and prioritization of interventions to protect this important workforce. We report detailed characteristics of HCP who tested positive for SARS-CoV-2 from April 20, 2020 through December 31, 2021.
Methods:
CDC collaborated with Emerging Infections Program sites in 10 states to interview HCP with SARS-CoV-2 infection (case-HCP) about their demographics, underlying medical conditions, healthcare roles, exposures, personal protective equipment (PPE) use, and COVID-19 vaccination status. We grouped case-HCP by healthcare role. To describe residential social vulnerability, we merged geocoded HCP residential addresses with CDC/ATSDR Social Vulnerability Index (SVI) values at the census tract level. We defined highest and lowest SVI quartiles as high and low social vulnerability, respectively.
Results:
Our analysis included 7,531 case-HCP. Most case-HCP with roles as certified nursing assistant (CNA) (444, 61.3%), medical assistant (252, 65.3%), or home healthcare worker (HHW) (225, 59.5%) reported their race and ethnicity as either non-Hispanic Black or Hispanic. More than one third of HHWs (166, 45.2%), CNAs (283, 41.7%), and medical assistants (138, 37.9%) reported a residential address in the high social vulnerability category. The proportion of case-HCP who reported using recommended PPE at all times when caring for patients with COVID-19 was lowest among HHWs compared with other roles.
Conclusions:
To mitigate SARS-CoV-2 infection risk in healthcare settings, infection prevention, and control interventions should be specific to HCP roles and educational backgrounds. Additional interventions are needed to address high social vulnerability among HHWs, CNAs, and medical assistants.
Clinical outcomes of repetitive transcranial magnetic stimulation (rTMS) for treatment of treatment-resistant depression (TRD) vary widely and there is no mood rating scale that is standard for assessing rTMS outcome. It remains unclear whether TMS is as efficacious in older adults with late-life depression (LLD) compared to younger adults with major depressive disorder (MDD). This study examined the effect of age on outcomes of rTMS treatment of adults with TRD. Self-report and observer mood ratings were measured weekly in 687 subjects ages 16–100 years undergoing rTMS treatment using the Inventory of Depressive Symptomatology 30-item Self-Report (IDS-SR), Patient Health Questionnaire 9-item (PHQ), Profile of Mood States 30-item, and Hamilton Depression Rating Scale 17-item (HDRS). All rating scales detected significant improvement with treatment; response and remission rates varied by scale but not by age (response/remission ≥ 60: 38%–57%/25%–33%; <60: 32%–49%/18%–25%). Proportional hazards models showed early improvement predicted later improvement across ages, though early improvements in PHQ and HDRS were more predictive of remission in those < 60 years (relative to those ≥ 60) and greater baseline IDS burden was more predictive of non-remission in those ≥ 60 years (relative to those < 60). These results indicate there is no significant effect of age on treatment outcomes in rTMS for TRD, though rating instruments may differ in assessment of symptom burden between younger and older adults during treatment.
Patients tested for Clostridioides difficile infection (CDI) using a 2-step algorithm with a nucleic acid amplification test (NAAT) followed by toxin assay are not reported to the National Healthcare Safety Network as a laboratory-identified CDI event if they are NAAT positive (+)/toxin negative (−). We compared NAAT+/toxin− and NAAT+/toxin+ patients and identified factors associated with CDI treatment among NAAT+/toxin− patients.
Design:
Retrospective observational study.
Setting:
The study was conducted across 36 laboratories at 5 Emerging Infections Program sites.
Patients:
We defined a CDI case as a positive test detected by this 2-step algorithm during 2018–2020 in a patient aged ≥1 year with no positive test in the previous 8 weeks.
Methods:
We used multivariable logistic regression to compare CDI-related complications and recurrence between NAAT+/toxin− and NAAT+/toxin+ cases. We used a mixed-effects logistic model to identify factors associated with treatment in NAAT+/toxin− cases.
Results:
Of 1,801 cases, 1,252 were NAAT+/toxin−, and 549 were NAAT+/toxin+. CDI treatment was given to 866 (71.5%) of 1,212 NAAT+/toxin− cases versus 510 (95.9%) of 532 NAAT+/toxin+ cases (P < .0001). NAAT+/toxin− status was protective for recurrence (adjusted odds ratio [aOR], 0.65; 95% CI, 0.55–0.77) but not CDI-related complications (aOR, 1.05; 95% CI, 0.87–1.28). Among NAAT+/toxin− cases, white blood cell count ≥15,000/µL (aOR, 1.87; 95% CI, 1.28–2.74), ≥3 unformed stools for ≥1 day (aOR, 1.90; 95% CI, 1.40–2.59), and diagnosis by a laboratory that provided no or neutral interpretive comments (aOR, 3.23; 95% CI, 2.23–4.68) were predictors of CDI treatment.
Conclusion:
Use of this 2-step algorithm likely results in underreporting of some NAAT+/toxin− cases with clinically relevant CDI. Disease severity and laboratory interpretive comments influence treatment decisions for NAAT+/toxin− cases.
Although knowledge of their fossil record continues to improve, multituberculates nonetheless remain one of the more poorly understood mammalian clades, which can be attributed to a record comprised of isolated teeth and fragmentary jaws. Fortunately, the p4 of multituberculates is the most common form of remains for this group and is a principal source of diagnostic characters in systematic studies, the p4 of cimolodontan multituberculates is both common and a source of diagnostic characters in systematic studies. The results of a recent morphometric study on the neoplagiaulacid Mesodma suggest that p4 size may be more useful than shape in diagnosing the various species referred to this genus. We tested this hypothesis by applying two different morphometric methods (2D geometric morphometrics and linear measurements) to two samples: (1) one including the p4s of four known species of Mesodma (M. ambigua, M. thompsoni, M. formosa, and M. pygmaea), and (2) a sample of unidentified p4s of Mesodma from the Bug Creek Anthills locality of northeastern Montana. Our results indicate that while form explains most of the morphological variation in p4s of the various species of Mesodma, linear-measurement data support differences in p4 morphology that are not recovered by form data alone. Depending on the methods used, we found evidence for the presence of one or more species of Mesodma in the Bug Creek Anthills fauna. Although shape and size both contribute to morphological variation in the p4 of Mesodma, our results suggest that the diagnostic power of each varies with the type of methodology employed.
This paper introduces a generalization of the $dd^c$-condition for complex manifolds. Like the $dd^c$-condition, it admits a diverse collection of characterizations, and is hereditary under various geometric constructions. Most notably, it is an open property with respect to small deformations. The condition is satisfied by a wide range of complex manifolds, including all compact complex surfaces, and all compact Vaisman manifolds. We show there are computable invariants of a real homotopy type which in many cases prohibit it from containing any complex manifold satisfying such $dd^c$-type conditions in low degrees. This gives rise to numerous examples of almost complex manifolds which cannot be homotopy equivalent to any of these complex manifolds.
Background: Healthcare facilities have experienced many challenges during the COVID-19 pandemic, including limited personal protective equipment (PPE) supplies. Healthcare personnel (HCP) rely on PPE, vaccines, and other infection control measures to prevent SARS-CoV-2 infections. We describe PPE concerns reported by HCP who had close contact with COVID-19 patients in the workplace and tested positive for SARS-CoV-2. Method: The CDC collaborated with Emerging Infections Program (EIP) sites in 10 states to conduct surveillance for SARS-CoV-2 infections in HCP. EIP staff interviewed HCP with positive SARS-CoV-2 viral tests (ie, cases) to collect data on demographics, healthcare roles, exposures, PPE use, and concerns about their PPE use during COVID-19 patient care in the 14 days before the HCP’s SARS-CoV-2 positive test. PPE concerns were qualitatively coded as being related to supply (eg, low quality, shortages); use (eg, extended use, reuse, lack of fit test); or facility policy (eg, lack of guidance). We calculated and compared the percentages of cases reporting each concern type during the initial phase of the pandemic (April–May 2020), during the first US peak of daily COVID-19 cases (June–August 2020), and during the second US peak (September 2020–January 2021). We compared percentages using mid-P or Fisher exact tests (α = 0.05). Results: Among 1,998 HCP cases occurring during April 2020–January 2021 who had close contact with COVID-19 patients, 613 (30.7%) reported ≥1 PPE concern (Table 1). The percentage of cases reporting supply or use concerns was higher during the first peak period than the second peak period (supply concerns: 12.5% vs 7.5%; use concerns: 25.5% vs 18.2%; p Conclusions: Although lower percentages of HCP cases overall reported PPE concerns after the first US peak, our results highlight the importance of developing capacity to produce and distribute PPE during times of increased demand. The difference we observed among selected groups of cases may indicate that PPE access and use were more challenging for some, such as nonphysicians and nursing home HCP. These findings underscore the need to ensure that PPE is accessible and used correctly by HCP for whom use is recommended.
Recent well-powered genome-wide association studies have enhanced prediction of substance use outcomes via polygenic scores (PGSs). Here, we test (1) whether these scores contribute to prediction over-and-above family history, (2) the extent to which PGS prediction reflects inherited genetic variation v. demography (population stratification and assortative mating) and indirect genetic effects of parents (genetic nurture), and (3) whether PGS prediction is mediated by behavioral disinhibition prior to substance use onset.
Methods
PGSs for alcohol, cannabis, and nicotine use/use disorder were calculated for Minnesota Twin Family Study participants (N = 2483, 1565 monozygotic/918 dizygotic). Twins' parents were assessed for histories of substance use disorder. Twins were assessed for behavioral disinhibition at age 11 and substance use from ages 14 to 24. PGS prediction of substance use was examined using linear mixed-effects, within-twin pair, and structural equation models.
Results
Nearly all PGS measures were associated with multiple types of substance use independently of family history. However, most within-pair PGS prediction estimates were substantially smaller than the corresponding between-pair estimates, suggesting that prediction is driven in part by demography and indirect genetic effects of parents. Path analyses indicated the effects of both PGSs and family history on substance use were mediated via disinhibition in preadolescence.
Conclusions
PGSs capturing risk of substance use and use disorder can be combined with family history measures to augment prediction of substance use outcomes. Results highlight indirect sources of genetic associations and preadolescent elevations in behavioral disinhibition as two routes through which these scores may relate to substance use.
Major depressive disorder (MDD) and chronic pain are highly comorbid, and pain symptoms are associated with a poorer response to antidepressant medication treatment. It is unclear whether comorbid pain also is associated with a poorer response to treatment with repetitive transcranial magnetic stimulation (rTMS).
Methods
162 MDD subjects received 30 sessions of 10 Hz rTMS treatment administered to the left dorsolateral prefrontal cortex (DLPFC) with depression and pain symptoms measured before and after treatment. For a subset of 96 patients, a resting-state electroencephalogram (EEG) was recorded at baseline. Clinical outcome was compared between subjects with and without comorbid pain, and the relationships among outcome, pain severity, individual peak alpha frequency (PAF), and PAF phase-coherence in the EEG were examined.
Results
64.8% of all subjects reported pain, and both depressive and pain symptoms were significantly reduced after rTMS treatment, irrespective of age or gender. Patients with severe pain were 27% less likely to respond to MDD treatment than pain-free individuals. PAF was positively associated with pain severity. PAF phase-coherence in the somatosensory and default mode networks was significantly lower for MDD subjects with pain who failed to respond to MDD treatment.
Conclusions
Pain symptoms improved after rTMS to left DLPFC in MDD irrespective of age or gender, although the presence of chronic pain symptoms reduced the likelihood of treatment response. Individual PAF and baseline phase-coherence in the sensorimotor and midline regions may represent predictors of rTMS treatment outcome in comorbid pain and MDD.
Around 40% of people with bipolar disorder (BD) are non-adherent to medication leading to relapse, hospitalisation and increased suicide risk. Limited progress in addressing non-adherence may be partly attributable to insufficient understanding of the modifiable determinants of adherence that require targeting in interventions. We synthesised the modifiable determinants of adherence in BD and map them to the theoretical domains framework (TDF).
Method
We searched CINAHL, Cochrane Library, Embase, LILACS, Medline, PsychINFO and PubMed until February 2020. We included studies reporting modifiable determinants of adherence in BD. Two reviewers independently screened studies, assessed quality, extracted modifiable determinants and mapped them to TDF.
Results
We included 57 studies involving 32 894 participants. Determinants reported by patients spanned 11 of the 14 TDF domains compared to six domains represented by clinician/researcher. The TDF domains most commonly represented (% and example) in studies were: ‘Environmental context and resources’ (63%, e.g. experiencing side effects), ‘Beliefs about consequences’ (63%, e.g. beliefs about medication effects), ‘Knowledge’ (40%, e.g. knowledge about disorder), ‘Social influences’ (33%, e.g. support from family/clinicians), ‘Memory, attention and decision processes’ (33%, e.g. forgetfulness), ‘Emotion’ (21%, e.g. fear of addiction) and ‘Intentions’ (21%, e.g. wanting alternative treatment). ‘Intentions’, ‘Memory, attention and decision processes’ and ‘Emotion’ domains were only reported by patients but not clinicians.
Conclusions
Clinicians may be underappreciating the full range of modifiable determinants of adherence and thus not providing adherence support reflective of patients' needs. Reporting of modifiable determinants in behavioural terms facilitates developing theory-based interventions to address non-adherence in BD.
To assess the potential for contamination of personnel, patients, and the environment during use of contaminated N95 respirators and to compare the effectiveness of interventions to reduce contamination.
Design:
Simulation study of patient care interactions using N95 respirators contaminated with a higher and lower inocula of the benign virus bacteriophage MS2.
Methods:
In total, 12 healthcare personnel performed 3 standardized examinations of mannequins including (1) control with suboptimal respirator handling technique, (2) improved technique with glove change after each N95 contact, and (3) control with 1-minute ultraviolet-C light (UV-C) treatment prior to donning. The order of the examinations was randomized within each subject. The frequencies of contamination were compared among groups. Observations and simulations with fluorescent lotion were used to assess routes of transfer leading to contamination.
Results:
With suboptimal respirator handling technique, bacteriophage MS2 was frequently transferred to the participants, mannequin, and environmental surfaces and fomites. Improved technique resulted in significantly reduced transfer of MS2 in the higher inoculum simulations (P < .01), whereas UV-C treatment reduced transfer in both the higher- and lower-inoculum simulations (P < .01). Observations and simulations with fluorescent lotion demonstrated multiple potential routes of transfer to participants, mannequin, and surfaces, including both direct contact with the contaminated respirator and indirect contact via contaminated gloves.
Conclusion:
Reuse of contaminated N95 respirators can result in contamination of personnel and the environment even when correct technique is used. Decontamination technologies, such as UV-C, could reduce the risk for transmission.
Background: Certain nursing home (NH) resident care tasks have a higher risk for multidrug-resistant organisms (MDRO) transfer to healthcare personnel (HCP), which can result in transmission to residents if HCPs fail to perform recommended infection prevention practices. However, data on HCP-resident interactions are limited and do not account for intrafacility practice variation. Understanding differences in interactions, by HCP role and unit, is important for informing MDRO prevention strategies in NHs. Methods: In 2019, we conducted serial intercept interviews; each HCP was interviewed 6–7 times for the duration of a unit’s dayshift at 20 NHs in 7 states. The next day, staff on a second unit within the facility were interviewed during the dayshift. HCP on 38 units were interviewed to identify healthcare personnel (HCP)–resident care patterns. All unit staff were eligible for interviews, including certified nursing assistants (CNAs), nurses, physical or occupational therapists, physicians, midlevel practitioners, and respiratory therapists. HCP were asked to list which residents they had cared for (within resident rooms or common areas) since the prior interview. Respondents selected from 14 care tasks. We classified units into 1 of 4 types: long-term, mixed, short stay or rehabilitation, or ventilator or skilled nursing. Interactions were classified based on the risk of HCP contamination after task performance. We compared proportions of interactions associated with each HCP role and performed clustered linear regression to determine the effect of unit type and HCP role on the number of unique task types performed per interaction. Results: Intercept-interviews described 7,050 interactions and 13,843 care tasks. Except in ventilator or skilled nursing units, CNAs have the greatest proportion of care interactions (interfacility range, 50%–60%) (Fig. 1). In ventilator and skilled nursing units, interactions are evenly shared between CNAs and nurses (43% and 47%, respectively). On average, CNAs in ventilator and skilled nursing units perform the most unique task types (2.5 task types per interaction, Fig. 2) compared to other unit types (P < .05). Compared to CNAs, most other HCP types had significantly fewer task types (0.6–1.4 task types per interaction, P < .001). Across all facilities, 45.6% of interactions included tasks that were higher-risk for HCP contamination (eg, transferring, wound and device care, Fig. 3). Conclusions: Focusing infection prevention education efforts on CNAs may be most efficient for preventing MDRO transmission within NH because CNAs have the most HCP–resident interactions and complete more tasks per visit. Studies of HCP-resident interactions are critical to improving understanding of transmission mechanisms as well as target MDRO prevention interventions.
Funding: Centers for Disease Control and Prevention (grant no. U01CK000555-01-00)
Disclosures:Scott Fridkin, consulting fee, vaccine industry (spouse)
Background: Effective medical device reprocessing (MDR) is essential in preventing the spread of microorganisms and maintaining patient safety. Alberta Health Services (AHS) is an Alberta-wide, integrated health system, responsible for delivering health services to >4.3 million people living in the province. In 2010, periodic province-wide MDR reviews were initiated by the provincial health system to verify that the cleaning, disinfection, and sterilization of reusable critical and semicritical medical devices met established standards. To date, there have been 3 review cycles; in cycle 3, a follow-up process for tracking and reporting corrective actions was initiated. Methods: As in previous MDR review cycles, cycle 3 included the use of a standardized suite of tools to measure compliance with standards set by Accreditation Canada, the Canadian Standards Association, and the Government of Alberta. Each cycle involved a review of MDR areas completed by trained reviewers. Interrater reliability among reviewers was maintained through training and debriefings following reviews to ensure agreement. Following reviews, reports were generated for areas, zones, and AHS. As part of the corrective actions and follow-up process, identified deficiencies were categorized into 5 themes. Corrective actions were tracked and periodic reports were generated showing the progress of deficiency resolution. Resolution rates (number of resolved deficiencies divided by total number of of deficiencies) were calculated for each of the identified themes as well as overall for cycle 3. Results: Overall compliance for cycle 3 was 93%. Cycle 3 reviews revealed that more than half of the deficiencies (58%) were identified previously in cycle 2. The resolution rates ranged from 78% to 95% for identified deficiencies for 4 of the 5 themes: documentation, technique, PPE/attire/hand hygiene, and other. The theme related to physical infrastructure showed a considerably lower resolution rate of 49%. The corrective action follow-up process showed increased overall resolution rate from 59% at the start of the follow-up process to 82% at its completion. When this resolution rate was applied to the initial survey compliance rate for cycle 3, overall compliance increased to 99%. Conclusions: Monitoring quality of MDR practices is essential in maintaining and improving patient safety. The standardized provincial review process identified common themes and a coordinated approach to support the resolution of many identified deficiencies. Most of those deficiencies were resolved; however, those deficiencies related to physical infrastructure of the MDR department continue to be seen across review cycles. This review process with follow up of these deficiencies can help bring attention to organization leadership and Funding: authorities during budget cycles.
The Minnesota Center for Twin and Family Research (MCTFR) comprises multiple longitudinal, community-representative investigations of twin and adoptive families that focus on psychological adjustment, personality, cognitive ability and brain function, with a special emphasis on substance use and related psychopathology. The MCTFR includes the Minnesota Twin Registry (MTR), a cohort of twins who have completed assessments in middle and older adulthood; the Minnesota Twin Family Study (MTFS) of twins assessed from childhood and adolescence into middle adulthood; the Enrichment Study (ES) of twins oversampled for high risk for substance-use disorders assessed from childhood into young adulthood; the Adolescent Brain (AdBrain) study, a neuroimaging study of adolescent twins; and the Siblings Interaction and Behavior Study (SIBS), a study of adoptive and nonadoptive families assessed from adolescence into young adulthood. Here we provide a brief overview of key features of these established studies and describe new MCTFR investigations that follow up and expand upon existing studies or recruit and assess new samples, including the MTR Study of Relationships, Personality, and Health (MTR-RPH); the Colorado-Minnesota (COMN) Marijuana Study; the Adolescent Brain Cognitive Development (ABCD) study; the Colorado Online Twins (CoTwins) study and the Children of Twins (CoT) study.
To assess the impact of a newly developed Central-Line Insertion Site Assessment (CLISA) score on the incidence of local inflammation or infection for CLABSI prevention.
Design:
A pre- and postintervention, quasi-experimental quality improvement study.
Setting and participants:
Adult inpatients with central venous catheters (CVCs) hospitalized in an intensive care unit or oncology ward at a large academic medical center.
Methods:
We evaluated CLISA score impact on insertion site inflammation and infection (CLISA score of 2 or 3) incidence in the baseline period (June 2014–January 2015) and the intervention period (April 2015–October 2017) using interrupted times series and generalized linear mixed-effects multivariable analyses. These were run separately for days-to-line removal from identification of a CLISA score of 2 or 3. CLISA score interrater reliability and photo quiz results were evaluated.
Results:
Among 6,957 CVCs assessed 40,846 times, percentage of lines with CLISA score of 2 or 3 in the baseline and intervention periods decreased by 78.2% (from 22.0% to 4.7%), with a significant immediate decrease in the time-series analysis (P < .001). According to the multivariable regression, the intervention was associated with lower percentage of lines with a CLISA score of 2 or 3, after adjusting for age, gender, CVC body location, and hospital unit (odds ratio, 0.15; 95% confidence interval, 0.06–0.34; P < .001). According to the multivariate regression, days to removal of lines with CLISA score of 2 or 3 was 3.19 days faster after the intervention (P < .001). Also, line dwell time decreased 37.1% from a mean of 14 days (standard deviation [SD], 10.6) to 8.8 days (SD, 9.0) (P < .001). Device utilization ratios decreased 9% from 0.64 (SD, 0.08) to 0.58 (SD, 0.06) (P = .039).
Conclusions:
The CLISA score creates a common language for assessing line infection risk and successfully promotes high compliance with best practices in timely line removal.
Consumption of certain berries appears to slow postprandial glucose absorption, attributable to polyphenols, which may benefit exercise and cognition, reduce appetite and/or oxidative stress. This randomised, crossover, placebo-controlled study determined whether polyphenol-rich fruits added to carbohydrate-based foods produce a dose-dependent moderation of postprandial glycaemic, glucoregulatory hormone, appetite and ex vivo oxidative stress responses. Twenty participants (eighteen males/two females; 24 (sd 5) years; BMI: 27 (sd 3) kg/m2) consumed one of five cereal bars (approximately 88 % carbohydrate) containing no fruit ingredients (reference), freeze-dried black raspberries (10 or 20 % total weight; LOW-Rasp and HIGH-Rasp, respectively) and cranberry extract (0·5 or 1 % total weight; LOW-Cran and HIGH-Cran), on trials separated by ≥5 d. Postprandial peak/nadir from baseline (Δmax) and incremental postprandial AUC over 60 and 180 min for glucose and other biochemistries were measured to examine the dose-dependent effects. Glucose AUC0–180 min trended towards being higher (43 %) after HIGH-Rasp v. LOW-Rasp (P=0·06), with no glucose differences between the raspberry and reference bars. Relative to reference, HIGH-Rasp resulted in a 17 % lower Δmax insulin, 3 % lower C-peptide (AUC0–60 min and 3 % lower glucose-dependent insulinotropic polypeptide (AUC0–180 min) P<0·05. No treatment effects were observed for the cranberry bars regarding glucose and glucoregulatory hormones, nor were there any treatment effects for either berry type regarding ex vivo oxidation, appetite-mediating hormones or appetite. Fortification with freeze-dried black raspberries (approximately 25 g, containing 1·2 g of polyphenols) seems to slightly improve the glucoregulatory hormone and glycaemic responses to a high-carbohydrate food item in young adults but did not affect appetite or oxidative stress responses at doses or with methods studied herein.