To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Paleontology provides insights into the history of the planet, from the origins of life billions of years ago to the biotic changes of the Recent. The scope of paleontological research is as vast as it is varied, and the field is constantly evolving. In an effort to identify “Big Questions” in paleontology, experts from around the world came together to build a list of priority questions the field can address in the years ahead. The 89 questions presented herein (grouped within 11 themes) represent contributions from nearly 200 international scientists. These questions touch on common themes including biodiversity drivers and patterns, integrating data types across spatiotemporal scales, applying paleontological data to contemporary biodiversity and climate issues, and effectively utilizing innovative methods and technology for new paleontological insights. In addition to these theoretical questions, discussions touch upon structural concerns within the field, advocating for an increased valuation of specimen-based research, protection of natural heritage sites, and the importance of collections infrastructure, along with a stronger emphasis on human diversity, equity, and inclusion. These questions offer a starting point—an initial nucleus of consensus that paleontologists can expand on—for engaging in discussions, securing funding, advocating for museums, and fostering continued growth in shared research directions.
Describe the hemodynamic implications of anaesthetic choice among children with heart disease undergoing cardiac catheterisation.
Methods:
Study 1 was a secondary analysis of data obtained during catheterisation-based hemodynamic assessment of infants with hypoplastic left heart syndrome following Stage 1 palliation, randomised in the Single Ventricle Reconstruction trial. Measured and calculated hemodynamics including pulmonary and systemic vascular resistance indexed to body surface area (PVRi and SVRi respectively) and pulmonary/systemic blood flow (Qp/Qs) were analysed with respect to anaesthetic employed during catheterisation, classified as moderate sedation or general anaesthesia. Study 2 consisted of a single centre, prospective analysis of patients requiring percutaneous closure of a patent ductus arteriosus or endomyocardial biopsy after orthotopic heart transplant. Participants underwent hemodynamic assessment first using inhaled volatile anaesthesia (IA), and then transitioned to total intravenous anaesthesia, comparing hemodynamic measures with respect to anaesthetic approach.
Results:
In Study 1, independent of shunt type, PVRi, and patient size, moderate sedation was associated with a greater than two-fold odds of a Qp/Qs >1 (OR 2.12, 95%CI 1.18–3.87, p = 0.013). In Study 2, while PVRi was similar, SVRi was significantly higher using total intravenous anaesthesia. Among the patent ductus arteriosus subgroup, Qp/Qs increased significantly with a total intravenous anaesthesia relative to IA (p = 0.003); additionally, among the orthotopic heart transplant subgroup, left ventricular end diastolic pressure increased following a transition to total intravenous anaesthesia (p = 0.002).
Conclusions:
Analyses of hemodynamics during catheterisation support a significant impact of anaesthetic type on hemodynamic values including SVRi, left ventricular end diastolic pressure, and Qp/Qs. Anaesthesia choice and intraprocedural management of SVRi are important considerations when making clinical decisions based on hemodynamic data.
Functional impairment in daily activities, such as work and socializing, is part of the diagnostic criteria for major depressive disorder and most anxiety disorders. Despite evidence that symptom severity and functional impairment are partially distinct, functional impairment is often overlooked. To assess whether functional impairment captures diagnostically relevant genetic liability beyond that of symptoms, we aimed to estimate the heritability of, and genetic correlations between, key measures of current depression symptoms, anxiety symptoms, and functional impairment.
Methods
In 17,130 individuals with lifetime depression or anxiety from the Genetic Links to Anxiety and Depression (GLAD) Study, we analyzed total scores from the Patient Health Questionnaire-9 (depression symptoms), Generalized Anxiety Disorder-7 (anxiety symptoms), and Work and Social Adjustment Scale (functional impairment). Genome-wide association analyses were performed with REGENIE. Heritability was estimated using GCTA-GREML and genetic correlations with bivariate-GREML.
Results
The phenotypic correlations were moderate across the three measures (Pearson’s r = 0.50–0.69). All three scales were found to be under low but significant genetic influence (single-nucleotide polymorphism-based heritability [h2SNP] = 0.11–0.19) with high genetic correlations between them (rg = 0.79–0.87).
Conclusions
Among individuals with lifetime depression or anxiety from the GLAD Study, the genetic variants that underlie symptom severity largely overlap with those influencing functional impairment. This suggests that self-reported functional impairment, while clinically relevant for diagnosis and treatment outcomes, does not reflect substantial additional genetic liability beyond that captured by symptom-based measures of depression or anxiety.
Accurate diagnosis of bipolar disorder (BPD) is difficult in clinical practice, with an average delay between symptom onset and diagnosis of about 7 years. A depressive episode often precedes the first manic episode, making it difficult to distinguish BPD from unipolar major depressive disorder (MDD).
Aims
We use genome-wide association analyses (GWAS) to identify differential genetic factors and to develop predictors based on polygenic risk scores (PRS) that may aid early differential diagnosis.
Method
Based on individual genotypes from case–control cohorts of BPD and MDD shared through the Psychiatric Genomics Consortium, we compile case–case–control cohorts, applying a careful quality control procedure. In a resulting cohort of 51 149 individuals (15 532 BPD patients, 12 920 MDD patients and 22 697 controls), we perform a variety of GWAS and PRS analyses.
Results
Although our GWAS is not well powered to identify genome-wide significant loci, we find significant chip heritability and demonstrate the ability of the resulting PRS to distinguish BPD from MDD, including BPD cases with depressive onset (BPD-D). We replicate our PRS findings in an independent Danish cohort (iPSYCH 2015, N = 25 966). We observe strong genetic correlation between our case–case GWAS and that of case–control BPD.
Conclusions
We find that MDD and BPD, including BPD-D are genetically distinct. Our findings support that controls, MDD and BPD patients primarily lie on a continuum of genetic risk. Future studies with larger and richer samples will likely yield a better understanding of these findings and enable the development of better genetic predictors distinguishing BPD and, importantly, BPD-D from MDD.
To investigate the symptoms of SARS-CoV-2 infection, their dynamics and their discriminatory power for the disease using longitudinally, prospectively collected information reported at the time of their occurrence. We have analysed data from a large phase 3 clinical UK COVID-19 vaccine trial. The alpha variant was the predominant strain. Participants were assessed for SARS-CoV-2 infection via nasal/throat PCR at recruitment, vaccination appointments, and when symptomatic. Statistical techniques were implemented to infer estimates representative of the UK population, accounting for multiple symptomatic episodes associated with one individual. An optimal diagnostic model for SARS-CoV-2 infection was derived. The 4-month prevalence of SARS-CoV-2 was 2.1%; increasing to 19.4% (16.0%–22.7%) in participants reporting loss of appetite and 31.9% (27.1%–36.8%) in those with anosmia/ageusia. The model identified anosmia and/or ageusia, fever, congestion, and cough to be significantly associated with SARS-CoV-2 infection. Symptoms’ dynamics were vastly different in the two groups; after a slow start peaking later and lasting longer in PCR+ participants, whilst exhibiting a consistent decline in PCR- participants, with, on average, fewer than 3 days of symptoms reported. Anosmia/ageusia peaked late in confirmed SARS-CoV-2 infection (day 12), indicating a low discrimination power for early disease diagnosis.
Cognitive training has shown promise for improving cognition in older adults. Aging involves a variety of neuroanatomical changes that may affect response to cognitive training. White matter hyperintensities (WMH) are one common age-related brain change, as evidenced by T2-weighted and Fluid Attenuated Inversion Recovery (FLAIR) MRI. WMH are associated with older age, suggestive of cerebral small vessel disease, and reflect decreased white matter integrity. Higher WMH load associates with reduced threshold for clinical expression of cognitive impairment and dementia. The effects of WMH on response to cognitive training interventions are relatively unknown. The current study assessed (a) proximal cognitive training performance following a 3-month randomized control trial and (b) the contribution of baseline whole-brain WMH load, defined as total lesion volume (TLV), on pre-post proximal training change.
Participants and Methods:
Sixty-two healthy older adults ages 65-84 completed either adaptive cognitive training (CT; n=31) or educational training control (ET; n=31) interventions. Participants assigned to CT completed 20 hours of attention/processing speed training and 20 hours of working memory training delivered through commercially-available Posit Science BrainHQ. ET participants completed 40 hours of educational videos. All participants also underwent sham or active transcranial direct current stimulation (tDCS) as an adjunctive intervention, although not a variable of interest in the current study. Multimodal MRI scans were acquired during the baseline visit. T1- and T2-weighted FLAIR images were processed using the Lesion Segmentation Tool (LST) for SPM12. The Lesion Prediction Algorithm of LST automatically segmented brain tissue and calculated lesion maps. A lesion threshold of 0.30 was applied to calculate TLV. A log transformation was applied to TLV to normalize the distribution of WMH. Repeated-measures analysis of covariance (RM-ANCOVA) assessed pre/post change in proximal composite (Total Training Composite) and sub-composite (Processing Speed Training Composite, Working Memory Training Composite) measures in the CT group compared to their ET counterparts, controlling for age, sex, years of education and tDCS group. Linear regression assessed the effect of TLV on post-intervention proximal composite and sub-composite, controlling for baseline performance, intervention assignment, age, sex, years of education, multisite scanner differences, estimated total intracranial volume, and binarized cardiovascular disease risk.
Results:
RM-ANCOVA revealed two-way group*time interactions such that those assigned cognitive training demonstrated greater improvement on proximal composite (Total Training Composite) and sub-composite (Processing Speed Training Composite, Working Memory Training Composite) measures compared to their ET counterparts. Multiple linear regression showed higher baseline TLV associated with lower pre-post change on Processing Speed Training sub-composite (ß = -0.19, p = 0.04) but not other composite measures.
Conclusions:
These findings demonstrate the utility of cognitive training for improving postintervention proximal performance in older adults. Additionally, pre-post proximal processing speed training change appear to be particularly sensitive to white matter hyperintensity load versus working memory training change. These data suggest that TLV may serve as an important factor for consideration when planning processing speed-based cognitive training interventions for remediation of cognitive decline in older adults.
Interventions using a cognitive training paradigm called the Useful Field of View (UFOV) task have shown to be efficacious in slowing cognitive decline. However, no studies have looked at the engagement of functional networks during UFOV task completion. The current study aimed to (a) assess if regions activated during the UFOV fMRI task were functionally connected and related to task performance (henceforth called the UFOV network), (b) compare connectivity of the UFOV network to 7 resting-state functional connectivity networks in predicting proximal (UFOV) and near-transfer (Double Decision) performance, and (c) explore the impact of network segregation between higher-order networks and UFOV performance.
Participants and Methods:
336 healthy older adults (mean age=71.6) completed the UFOV fMRI task in a Siemens 3T scanner. UFOV fMRI accuracy was calculated as the number of correct responses divided by 56 total trials. Double Decision performance was calculated as the average presentation time of correct responses in log ms, with lower scores equating to better processing speed. Structural and functional MRI images were processed using the default pre-processing pipeline within the CONN toolbox. The Artifact Rejection Toolbox was set at a motion threshold of 0.9mm and participants were excluded if more than 50% of volumes were flagged as outliers. To assess connectivity of regions associated with the UFOV task, we created 10 spherical regions of interest (ROIs) a priori using the WFU PickAtlas in SPM12. These include the bilateral pars triangularis, supplementary motor area, and inferior temporal gyri, as well as the left pars opercularis, left middle occipital gyrus, right precentral gyrus and right superior parietal lobule. We used a weighted ROI-to-ROI connectivity analysis to model task-based within-network functional connectivity of the UFOV network, and its relationship to UFOV accuracy. We then used weighted ROI-to-ROI connectivity analysis to compare the efficacy of the UFOV network versus 7 resting-state networks in predicting UFOV fMRI task performance and Double Decision performance. Finally, we calculated network segregation among higher order resting state networks to assess its relationship with UFOV accuracy. All functional connectivity analyses were corrected at a false discovery threshold (FDR) at p<0.05.
Results:
ROI-to-ROI analysis showed significant within-network functional connectivity among the 10 a priori ROIs (UFOV network) during task completion (all pFDR<.05). After controlling for covariates, greater within-network connectivity of the UFOV network associated with better UFOV fMRI performance (pFDR=.008). Regarding the 7 resting-state networks, greater within-network connectivity of the CON (pFDR<.001) and FPCN (pFDR=. 014) were associated with higher accuracy on the UFOV fMRI task. Furthermore, greater within-network connectivity of only the UFOV network associated with performance on the Double Decision task (pFDR=.034). Finally, we assessed the relationship between higher-order network segregation and UFOV accuracy. After controlling for covariates, no significant relationships between network segregation and UFOV performance remained (all p-uncorrected>0.05).
Conclusions:
To date, this is the first study to assess task-based functional connectivity during completion of the UFOV task. We observed that coherence within 10 a priori ROIs significantly predicted UFOV performance. Additionally, enhanced within-network connectivity of the UFOV network predicted better performance on the Double Decision task, while conventional resting-state networks did not. These findings provide potential targets to optimize efficacy of UFOV interventions.
Nonpathological aging has been linked to decline in both verbal and visuospatial memory abilities in older adults. Disruptions in resting-state functional connectivity within well-characterized, higherorder cognitive brain networks have also been coupled with poorer memory functioning in healthy older adults and in older adults with dementia. However, there is a paucity of research on the association between higherorder functional connectivity and verbal and visuospatial memory performance in the older adult population. The current study examines the association between resting-state functional connectivity within the cingulo-opercular network (CON), frontoparietal control network (FPCN), and default mode network (DMN) and verbal and visuospatial learning and memory in a large sample of healthy older adults. We hypothesized that greater within-network CON and FPCN functional connectivity would be associated with better immediate verbal and visuospatial memory recall. Additionally, we predicted that within-network DMN functional connectivity would be associated with improvements in delayed verbal and visuospatial memory recall. This study helps to glean insight into whether within-network CON, FPCN, or DMN functional connectivity is associated with verbal and visuospatial memory abilities in later life.
Participants and Methods:
330 healthy older adults between 65 and 89 years old (mean age = 71.6 ± 5.2) were recruited at the University of Florida (n = 222) and the University of Arizona (n = 108). Participants underwent resting-state fMRI and completed verbal memory (Hopkins Verbal Learning Test - Revised [HVLT-R]) and visuospatial memory (Brief Visuospatial Memory Test - Revised [BVMT-R]) measures. Immediate (total) and delayed recall scores on the HVLT-R and BVMT-R were calculated using each test manual’s scoring criteria. Learning ratios on the HVLT-R and BVMT-R were quantified by dividing the number of stimuli (verbal or visuospatial) learned between the first and third trials by the number of stimuli not recalled after the first learning trial. CONN Toolbox was used to extract average within-network connectivity values for CON, FPCN, and DMN. Hierarchical regressions were conducted, controlling for sex, race, ethnicity, years of education, number of invalid scans, and scanner site.
Results:
Greater CON connectivity was significantly associated with better HVLT-R immediate (total) recall (ß = 0.16, p = 0.01), HVLT-R learning ratio (ß = 0.16, p = 0.01), BVMT-R immediate (total) recall (ß = 0.14, p = 0.02), and BVMT-R delayed recall performance (ß = 0.15, p = 0.01). Greater FPCN connectivity was associated with better BVMT-R learning ratio (ß = 0.13, p = 0.04). HVLT-R delayed recall performance was not associated with connectivity in any network, and DMN connectivity was not significantly related to any measure.
Conclusions:
Connectivity within CON demonstrated a robust relationship with different components of memory function as well across verbal and visuospatial domains. In contrast, FPCN only evidenced a relationship with visuospatial learning, and DMN was not significantly associated with memory measures. These data suggest that CON may be a valuable target in longitudinal studies of age-related memory changes, but also a possible target in future non-invasive interventions to attenuate memory decline in older adults.
Drawing on early work on ethical leadership, we argue that when leaders engage in leader moral hypocrisy (i.e., ethical promotion without ethical demonstration), followers can experience psychological reactance—a negative response to a perceived restriction of freedom—which can have negative downstream consequences. In a survey of employee–manager dyads (study 1), we demonstrate that leader moral hypocrisy is positively associated with follower psychological reactance, which increases follower deviance. In two subsequent laboratory experiments, we find similar patterns of results (study 2) and explore potential alternative mechanisms (study 3). We demonstrate in a final experiment with working adults that the relationship between leader moral hypocrisy and psychological reactance is partly explained by increased perceptions of a leader’s use of power (study 4). We discuss the implications of our findings and advocate for further understanding of the risks associated with psychological reactance in response to leaders and other workplace situations.
The U.S. Department of Agriculture–Agricultural Research Service (USDA-ARS) has been a leader in weed science research covering topics ranging from the development and use of integrated weed management (IWM) tactics to basic mechanistic studies, including biotic resistance of desirable plant communities and herbicide resistance. ARS weed scientists have worked in agricultural and natural ecosystems, including agronomic and horticultural crops, pastures, forests, wild lands, aquatic habitats, wetlands, and riparian areas. Through strong partnerships with academia, state agencies, private industry, and numerous federal programs, ARS weed scientists have made contributions to discoveries in the newest fields of robotics and genetics, as well as the traditional and fundamental subjects of weed–crop competition and physiology and integration of weed control tactics and practices. Weed science at ARS is often overshadowed by other research topics; thus, few are aware of the long history of ARS weed science and its important contributions. This review is the result of a symposium held at the Weed Science Society of America’s 62nd Annual Meeting in 2022 that included 10 separate presentations in a virtual Weed Science Webinar Series. The overarching themes of management tactics (IWM, biological control, and automation), basic mechanisms (competition, invasive plant genetics, and herbicide resistance), and ecosystem impacts (invasive plant spread, climate change, conservation, and restoration) represent core ARS weed science research that is dynamic and efficacious and has been a significant component of the agency’s national and international efforts. This review highlights current studies and future directions that exemplify the science and collaborative relationships both within and outside ARS. Given the constraints of weeds and invasive plants on all aspects of food, feed, and fiber systems, there is an acknowledged need to face new challenges, including agriculture and natural resources sustainability, economic resilience and reliability, and societal health and well-being.
A 2016 study season and 2017 excavation season at the 95-hectare walled site of Kurd Qaburstan on the Erbil plain have generated a variety of new results. Geophysical survey on the lower town revealed details of the Middle Bronze occupation in the southeast part of the site, including the city wall, a large open area, streets, houses, and a monumental temple comparable to examples from Tell al Rimah, Aššur, and Larsa. Excavations confirmed the Middle Bronze date of the temple and explored further Middle Bronze contexts elsewhere on the lower town. On the High Mound North Slope, Middle Bronze occupation included a fortification wall and large-scale architecture inside it. On the High Mound East, Late Bronze architecture of apparent elite character was documented. Archaeobotanical analyses complementing the excavations reveal the existence of naan-style bread in both Middle and Late Bronze contexts. Given radiocarbon and ceramic results, the Middle Bronze occupation at Kurd Qaburstan is datable to c. 1800 B.C., while the Late Bronze phases on the High Mound East belong to an early LB horizon in the 16–15th centuries B.C., perhaps predating the imposition of Mittani political authority in the region.
In this poster, using the POSYDON code, we present results on binary progenitors of stripped-envelope SNe and their companions. We find that most progenitors are expected to explode, according to typical SN prescriptions (in contrast to single star progenitors). We also show the expected masses and position in the HR diagram of the companions of these SNe at the moment of explosion, allowing us to do a first statistical comparison with the compiled sample of observational detections (or upper limits) on these companions.
Studying phenotypic and genetic characteristics of age at onset (AAO) and polarity at onset (PAO) in bipolar disorder can provide new insights into disease pathology and facilitate the development of screening tools.
Aims
To examine the genetic architecture of AAO and PAO and their association with bipolar disorder disease characteristics.
Method
Genome-wide association studies (GWASs) and polygenic score (PGS) analyses of AAO (n = 12 977) and PAO (n = 6773) were conducted in patients with bipolar disorder from 34 cohorts and a replication sample (n = 2237). The association of onset with disease characteristics was investigated in two of these cohorts.
Results
Earlier AAO was associated with a higher probability of psychotic symptoms, suicidality, lower educational attainment, not living together and fewer episodes. Depressive onset correlated with suicidality and manic onset correlated with delusions and manic episodes. Systematic differences in AAO between cohorts and continents of origin were observed. This was also reflected in single-nucleotide variant-based heritability estimates, with higher heritabilities for stricter onset definitions. Increased PGS for autism spectrum disorder (β = −0.34 years, s.e. = 0.08), major depression (β = −0.34 years, s.e. = 0.08), schizophrenia (β = −0.39 years, s.e. = 0.08), and educational attainment (β = −0.31 years, s.e. = 0.08) were associated with an earlier AAO. The AAO GWAS identified one significant locus, but this finding did not replicate. Neither GWAS nor PGS analyses yielded significant associations with PAO.
Conclusions
AAO and PAO are associated with indicators of bipolar disorder severity. Individuals with an earlier onset show an increased polygenic liability for a broad spectrum of psychiatric traits. Systematic differences in AAO across cohorts, continents and phenotype definitions introduce significant heterogeneity, affecting analyses.
Optical tracking systems typically trade off between astrometric precision and field of view. In this work, we showcase a networked approach to optical tracking using very wide field-of-view imagers that have relatively low astrometric precision on the scheduled OSIRIS-REx slingshot manoeuvre around Earth on 22 Sep 2017. As part of a trajectory designed to get OSIRIS-REx to NEO 101955 Bennu, this flyby event was viewed from 13 remote sensors spread across Australia and New Zealand to promote triangulatable observations. Each observatory in this portable network was constructed to be as lightweight and portable as possible, with hardware based off the successful design of the Desert Fireball Network. Over a 4-h collection window, we gathered 15 439 images of the night sky in the predicted direction of the OSIRIS-REx spacecraft. Using a specially developed streak detection and orbit determination data pipeline, we detected 2 090 line-of-sight observations. Our fitted orbit was determined to be within about 10 km of orbital telemetry along the observed 109 262 km length of OSIRIS-REx trajectory, and thus demonstrating the impressive capability of a networked approach to Space Surveillance and Tracking.
Determining infectious cross-transmission events in healthcare settings involves manual surveillance of case clusters by infection control personnel, followed by strain typing of clinical/environmental isolates suspected in said clusters. Recent advances in genomic sequencing and cloud computing now allow for the rapid molecular typing of infecting isolates.
Objective:
To facilitate rapid recognition of transmission clusters, we aimed to assess infection control surveillance using whole-genome sequencing (WGS) of microbial pathogens to identify cross-transmission events for epidemiologic review.
Methods:
Clinical isolates of Staphylococcus aureus, Enterococcus faecium, Pseudomonas aeruginosa, and Klebsiella pneumoniae were obtained prospectively at an academic medical center, from September 1, 2016, to September 30, 2017. Isolate genomes were sequenced, followed by single-nucleotide variant analysis; a cloud-computing platform was used for whole-genome sequence analysis and cluster identification.
Results:
Most strains of the 4 studied pathogens were unrelated, and 34 potential transmission clusters were present. The characteristics of the potential clusters were complex and likely not identifiable by traditional surveillance alone. Notably, only 1 cluster had been suspected by routine manual surveillance.
Conclusions:
Our work supports the assertion that integration of genomic and clinical epidemiologic data can augment infection control surveillance for both the identification of cross-transmission events and the inclusion of missed and exclusion of misidentified outbreaks (ie, false alarms). The integration of clinical data is essential to prioritize suspect clusters for investigation, and for existing infections, a timely review of both the clinical and WGS results can hold promise to reduce HAIs. A richer understanding of cross-transmission events within healthcare settings will require the expansion of current surveillance approaches.
Excavations at the 109 hectare site of Kurd Qaburstan on the Erbil plain in the Kurdistan Region of Iraq were conducted by the Johns Hopkins University in 2013 and 2014. The Middle Bronze Age (Old Babylonian period) is the main period of occupation evident on the site, and the project therefore aims to study the character of a north Mesopotamian urban centre of the early second millennium b.c. On the high mound, excavations revealed three phases of Mittani (Late Bronze) period occupation, including evidence of elite residential architecture. On the low mound and the south slope of the high mound, Middle Bronze evidence included domestic remains with numerous ceramic vessels left in situ. Also dating to the Middle Bronze period is evidence of a city wall on the site edges. Later occupations include a cemetery, perhaps of Achaemenid date, on the south slope of the high mound and a Middle Islamic settlement on the southern lower town. Faunal and archaeobotanical analysis provide information on the plant and animal economy of the second millennium b.c. occupations, and geophysical results have documented a thirty-one hectare expanse of dense Middle Bronze Age architecture in the northern lower town.
Management of Italian ryegrass in cereal-based cropping systems continues to be a major production constraint in areas of the United States, including the soft white winter wheat producing regions of the Pacific Northwest. Pyroxasulfone is a soil-applied herbicide with the potential to control broadleaf and grass weed species, including grass weed biotypes resistant to group 1, 2, and 7 herbicides, in several crops for which registration has been completed or is pending, including wheat, corn, sunflower, dry bean, and soybean. Field experiments were conducted from 2006 through 2009 near Corvallis, OR, to evaluate the potential for Italian ryegrass control in winter wheat with applications of pyroxasulfone. Application rates of PRE treatments ranged from 0.05 to 0.15 kg ai ha−1. All treatments were compared to standard Italian ryegrass soil-applied herbicides used in winter wheat, including diuron, flufenacet, and flufenacet + metribuzin. Visual evaluations of Italian ryegrass and ivyleaf speedwell control and winter wheat injury were made at regular intervals following applications. Winter wheat yields were quantified at grain maturity. Ivyleaf speedwell control was variable, and Italian ryegrass control following pyroxasulfone applications ranged from 65 to 100% and was equal to control achieved with flufenacet and flufenacet + metribuzin treatments and greater than that achieved with diuron applications. Winter wheat injury from pyroxasulfone ranged from 0 to 8% and was most associated with the 0.15–kg ha−1 application rate. However, this early-season injury did not negatively impact winter wheat yield. Pyroxasulfone applied at the application rates and timings in these studies resulted in high levels of activity on Italian ryegrass and excellent winter wheat safety. Based on the results, pyroxasulfone has the potential to be used as a soil-applied herbicide in winter wheat for Italian ryegrass management and its utility for management of other important grass and broadleaf weeds of cereal-based cropping systems should be evaluated.
Greenhouse studies were conducted to evaluate the influence of selected adjuvants on glyphosate efficacy on yellow nutsedge and tuber production. Glyphosate was applied at 0, 0.25, 0.43, 0.87, 1.26 (1× rate), and 1.74 kg ae ha−1 at 31 d after yellow nutsedge was planted. Each rate was mixed with one of the following adjuvants: ammonium sulfate (AMS), AMS plus nonionic surfactant (NIS), or AMS plus an experimental adjuvant (W-7995) plus NIS. Plants were evaluated for injury and for the number and size of tubers produced. Dose–response curves based on log-logistic models were used to determine the effective glyphosate rate plus adjuvant that provided both 90% effective dose (ED90) for yellow nutsedge injury and reduced tuber production. Addition of NIS to glyphosate plus AMS resulted in the greatest yellow nutsedge injury at 28 d after treatment (DAT). Addition of the experimental adjuvant plus NIS resulted in injury similar to NIS alone. The ED90 for injury at 28 DAT was 2.12 kg ha−1 with glyphosate plus AMS and NIS compared with 2.18 kg ha−1 for W-7995 plus NIS and 3.06 kg ha−1 with AMS alone. The ED90 rates with different adjuvants represent 168%, 173%, and 243% of the highest glyphosate rate (1.26 kg ha−1) labeled for application on many glyphosate-resistant crops. However, the estimated ED90 to reduce small, medium, large, and total tubers were 1.60, 1.50, 1.63, and 1.66 kg ha−1, respectively. Increases in labeled rates of glyphosate may be required to reduce yellow nutsedge tuber production in field conditions. Use of lower glyphosate rates should be discouraged because it may increase tuber production and exacerbate yellow nutsedge expansion in infested fields.
Molecular data are becoming an indispensable tool for the reconstruction of phylogenies. Fossil molecular data remain scarce, but have the potential to resolve patterns of deep branching and provide empirical tests of tree reconstruction techniques. A total evidence approach, combining and comparing complementary morphological, molecular and stratigraphical data from both recent and fossil taxa, is advocated as the most promising way forward because there are several well-established problems that can afflict the analysis of molecular sequence data sometimes resulting in spurious tree topologies. The integration of evidence allows us to: (1) choose suitable taxa for molecular phylogenetic analysis for the question at hand; (2) discriminate between conflicting hypotheses of taxonomic relationship and phylogeny; (3) evaluate procedures and assumptions underlying methods of building trees; and (4) estimate rates of molecular evolution in the geological past. Paleontology offers a set of independent data for comparison and corroboration of analyses and provides the only direct means of calibrating molecular trees, thus giving insight into rates of molecular evolution in the geological past.