Introduction
In large-scale surveys with objective measures of cognitive functioning, constraints on time and resources underscore the importance of optimizing the quantification of cognition given available information. Increased adoption and use of computer-assisted personal interview (CAPI) methods for cognitive testing has led to increases in the availability of new types of survey metadata, including information on the time spent on cognitive testing (Banerjee et al., Reference Banerjee, Jain, Khobragade, Weerman, Hu, Chien, Dey, Chatterjee, Saxton, Keller, Crimmins, Toga, Jain, Shanthi, Kurup, Raman, Chakrabarti, Varghese, John, Joshi and Rajguru2020; Humphreys et al., Reference Humphreys, Duta, Montana, Demeyere, McCrory, Rohr, Kahn, Tollman and Berkman2017; Shega et al., Reference Shega, Sunkara, Kotwal, Kern, Henning, McClintock, Schumm, Waite and Dale2014). The use of speeded tests, such as the Trail-Making Test, to assess cognition supports the general notion that response times on cognitive tests contain meaningful information about cognitive functioning (Bowie & Harvey, Reference Bowie and Harvey2006; Salthouse, Reference Salthouse1996). However, in large-scale community-based survey research on cognitive aging and dementia, this information is not typically used beyond general data monitoring procedures and quality control.
A substantial body of evidence from psychology and educational testing provides further support for the utility of information on timing and suggests a tradeoff between timing and accuracy or score; those with the fastest times may be guessing or otherwise sacrificing accuracy for speed (Kyllonen & Zu, Reference Kyllonen and Zu2016; Wickelgren, Reference Wickelgren1977). Longer response times may indicate more careful attention to the question or may be a sign of increased struggle in understanding and answering a given question. The complexities of the combined effects of guessing, careful consideration of questions, and observed difficulties with challenging questions likely help explain observed nonmonotonic, nonlinear relationships between response times and ability (Chen, De Boeck, Grady, Yang, & Waldschmidt, Reference Chen, De Boeck, Grady, Yang and Waldschmidt2018; Dodonov & Dodonova, Reference Dodonov and Dodonova2012). Evidence also suggests that the impacts of the different factors affecting response times (e.g., guessing, processes around critical reasoning) may vary by item type (DiTrapani et al., Reference DiTrapani, Jeon, De Boeck and Partchev2016; Goldhammer et al., Reference Goldhammer, Naumann, Stelter, Tóth, Rölke and Klieme2014; Hess et al., Reference Hess, Johnston and Lipner2013). For example, one study of items on the Dental Admission Test found that the maximum point on the curve describing the relationship between score and response time (i.e., the time taken associated with the highest ability level) was earlier for knowledge-based items than items requiring controlled processes (Chen et al., Reference Chen, De Boeck, Grady, Yang and Waldschmidt2018).
Despite robust existing research on response times in psychology and educational testing, relatively few studies have examined the use of information on response times from non-speeded tests to inform the estimation of cognitive functioning and classification of mild cognitive impairment (MCI) and dementia in older adults in large-scale population-based research. Issues associated with challenges with measuring cognitive functioning in participants’ homes in large-scale survey research also raises questions about the potential value of timing data in this context. However, there is growing evidence that response times on non-cognitive survey questions are associated with future cognitive functioning, MCI, and dementia (Junghaenel et al., Reference Junghaenel, Schneider, Orriens, Jin, Lee, Kapteyn, Stone, Meijer, Zelinski, Hernandez and Stone2023; Schneider et al., Reference Schneider, Junghaenel, Meijer, Stone, Orriens, Jin, Zelinski, Lee, Hernandez and Kapteyn2023; Seelye et al., Reference Seelye, Mattek, Sharma, Riley, Austin, Wild, Dodge, Lore and Kaye2018), which may be useful for the estimation of cognitive functioning in surveys without formal cognitive assessments. However, in studies with objective cognitive testing, timing data may also be used to enhance the available information on cognition from test scores.
To understand the potential benefit of using timing data, it is first necessary to assess whether information on timing from cognitive testing provides useful information independently of the already available information on performance from cognitive test scores. One prior study found that longer and more variable response times for items included in the Montreal Cognitive Assessment (MoCA) were significantly associated with concurrent and future cognitive functioning (Sanders et al., Reference Sanders, Schofield, Schumm and Waite2024); however, the brevity of the MoCA instrument precludes examination of item-level heterogeneity, and nonlinearities in associations were not considered. Additionally, despite increasing research interest in cross-national and global research on cognitive aging and dementia, this work and other prior research on timing has largely been conducted in the United States. Differences in testing environments at respondent’s houses, differences in the distribution and range of socioeconomic status across respondents, and cultural perceptions of time in low- and middle-income countries (LMICs) may influence observed findings.
We sought to build on this existing evidence by using data from the Longitudinal Aging Study in India Diagnostic Assessment of Dementia (LASI-DAD) study to examine the added value of timing data from cognitive testing for the assessment of cognitive functioning in large-scale survey research in the Indian context. We leveraged the large cognitive battery included in the LASI-DAD study to examine associations between cognitive functioning and time spent on cognitive test items across 36 items, considering item-level heterogeneity and allowing for nonlinearities in associations. We also evaluated associations with MCI and dementia as key clinical outcomes.
Method
Sample
LASI-DAD is a nationally representative sample of 4,096 adults 60 years and older in India (Lee, Banerjee, Khobragade, Angrisani, & Dey, Reference Lee, Banerjee, Khobragade, Angrisani and Dey2019; Lee, Khobragade, et al., Reference Lee, Khobragade, Banerjee, Chien, Angrisani, Perianayagam, Bloom and Dey2020). LASI-DAD was designed as a sub-study of the broader Longitudinal Aging Study in India (LASI) study with focus on the assessment of cognition and dementia. LASI-DAD uses stratified sampled procedures to select LASI participants from 18 states and union territories with equal numbers of those at high and low risk of cognitive impairment based on cognitive testing in LASI. We used data from N = 4,091 respondents with available information on the timing of cognitive tests from the CAPI survey instrument. All participants gave informed consent for participation (either written or thumbprint). We obtained ethics approval from the Indian Council of Medical Research (2202-16741/F1) and all collaborating institutions and research was conducted in accordance with the Helsinki Declaration.
Cognitive functioning
The LASI-DAD study administered the Harmonized Cognitive Assessment Protocol (HCAP) battery of cognitive tests and informant reports, which has been described in detail elsewhere (Langa et al., Reference Langa, Ryan, McCammon, Jones, Manly, Levine, Sonnega, Farron and Weir2020; Lee, Khobragade, et al., Reference Lee, Khobragade, Banerjee, Chien, Angrisani, Perianayagam, Bloom and Dey2020). The cognitive battery was designed to take approximately 1 hour, and included tests assessing orientation, memory, language, executive functioning, and visuospatial functioning (Supplementary Materials 1). Some adaptations to the original HCAP battery were necessary to ensure cultural appropriateness and adequate performance in a population with low education and literacy (Banerjee et al., Reference Banerjee, Jain, Khobragade, Weerman, Hu, Chien, Dey, Chatterjee, Saxton, Keller, Crimmins, Toga, Jain, Shanthi, Kurup, Raman, Chakrabarti, Varghese, John, Joshi and Rajguru2020). The battery was translated into 12 languages and was administered in each participant’s preferred spoken language. A continuous measure of cognitive functioning based on all available cognitive tests was previously estimated using confirmatory factor analysis, and was shown to perform equivalently across language of administration (Gross et al., Reference Gross, Khobragade, Meijer and Saxton2020). The measure is scaled to have a mean of 0 and standard deviation of 1 in the full LASI-DAD sample.
Cognitive test timing
The time spent on each cognitive task was captured by the CAPI survey system as the time spent on each screen. Each screen typically included a combination of related instructions and one or more items. For cognitive tests that assess a specific cognitive process, and which are not typically disaggregated at the item level (e.g., Raven’s progressive matrices), time spent on multiple individual items belonging to the same cognitive test was aggregated to the test level. To simplify the text in the present study, we use the term “item” to refer to the mixture of both items and tests considered (test-level scores and timing from tests of specific cognitive processes and item-level scores and timing from more heterogeneous test batteries such as the Hindi Mental State Examination) (Ganguli et al., Reference Ganguli, Ratcliff, Chandra, Sharma, Gilby, Pandav, Belle, Ryan, Baker, Seaberg and Dekosky1995).
We excluded the first items of the judgement and problem-solving item set, Token test, and Raven’s progressive matrices, as these items appeared on the same screen as large amounts of introductory text and instructions and it was therefore impossible to disentangle the time spent on instructions from the time spent answering these items. Items for orientation to time (what is the month, year, day of week, and date) were presented on the same screen and were therefore summed together. We excluded items conducted with pencil and paper (write sentence, copy pentagons, constructional praxis, clock drawing), as timing data was not collected. We also excluded delayed story recall, as each story point was assessed on a separate screen; due to the brief nature of the items in contrast to the time taken to switch screens, the captured timing measure would likely be less reflective of respondent ability. Finally, we excluded the item for floor of building as it was not administered to over 75% of respondents. It is important to acknowledge that time as captured by the CAPI instrument includes the interviewer’s delivery of the question and could also include any interruptions to the interview that may have occurred while on a specific screen.
To reduce the number of outliers due to testing interruptions, we examined the distributions of timing data and set extreme observations to missing. The items varied in the skewness of the observed timing data with a bimodal distribution of variable skewness across tests. Therefore, we defined extreme as the 99th percentile of the data distribution for items with low skew, and the 98th percentile of the data distribution for items with high skew (Supplementary Materials 2). Because this method was designed to be conservative in removing data, we selected analytic approaches (analysis of quintiles, splines with boundary knots) that are robust to outliers at the extremes.
We divided the timing data for each item into quintiles for use in primary analyses. In addition to considering timing data on individual test items, we also calculated quintiles of an overall timing variable to characterize broad patterns in the timing data, considering the time spent on all items. The overall timing variable was calculated as the sum of time spent across all items considered in the analyses.
Cognitive impairment status
We used Clinical Dementia Rating (CDR) scale classifications from an online clinical consensus process. We compared CDR = 0 (no dementia), CDR = 0.5 (questionable dementia or MCI, here referred to as MCI), and CDR ≥1 (dementia). Details on the process for the online adjudication procedure are available elsewhere (Lee, Ganguli, et al., Reference Lee, Ganguli, Weerman, Chien, Lee, Varghese and Dey2020). Because CDR adjudication was not completed for participants in Phase 1 of data collection, analyses using CDR classifications were conducted on a subset of respondents (N = 2,525).
Other measured covariates
We also used self-reported information on age, sex/gender (hereafter gender), education (none/less than secondary/secondary or higher), marital status (married or partnered/other), and hearing impairment (whether the respondent ever had hearing/ear-related problems or conditions). Rural/urban residence was determined based on the classification of locations in the 2011 Census.
Statistical analysis
We examined the characteristics of the sample by demographic variables, both overall and by quintiles of the overall timing variable (timing summed across items). We also assessed the means and distributions of cognitive scores across quintiles of the overall timing variable.
Primary analyses consisted of regression models for each item (or set of items appearing together on the CAPI screen) to estimate associations between general cognitive functioning, the dependent variable, and indicators for quintile of time spent on the test item, the independent variable. All models adjusted for age, gender, interviewer, and score on the specific item. By adjusting for score on the specific item, models capture the added value of timing data for the measurement of cognitive functioning, beyond traditional test scores. We used five quintiles to ensure that models had sufficient flexibility to capture nonlinearities while prioritizing the interpretability of estimates. To ensure findings were not sensitive to our choice of the number of bins included, we also estimated item-specific models using flexible splines for item timing to assess the consistency of findings. We used restricted cubic splines with 3 degrees of freedom and boundary knots at the 5th and 90th percentiles, which constrain the tail segments to be linear to prevent potential outliers from having undue influence. We standardized all continuous item-specific timing data prior to estimating spline models to facilitate comparisons across items. To evaluate the statistical significance of comparisons between different points on the estimated spline terms (analogous to the comparison of quintiles in primary analyses), we used a non-parametric bootstrap (details in Supplementary Materials 3).
We then assessed whether item difficulty could explain heterogeneity in findings across items by numerically and visually evaluating correlations between item difficulty and the item-level differences in cognitive functioning between the third and first quintiles of time spent on a specific item. We used item response theory methods to quantify the average item difficulty across item-specific thresholds (additional details in Supplementary Materials 4). Item difficulties for each item are in Supplementary Materials 5.
To evaluate the utility of information on timing for classification purposes, we first estimated multinomial logistic models of CDR MCI and dementia on quintile of time spent on each item. Models were analogous to those used in the primary analyses; however, we used multinomial logistic regression given the categorical nature of the CDR outcome. To further assess whether data on timing may contain information on incipient impairments that are present before changes can be observed in cognitive test scores, we estimated the discordance between cognitive functioning and timing data. Specifically, for each item we calculated the following proportion:
We defined timing data suggestive of cognition below the mean as timing data for which the observed time spent was more extreme than the time for which the upper bound of the mean predicted cognitive functioning crossed 0. We only considered times in the tails of the distributions, as the timing data in these regions showed the largest associations with cognitive functioning in primary analyses. These criteria were designed to identify respondents who scored well on cognitive testing, but who, based on timing data, completed cognitive testing in a manner that would suggest lower cognitive functioning. For these respondents, timing data may indicate subtle cognitive deficits that are unobservable using the existing cognitive tests.
We conducted sensitivity analyses to evaluate the stability of our primary results to additional adjustment for language of administration, education, urbanicity, and hearing impairment. We considered adjustment for language of administration as a sensitivity analysis because language was somewhat collinear with interviewer and added a significant number of terms to models, reducing precision. Adjustments for education, urbanicity as a proxy for noise and interruptions, and hearing impairment were considered sensitivity analyses because while these factors may impact speed independently of cognitive functioning through mechanisms such as familiarity with test taking or ability to hear instructions better, evidence suggests these factors may also causally affect cognitive functioning and variation in cognition due to these potentially causal links should not be adjusted away. We also estimated models with executive functioning as the cognitive outcome, as attention/speed was considered a component of executive functioning in factor analyses (Gross et al., Reference Gross, Khobragade, Meijer and Saxton2020) and therefore timing data may be more closely related with executive functioning than general cognitive functioning.
All analyses were conducted in R version 4.2.2.
Results
Sample characteristics
The LASI-DAD sample used for this analysis included 4,091 respondents with a mean age of 69.7 years (Interquartile Range [IQR] 64–74); 53.9% were women (Table 1). The majority (61.9%) of respondents lived in rural settings, and almost half (49.0%) had no formal education. There were some differences in demographics by quintile of overall time spent on items; for example, respondents in Quintile 2 were on average younger (69.0 years) than those in the fastest (Quintiles 1; 69.7 years) or slowest (Quintile 5; 70.4 years)) quintiles. A smaller proportion of respondents in Quintile 2 had no education (39.3%) compared to Quintiles 1 (50.9%) and 5 (59.3%).
Univariate associations between cognition and overall time on cognitive items
The distributions of cognitive functioning were largely overlapping, comparing across quintiles of overall time spent on cognitive tasks, but there was a clear pattern when examining means of distributions (Table 1, Supplementary Materials 6). The overall pattern was nonmonotonic: increasing between Quintiles 1 (Mean: −0.09 SD) and 2 (0.20 SD), and then decreasing across the remaining 3 Quintiles (3: 0.13 SD; 4: −0.05 SD; 5: −0.25 SD). A similar pattern was observed for executive functioning.
Item-specific associations between timing and cognitive functioning
To examine item-specific associations, we assessed models in which we predicted general cognitive functioning (the factor score) as a function of an item score and the time spent on the particular item, again categorized into quintiles, with Quintile 3 the reference category. Compared to Quintile 3, respondents who spent slightly less time on cognitive test items (Quintile 2) tended to have higher cognitive functioning on average, holding their observed scores on the specific item constant (Figure 1). Of the 36 items considered, 13 (36%) had significantly positive associations, suggesting that slightly faster times to complete these 18 items is associated with better cognition; only 1 had a significantly negative association.
In contrast, respondents who spent more time on items compared to those in Quintile 3 tended to have lower cognitive functioning holding observed scores constant. Results were stronger for Quintile 5 (24 items with negative associations, 0 positive) than Quintile 4 (10 items with negative associations, 1 positive). For Quintile 1, results were less consistent, with time spent on 7 items having positive and 11 negative associations with cognitive functioning, holding score constant.
Overall findings were consistent in models with executive functioning as the outcome instead of general cognitive performance, and in models controlling for either education, urbanicity, language of administration, or hearing impairment (Supplementary Materials 7–11). Use of flexible cubic splines in lieu of quintiles also yielded results with consistent overall findings: for most items, predicted cognitive functioning given timing data suggested low levels of cognitive functioning for the shortest response times and as the time spent on items increased, predicted cognitive functioning then peaked quickly and tapered off after 0.5 standard deviation units of response time with a long right tail (Figure 2). Comparisons between predictions at different points on the curves were largely statistically significant (Supplementary Materials 3). We observed the most heterogeneity across items in the left tails. This aligns with our findings regarding Quintile 1 in the main analyses: for some items, after adjusting for item score, answering very quickly was associated with high cognitive functioning, while for other items this is associated with low cognitive functioning after adjusting for item score.
Item difficulty and item-level differences
Item difficulty was negatively correlated with coefficients for Quintile 1 (p = 0.002), explaining some of the observed heterogeneity (Supplementary Materials 12). A negative correlation suggests that for easier items with lower difficulty, coefficients for Quintile 1 are more likely to be positive, suggesting that quicker response times are associated with higher cognition, controlling for item score. However, for more difficult, complex items (higher difficulty), coefficients for Quintile 1 are more likely to be negative, suggesting that quicker response times are associated with lower cognition, controlling for item score, potentially due to the respondent guessing or giving up.
Item-specific associations between timing and MCI and dementia
Associations between item-level timing and odds of MCI or dementia were consistent with findings using continuous cognition, albeit somewhat weaker (Figure 3). Of the 36 items, slower test performance was associated with elevated odds of MCI for 6 and 14 items for Quintiles 4 and 5, respectively. Similarly, slower test performance was associated with elevated odds of dementia for 4 and 13 items for Quintiles 4 and 5, respectively. Although observed effect sizes were larger for dementia, the number of statistically significant associations in the expected direction (based on primary analyses) was similar between MCI and dementia for Quintiles 4 and 5. However, findings for MCI and dementia diverged for Quintiles 1 and 2, with, for example, 6 significant negative associations and 0 positive associations for MCI and 7 significant positive associations and 0 negative associations for dementia in Quintile 1. This suggests that quick answers were associated with lower odds of MCI but higher odds of dementia, perhaps indicating that the effect of cognitive status on giving up and therefore answering quickly is only present at later stages of disease.
Proportion of respondents with mismatched cognition and timing data
The proportion of respondents who had above average cognitive functioning (factor scores above 0) but timing data suggestive of below average cognitive functioning varied between items from 43% (token test) to 0% (serial 7s, 4 other items) (Figure 4, Supplementary Materials 13). The proportion of respondents with discordant cognitive performance and timing data was above 20% for 25/36 items, suggesting a substantial number of individuals with intact cognition as measured by cognitive test scores, but for whom timing data may suggest subtle cognitive deficits.
Discussion
Using representative data on older adults from the LASI-DAD study, we found that data on timing from cognitive tests may be useful in improving or augmenting existing measures of cognitive functioning from large-scale community-based surveys. Analyses highlighted evidence that time spent on items was associated with cognitive functioning and cognitive outcomes after controlling for the traditional score on the same cognitive test. Longer time (than the central quintile) spent on items was associated with lower cognition, whereas shorter time (than the central quintile) spent on items was associated with either higher or lower cognition depending on item difficulty. For more difficult items, shorter completion times were associated with lower cognition, consistent with guessing or rushing through the interview. In contrast, for easier items, shorter completion times were associated with higher cognition. We also identified respondents with measures of objective cognitive functioning above the mean, but timing data suggestive of cognitive functioning significantly below the mean, which may indicate that timing data could be valuable for the identification of subtle cognitive deficits.
Our findings align with prior results from the National Social, Health and Aging Project (NSHAP), which highlighted initial evidence that information on timing captured using CAPI survey instruments may contain important information on cognition (Sanders et al., Reference Sanders, Schofield, Schumm and Waite2024). Findings from NSHAP also provided longitudinal evidence that information on timing is associated with future cognitive functioning independently of cognitive test scores. As additional waves of LASI-DAD data become available, these findings should be replicated and extended using the larger included cognitive battery in the LASI-DAD study. Our findings also follow from prior evidence on the potential utility of timing data in related contexts. One prior web-based study of 138 older adults found that total completion time of a novel cognitive assessment tool, including time for instructions, practice items, and test items, was highly predictive of cognitive status, lending initial support for the use of meta-data from cognitive testing in assessing cognition (Dorociak et al., Reference Dorociak, Mattek, Lee, Leese, Bouranis, Imtiaz, Doane, Bernstein, Kaye and Hughes2021). Our results, in tandem with results from NSHAP, extend this initial evidence by using larger, population-representative samples with more extensive cognitive batteries administered via a CAPI survey instrument.
In comparison to these prior studies, the length of the LASI-DAD battery allowed for a more detailed examination of item-level heterogeneity. Though we focused on larger patterns of results in overall findings to protect against concerns of type I error given the large number of hypothesis tests and models, identification of items with consistently strong associations across various analyses may highlight specific items for which timing data may be particularly indicative of cognitive status. In current analyses, word list recognition, digit span backwards, and the judgement questions were all significantly associated with multiple cognitive outcomes considered, suggesting timing data from these items may be particularly meaningful.
Other existing literature has focused specifically on response times for non-cognitive survey items (Junghaenel et al., Reference Junghaenel, Schneider, Orriens, Jin, Lee, Kapteyn, Stone, Meijer, Zelinski, Hernandez and Stone2023; Schneider et al., Reference Schneider, Junghaenel, Meijer, Stone, Orriens, Jin, Zelinski, Lee, Hernandez and Kapteyn2023; Seelye et al., Reference Seelye, Mattek, Sharma, Riley, Austin, Wild, Dodge, Lore and Kaye2018). These studies have found that the time spent on non-cognitive survey items is predictive of future cognitive functioning, MCI, and dementia (Junghaenel et al., Reference Junghaenel, Schneider, Orriens, Jin, Lee, Kapteyn, Stone, Meijer, Zelinski, Hernandez and Stone2023; Schneider et al., Reference Schneider, Junghaenel, Meijer, Stone, Orriens, Jin, Zelinski, Lee, Hernandez and Kapteyn2023; Seelye et al., Reference Seelye, Mattek, Sharma, Riley, Austin, Wild, Dodge, Lore and Kaye2018). Findings from these studies motivate the need for replication efforts in diverse settings such as LASI and LASI-DAD and have important implications for the potential utility of existing data sources without objective cognitive testing, as response times may allow for estimation of cognition in these studies. However, beyond this potential application, findings also more broadly support the notion that information on the time spent on tasks requiring cognitive effort yields meaningful information about cognition.
Our analyses indicated not only that response times from cognitive testing were associated with cognitive functioning, but that they uniquely contributed to explaining underlying cognitive functioning after accounting for scores from standard scoring procedures. These findings suggest that use of information on time spent on cognitive tests may lead to improvements in the precision and quality of derived measures of cognitive functioning based on survey data. Researchers in the field of educational testing have previously developed combined models that either incorporate response times into item-level item-response theory models (Wang & Hanson, Reference Wang and Hanson2005) or into joint models that estimate both ability and speed as correlated latent traits (van der Linden, Reference van der Linden2008). The correlated traits model was shown to improve precision in a test of fluid reasoning administered to a highly educated population, such that the battery could be reduced by 33% without a corresponding reduction in precision through the incorporation of information on response times (Bertling & Weeks, Reference Bertling and Weeks2018). The current results suggest that the use of such joint models in cognitive aging research has the potential to improve measurement and should be explored in future work.
Our findings of discordance between objective testing and implied cognition based on timing data suggest that information on timing may lend insights into subtle cognitive deficits that may not be apparent through objective cognitive testing. Of those with estimates of cognitive functioning above the mean, the proportion of respondents with discordant scores on objective testing and timing data on the three items identified with strong associations between timing data and cognitive outcomes was 28% (digit span backwards), 23% (word list recognition), and 21% (judgement questions), indicating sizable proportions for these tasks with strong associations. The notion that this observed discordance may signal subtle cognitive deficits is supported by findings from a prior study of survey meta-data from non-cognitive tests, which found that individuals with incident MCI had increased completion times one year prior to formal diagnoses compared to those who remained cognitively intact (Seelye et al., Reference Seelye, Mattek, Sharma, Riley, Austin, Wild, Dodge, Lore and Kaye2018). Other research has identified different subtle cognitive changes occurring prior to MCI diagnosis, including impaired response inhibition and differences in the patterns of response to neuropsychological tests (Schmid, Taylor, Foldi, Berres, & Monsch, Reference Schmid, Taylor, Foldi, Berres and Monsch2013; Wylie et al., Reference Wylie, Ridderinkhof, Eckerle and Manning2007). Protective factors, such as education, that may help respondents score well on cognitive testing in the face of subtle deficits may not protect against deficits that could show up in timing data or other metadata.
Taken together, our results and these prior findings indicate that metadata and other assessments of more subtle cognitive problems may be helpful for capturing mild impairments in those with largely intact cognitive functioning. Neuropsychological batteries designed for use in large-scale surveys often include a number of screening tests, such as the Mini-Mental State Examination, which do not have adequate precision for the detection of mild cognitive deficits (Gross, Jones, Fong, Tommet, & Inouye, Reference Gross, Jones, Fong, Tommet and Inouye2014; Koski et al., Reference Koski, Xie and Konsztowicz2011). Given the limitations of these batteries, particularly for the assessment of subtle cognitive changes among those with higher levels of cognitive functioning, augmentation of traditional scores with timing metadata could yield important improvements for the estimation of cognition across the entire spectrum of cognitive functioning. Future research should leverage longitudinal data to further evaluate whether individuals with mismatched data are more likely to decline in future waves, which would provide additional evidence that such discordance indicates early cognitive deficits. In addition to the potential research utility of timing information, the adoption of tablet-based or computerized testing in clinical settings would also facilitate the integration of metadata on time taken to complete cognitive tasks into clinical screening tools for subtle cognitive deficits in clinical settings. The development of such protocols was outside the scope of the current paper but should be the topic of follow-up research.
It is important to consider the ways in which the study context may impact results. Cultural attitudes towards time and experience with high-stakes testing have the potential to impact how respondents react to cognitive tests and prior research has shown this can impact timing data from speeded cognitive tests (Agranovich et al., Reference Agranovich, Panter, Puente and Touradji2011; Eizaguirre et al., Reference Eizaguirre, Vanotti, Aguayo Arelis, Rabago Barajas, Cores, Macías, Benedict and Cáceres2020). Social and cultural psychology has suggested LMICs often have a slower pace of life and less emphasis on timeliness compared to high-income contexts (Levine, Reference Levine2008). Given these patterns, overall lower education levels, and less exposure to high-stakes testing, respondents in LASI-DAD likely have less of an internal sense of time pressure than respondents in high-income contexts. Although differences in culture are important to consider in thinking about the generalizability of findings across settings and replication in other contexts is needed to confirm the consistency of patterns, the consistency of findings regarding the utility of timing data across both high-income contexts and India, despite vast differences in cultural attitudes toward time and educational experiences suggests that these data may be useful across diverse settings.
Additionally, considerations around the extent to which noise and distractions could impact testing are different across study and country contexts. The presence of robust associations based on timing data collected at respondents’ homes throughout India, where overall measurement error is likely higher due to higher levels of noise pollution, higher likelihood of interruptions, and a greater number of logistical challenges as compared to lab-based testing or even large-scale surveys in high-income contexts, lends even more strength to observed associations due to the expected attenuation of effects in the presence of this measurement error. Associations may be even stronger in settings with lower noise and fewer interruptions.
Key study strengths include the use of a large, nationally representative dataset with a comprehensive neuropsychological battery, and the use of many supplementary and sensitivity analyses to evaluate the robustness of findings to differences in model form and adjustments for a variety of factors that could impact the time spent on testing. Limitations should also be considered. First, data collection procedures were not optimized for the collection of timing data; information on the time spent on cognitive tests included time for instructions and time due to interruptions in testing and data collection. Respondents did not know in advance that data on response times would be considered by investigators and may have changed their behavior if presented with this information. Despite this lack of optimization, instruction time may contain relevant information; if interviewers slow down for those with cognitive impairments, this may contribute to some of the observed signal in analyses and may be helpful in characterizing cognition. However, to minimize the extent of these concerns in general, we made decisions on a post-hoc, case-by-case basis to exclude items on the same screens as blocks of instructions or groups of items on a single screen and used data-driven methods to identify outliers most likely due to interruptions. Some residual measurement error in the data due to inaccuracies in the collection of timing data likely remains; our results are conservative in comparison to what would be expected in the presence of less measurement error. With adjustments to the design of the CAPI instrument and interviewer training, it may be possible to reduce measurement error in the collection of timing data and increase its utility.
Second, in item-level analyses a large number of models and hypothesis tests are necessary, leading to some concern about inflated type 1 error. To alleviate these worries and limit concerns around over-interpreting individual findings, we focused our overall conclusions on broad patterns of results, rather than individual findings. Examination of these patterns across the range of items gives confidence that findings are not due to chance. For example, in primary regression models, 24/36 estimates for Quintile 5 were statistically significant and negative while there were no estimates that were statistically significant and positive; however, in contrast, if results were due solely to type 1 error, we would expect approximately 5% (∼2 models) to be statistically significant with a similar number of both positive and negative estimates. Finally, the LASI-DAD sample is currently cross-sectional; therefore, it was not possible to evaluate associations between timing data and future cognitive impairment or clinical status. Additional research is needed to extend the current analysis and evaluate associations with future cognitive outcomes when future waves become available.
Our results advance the evidence supporting the utility of information on the time spent on cognitive tasks, conditional on scores from traditional scoring procedures, even in studies conducted in real-world environments with the potential for interruptions and substantial measurement error. Findings support the integration of timing data from the administration of cognitive testing into the quantification of cognitive functioning in large-scale survey research on cognitive aging. Information may be particularly valuable for the assessment of subtle cognitive changes among those with higher levels of cognitive functioning. Future research is needed to bolster longitudinal evidence and to implement and test methods for the integration of information on timing into models for the quantification of cognitive functioning.
Supplementary material
The supplementary material for this article can be found at https://doi.org/10.1017/S1355617724000742.
Acknowledgements
We acknowledge the participants and families who participated in the LASI-DAD study, the staff at the study sites, as well as the personnel involved in the data collection and data release. We thank the entire LASI-DAD field team, with special mention to Dr Joyita Banerjee, for her insightful comments and contributions to the project and the paper.
Funding statement
This work was supported by the National Institutes of Health (EN, ALG, JL grant number 3R01 AG030153), (ALG, JL grant number R01AG070953), (EN, JL grant number R01AG051125).
Competing interests
None.