Abstract
PURPOSE Limited health literacy is increasingly recognized as a barrier to receiving adequate health care. Identifying patients at risk of poor health outcomes secondary to limited health literacy is currently the responsibility of clinicians. Our objective was to identify which screening questions and demographics independently predict limited health literacy and could thus help clinicians individualize their patient education.
METHODS Between August 2006 and July 2007, we asked 225 patients being treated for diabetes at an academic primary care office several questions regarding their reading ability as part of a larger study (57% response rate). We built a logistic regression model predicting limited health literacy to determine the independent predictive properties of these questions and demographic variables. Patients were classified as having limited health literacy if they had a Short Test of Functional Health Literacy in Adults (S-TOFHLA) score of less than 23. The potential predictors evaluated were self-rated reading ability, highest education level attained, Single-Item Literacy Screener (SILS) result, patients’ reading enjoyment, age, sex, and race.
RESULTS Overall, 15.1% of the patients had limited health literacy. In the final model, 5 of the potential predictors were independently associated with increased odds of having limited health literacy. Specifically, patients were more likely to have limited health literacy if they had a poorer self-rated reading ability (odds ratio [OR] per point increase in the model = 3.37; 95% confidence interval [CI], 1.71–6.63), more frequently needed help reading written health materials (assessed by the SILS) (OR = 2.03; 95% CI, 1.26–3.26), had a lower education level (OR = 1.89; 95% CI, 1.12–3.18), were male (OR = 4.46; 95% CI, 1.53–12.99), and were of nonwhite race (OR = 3.73; 95% CI, 1.04–13.40). These associations were not confounded by age. The area under the receiver operating characteristic curve was 0.9212.
CONCLUSIONS Self-rated reading ability, SILS result, highest education level attained, sex, and race independently predict whether a patient has limited health literacy. Clinicians should be aware of these associations and ask questions to identify patients at risk. We propose an “SOS” mnemonic based on these findings to help clinicians wishing to individualize patient education.
INTRODUCTION
Health literacy is defined as “the degree to which individuals have the capacity to obtain, process, and understand basic health information and services needed to make appropriate health decisions.”1 Limited health literacy, which refers to marginal health literacy, inadequate health literacy, or both depending on study definitions, has been independently associated with several undesirable health-related outcomes. Compared with individuals with adequate health literacy, those with limited health literacy have poorer understanding of their chronic diseases,2–5 physicians’ instructions,6–8 and health-related Web sites9; poorer disease management skills2; higher levels of disease indicators10,11; and worse self-reported health.12,13 Limited health literacy is also associated with less use of certain preventive services,14,15 increased hospitalizations,12,16 and increased health care costs.17 Not all of these associations have been found with perfect consistency, however.18–20 We refer the interested reader to larger literature reviews and summary analyses for a more comprehensive look at the effects of literacy on health-related outcomes.21,22
Interventions exist to aid persons with limited health literacy. Simplifying instruction forms is an effective means of ensuring better comprehension for entire patient populations.23–25 Patients’ health care teams may be of assistance by providing simplified education and ensuring that patients understand and retain what is being said.26–28 Although the use of such strategies should be encouraged for helping all patients, those with limited health literacy may receive a particularly strong benefit.27
Aggressive educational interventions have shown some additional benefit in outcomes among patients with limited health literacy,5,27 and at least 8 clinical trials of intensive behavioral interventions that may help this population have recently finished or are under way.29 Because most limited health literacy–focused interventions are only investigational, Paasche-Orlow and Wolf30 recently recommended against universal screening for health literacy difficulties. Instead, they recommend that clinicians take responsibility for assessing how well their individual patients understand health information.
Clinicians need to be able to recognize the “symptoms” of limited health literacy and other barriers to acquiring information. Unfortunately, physicians are sometimes poor estimators of literacy level, and the prevalence of limited health literacy varies widely between populations.31,32 Knowing what questions to ask about a patient’s learning styles would thus be helpful for clinicians seeking to personalize their patient education.
Recent studies have suggested that clinicians could use a few questions to identify patients with limited health literacy.33–36 Results of one of these studies were further refined to produce the Single-Item Literacy Screener (SILS), a 1-question test for adequate literacy.37 But these studies have not shown whether such questions are superior to other proxies, such as highest education level attained or self-rated reading ability, which they were intended to replace. Highest education level attained and self-rated reading ability, although not perfect predictors of literacy, may still be strongly associated with literacy level and would therefore merit study as risk factors for a potential literacy problem.38,39
The purpose of this study was to identify questions that could best indicate to a clinician that a patient may have low or marginal health literacy. Our hypothesis was that short screening questions and demographic information would help predict a patient’s literacy status. We also hoped to discover which questions are superior for predicting limited health literacy, and which predict it independently of the other questions and demographic risk factors.
METHODS
Data Collection
The Ohio State University Biomedical Institutional Review Board approved this study, which was conducted at the university’s Rardin Family Practice Center in Columbus, Ohio. At this practice site, 11 faculty family physicians and 15 family practice residents provide primary care to more than 9,100 individuals from the local community. Of this patient population, 60% are women, 53% are white, 38% are black, and more than 700 individuals have diabetes mellitus (95% have type 2).
Data for this study were collected as part of a larger diabetes cohort study intended to identify associations between health literacy test scores and diabetes outcomes. Patients with a prior diagnostic code for diabetes were recruited either in person at the office or by telephone a few days before a visit. Identified patients were asked to participate in a reading and diabetes knowledge survey. Recruitment and interviews for this study took place between August 2006 and July 2007.
After obtaining informed consent, trained research assistants interviewed participants, asking them for their age, sex, race, and highest education level completed. Participants were then asked 3 questions related to reading ability: (1) “How would you rate your ability to read?” (self-rated reading ability); (2) “On a scale of 1 to 10, where 1 is ‘not at all’ and 10 is ‘a great deal,’ how much do you like reading?” (reading enjoyment); and (3) “How often do you need to have someone help you when you read instructions, pamphlets, or other written material from your doctor or pharmacy?” (SILS).37 We selected these 3 questions for analysis because each was thought to capture some different aspect of what may predict an individual’s literacy level. A Likert scale was used to code answers to the first and last items.
Participants were then screened for visual acuity using a floating E eye chart. Those with sufficient visual acuity were administered the Short Test of Functional Health Literacy in Adults (S-TOFHLA).40 The S-TOFHLA is a 36-item multiple-choice test of reading comprehension that can be completed in 7 minutes. It is composed of 2 passages that have several missing words. Participants are asked to select the words that best fit into the passages given the context of surrounding words. This test is a health literacy metric that has good correlation with the full TOFHLA from which it was derived and has become the more popular of the 2 tests. After the interview, participants were awarded $15 grocery store gift cards for their participation.
For our analysis, we excluded participants if their visual acuity was less than 20/50 corrected, if they could not communicate with research staff in English, or if they had an obvious cognitive impairment that would interfere with testing (such as known dementia or mental retardation).
Analysis
We analyzed the data using binomial logistic regression analysis. An S-TOFHLA score of less than 23 (a generally accepted cut point) was regarded as a positive outcome, indicating limited health literacy. Limited health literacy thus included both marginal health literacy (a score of 17–22) and inadequate health literacy (a score of ≤16).
Covariates included for analysis were sex, race, age, highest education level attained, and responses to the self-rated reading ability, reading enjoyment, and SILS questions. If the patient reported more than 1 race, their race was coded as “other” for analysis. Answer options for the questions, as well as the codes applied for all of the variables, are found in Table 1⇓. For health literacy level, derived from the S-TOFHLA score, an adequate level was coded as 0, whereas marginal and inadequate levels were both coded as 1. In the regression analysis, for ordinal variables, we treated each difference between adjacent response options as a 1-point increase. Patients’ enjoyment of reading was coded using the original scale from 1 to 10.
We examined the distribution of answers for each variable, and performed analyses for crude associations with limited health literacy using exact tests for categorical independent variables and t tests or Wilcoxon rank-sum tests for continuous and ordinal variables. Each variable was then tested as a lone independent variable in a logistic regression model predicting limited health literacy. From these variables, we chose an initial univariate model using the lowest Akaike Information Criterion as a guideline for model superiority. This process was repeated, adding 1 variable at a time to build the multivariate model until there was a rise in the Akaike Information Criterion value for all new models being tested. Although a recent analysis by Vittinghoff and McCulloch41 suggests that the current data set could reliably support a model with up to 6 variables, we checked the stability of the model at each step of this process.
We checked for confounding by demographic variables and plausible interactions by comparison with the main effects model. We considered interactions to be significant at an individual type I error rate of .05, and confounding to be substantial if parameter coefficients changed by greater than 15%. Linearity in the logit was assessed using the fractional polynomial method for determining acceptable transforms.42 Goodness of fit was assessed using the Hosmer-Lemeshow test. The area under the receiver operating characteristic (ROC) curve was also calculated. To detect the presence of outliers, we obtained the Δ-β (change in coefficient) and Δ χ2 (change in significance) for each subject and plotted these values against predicted probability in the model. We considered variables included in the final model to be significant at an individual type I error rate of .05. Wald and likelihood ratio statistics were compared for model sensitivity and stability, but only results of Wald statistics are reported. All analyses were run using Stata SE, version 9.2 (Stata Corp, College Station, Texas).
RESULTS
Of the 396 patients invited to participate, 225 (57%) completed the interview and were included in analyses. Of the 171 who were not included, 13 (3% of those invited to participate) were found to have vision less acute than 20/50 during the interview, 99 (25%) declined to participate, and 59 (15%) were excluded before the interview (6 reported being legally blind, 14 were no longer patients at the office where the study was being conducted, 27 did not speak sufficient English, 4 had considerable dementia, 5 had considerable mental retardation, 2 refused to sign the Health Insurance Portability and Accountability Act waiver for the larger study, and 1 had gestational diabetes only). Of the 99 patients who declined to participate, 54 had demographic information available. Compared with the 238 patients who participated (regardless of visual acuity), these nonparticipants did not differ significantly by race (P = .85) or sex (P=.42); however, participants were on average 6.2 years younger (P=.002). This difference was attenuated but remained significant if age in the medical chart was used instead of self-reported age (difference of means: 5.0 years, P=.01). One patient failed to indicate how much he enjoyed reading, and this value was imputed as the median of the sample (8 out of 10). Otherwise, the data set used was complete. Distributions of participants’ responses and P values of crude associations of the variables studied with limited health literacy are shown in Table 1⇑.
On the basis of S-TOFHLA scores, 15.1% of participants had limited health literacy (14, or 6.2%, had marginal health literacy and 20, or 8.9%, had inadequate health literacy). The final model for predicting limited health literacy is given in Table 2⇓, and its development and properties are described below. Lower self-rated reading ability, lower educational attainment, and more frequent need for help with written health materials were all independently associated with limited health literacy. Male sex and nonwhite race were independently associated with this outcome as well.
The main effects model included the same variables as the final model. We excluded reading enjoyment from the model because its association with adequate literacy switched from positive to negative after adjusting for the other screening questions. This finding indicated that although reading enjoyment accounted for some residual variability, leaving it in the model would falsely amplify the effect of other variables. We combined black race and other race after fitting the model because their coefficients were very similar and did not warrant the addition of a second categorical variable for race. Four plausible interactions were tested: sex with self-rated reading ability, sex with SILS result, sex with race, and race with self-rated reading ability. None of these interactions were significant at the predefined level, and no confounding by age was observed. Multicollinearity was not a problem; the maximum variance inflation factor among explanatory variables was 1.36 (mean, 1.22). All ordinal variables were linear in the logit analysis (Table 2⇑), and the Hosmer-Lemeshow test yielded a P value of .19, indicating no significant lack of fit.
We noted that 2 participants had outlying results. These participants’ records were rechecked for accuracy, and the interviewer was asked to verify that the data were correct. We determined that these participants were representative of the participants generally, so we did not drop any of the outlying observations. This model had excellent discriminatory performance, with an area under the ROC curve of 0.9212 (Figure 1⇓). Sensitivities, specificities, and likelihood ratios of the model at various cutoffs are given in Table 3⇓. Depending on the probability cutoff value chosen, sensitivity of the model could range from 100% down to 49%, with a corresponding rise in specificity from 50% to 98%.
DISCUSSION
A strength of this study is that it is clinically based and provides health care workers with tools for better understanding their patients’ learning styles. Whereas most literacy screening tools can determine whether or not a patient has limited health literacy, they do little to define etiologies or direct management. Asking specific questions about how an individual understands health information will better elucidate interventions that can be used. For example, asking about self-rated reading ability may reveal a known diagnosis of dyslexia, which could subsequently be managed by teaching with pictures or reading aloud to patients. A positive SILS result may indicate that a patient should have a family member present during education sessions. At the very least, asking these questions can prompt a discussion about an individual’s learning styles and coping methods. It will also prevent clinicians from being ignorant of problems associated with limited health literacy, so that when more effective interventions become available, clinicians will understand their usefulness and be ready to use them.
Another strength of this study is that it combines several different screening questions and demographic information into 1 predictive model. Of note, self-rated reading ability was the single most reliable predictor of limited health literacy of the predictors tested; thus, although it is known to be biased, it is still the best stand-alone question we have for determining literacy level. That being said, the fact that self-rated reading ability, SILS result, highest education level attained, sex, and race were all significant predictors of limited health literacy even after adjustment for one another serves to clarify that they are independently helpful at predicting limited health literacy, and that the use of all these questions and demographics is superior to the use of any one of them.
Recommendation
We recommend that clinicians who wish to screen for limited health literacy ask about self-rated reading ability and highest education level attained, and use the SILS as part of a thorough social history. To help clinicians remember what specific questions to ask and which answers may be considered suspicious for a health literacy problem, we propose the mnemonic “SOS” (Table 4⇓). According to this mnemonic, an educational attainment of high school or less, a self-rated reading ability of “okay” or worse, and asking for help with reading health materials at least “sometimes” are all associated a higher likelihood of limited health literacy. Clinicians can use this mnemonic as a framework for discussing their patients’ potential barriers to learning.
Limitations
This study has several limitations, foremost of which is that the participants interviewed were all being treated for diabetes at a single academic family practice center. It is also unknown why patients who participated in this study were generally younger than patients who refused to participate. The generalizability of the results presented here should be tested in a more representative sample of patients nationwide. Also, because the patients knew they were going to be given reading tests, they may have been less likely to attempt to conceal a reading problem. This association would effectively remove a reporting bias that might be present outside of the study population. Our model should be validated to determine how well these results apply to other populations or persons who are present in offices for reasons other than literacy testing.
Another limitation of this study was the use of the S-TOFHLA as the reference standard for health literacy. Although the S-TOFHLA correlates well with the full version of the TOFHLA and the Rapid Estimate of Adult Literacy in Medicine (REALM), none of these is a comprehensive measure of health literacy as defined earlier.1 In fact, no simple reading test could realistically be a complete measure of the common definition of health literacy, which includes such skills as Internet or visual literacy.
For some of the questions used in this study, patient embarrassment may be of concern.7,43 We believe that the literacy questions are far less intrusive than common health-related questions concerning drug abuse or sexual behaviors, however, and may likewise provide information necessary for improving patient care. The general consensus among our research assistants was that these questions did not cause embarrassment. Other studies also indicate that patients are willing to divulge their literacy status and consider it to be important clinical information.44,45 Of course, as with any condition discovered in a physician’s office, limited health literacy should be kept confidential.
Self-rated reading ability, highest education level attained, and the SILS result can each provide clinicians with valuable information about a patient’s learning needs. Clinicians are advised to be aware of these associations and know what questions can help them identify patients who may need assistance with navigating the health care system or understanding health-related materials.
Acknowledgments
The authors would like to thank Andrea Jeppesen, Benjamin Hull, Stephen Wilkes, Mark Stevens, Charlie Pizanis, Maggie Ryan, Alan Sanderson, Ryan Hackett, David Williams, Julia Budde, Pooja Lahoti, Patricia S. Hatton, Philip Binkley, Joan Allen, and the very helpful staff at The Ohio State University Rardin Family Practice Center.
Footnotes
-
Conflicts of interest: Dr Miser is on the Speaker’s Bureau for Pfizer Corporation. Mr Jeppesen and Dr Coyle have no conflicts of interest to declare.
-
Funding support: This work was funded by The Ohio State University Crisafi-Monte Primary Care Cardiopulmonary Endowment, and by grant T32RR023260 from the National Center for Research Resources (NCRR), a component of the National Institutes of Health (NIH).
-
Disclaimer: This publication was made possible by The Ohio State University Crisafi-Monte Primary Care Cardiopulmonary Endowment and the NIH through the NIH Roadmap for Clinical Research. Its contents are solely the responsibility of the authors and do not necessarily represent the offficial views of Crisafi-Monte or NIH.
This work was previously presented as the Most Outstanding Research Presentation, Ohio Family Medicine Symposium on Research & Education, April 12, 2008, Newark, Ohio; and as a Distinguished Research Paper at the Society of Teachers of Family Medicine 41st Annual Spring Conference, April 30 through May 4, 2008, Baltimore, Maryland.
- Received for publication February 22, 2008.
- Revision received July 9, 2008.
- Accepted for publication July 24, 2008.
- © 2009 Annals of Family Medicine, Inc.