Keywords
health literacy, critical health literacy, evidence-informed decision-making, evidence-based practice, critical thinking, public health, shared decision making, public understanding of science
health literacy, critical health literacy, evidence-informed decision-making, evidence-based practice, critical thinking, public health, shared decision making, public understanding of science
We have revised the manuscript and considered all suggested feedback carefully.
The methods section has been revised to improve the description of the content and development of the questionnaires. Furthermore, the wording of the abstract and the results section has been revised to improve clarity. Typos were corrected were noted.
See the authors' detailed response to the review by Julia Lühnen
See the authors' detailed response to the review by Gerardine Doyle
Enabling people to make informed decisions about healthcare by improving their ability to think critically about such claims is an important public health initiative1. Health literacy has been defined in many ways. Commonly this includes functional, interactional, and critical health literacy skills2. Health literacy research efforts in Norway and elsewhere have mainly been directed at providing reliable health information and education in functional literacy, focusing on improving understanding of medical terminology and self-management of health conditions3. Very little research has targeted critical health literacy, focused on enabling members of the public to make informed healthcare choices4–6.
Training in critical thinking about treatment decisions includes courses in evidence-based practice and education in basic research methodology. Such training is usually directed at health professionals. Despite the importance of enabling patients and the public to make informed decisions about treatment choices, such training has rarely been offered4–8.
Over the past decade, there has been increasing interest in enabling patients and the public to think critically about healthcare4,5,9–14. For example, in Norway three websites have been developed that aim to empower patients and the public to assess the trustworthiness of health claims15–17. Critical thinking has also been the focus of popular television shows that target common claims about treatment effects and illustrate how such claims can be tested using rigorous study designs. In collaboration with the Norwegian national television network (NRK), we carried out randomised trials with the aim of educating the public about the need for fair comparison of treatments18–22. More recently, we have sought to improve critical thinking in school children through the Informed Health Choices (IHC) project, an international collaboration of researchers in Uganda, Rwanda, Kenya, the UK, and Norway. We developed an educational intervention to teach primary school children to assess healthcare claims and an educational podcast for their parents. Both interventions were shown to be effective in randomized trials in Uganda23,24.
Few studies have been done to evaluate this ability in the general public to assess the trustworthiness of claims about the effects of healthcare and to make informed health choices6. Most such studies have relied on self-assessment of critical health literacy skills. One such project is the HLS-EU survey, mapping health literacy skills in several European countries25.
The Claim Evaluation Tools item bank was originally developed in English for use in the two randomised trials described above23,24. Since the item bank was first created items have been translated and validated in several settings and languages, including Uganda (in Luganda and English), Mexico, and China26–30.
Our starting point for developing the educational interventions and the Claim Evaluation Tools item bank was a list of Key Concepts that people need to understand and apply to assess claims about treatment effects and make informed health choices (hereafter referred to as Key Concepts)31 (Table 1). We define “treatment” broadly to include any intervention (action) intended to improve health, including preventive, therapeutic and rehabilitative interventions, and public health or health system interventions. The list of concepts provides a framework for researchers, teachers and others to develop interventions and measure people’s ability to assess treatment claims and make informed choices. We review and amend the list yearly9.
The item bank is revised according to the Key Concepts list yearly. When we conducted this study, it included four multiple-choice questions (MCQs) for 32 of the Key Concepts shown in Table 1. The item bank also includes questions that assess intended behaviours and self-efficacy. The item bank is an open-access resource. Teachers and researchers can select items for questionnaires tailored to their specific needs.
To our knowledge only one survey has attempted to measure the ability to understand and apply any of the IHC Key Concepts in a representative sample of Norwegian adults32. That study only addressed four of the Key Concepts. The purpose of this study was to map the ability of Norwegian adults to assess treatment claims and make informed health choices, using MCQs from the Claim Evaluation Tools item bank. The findings can be used to inform the development of learning resources and communication of information to patients and the public, and for international comparisons. We also wanted to take the opportunity to compare the findings of this study with the results of our previous Ugandan study. Despite any differences that may exist between these two contexts, it is our experience from previous projects that the challenges associated with assessing claims in everyday life in Uganda and Norway are very similar. However, knowledge of specific Key Concepts and overall understanding may differ. This study will consequently provide us with comparative information of understanding of the concepts in Uganda and Norway, but also explore any differences in understanding for specific Key Concepts. Furthermore, the study in Uganda included children and thus offered us the opportunity to compare our results from the adult population in Norway with those of children in Uganda.
To map the ability of Norwegian adults to assess treatment claims and make informed choices, and to compare these results to the findings of two studies in Uganda23,24.
This study was considered for ethical approval by the Norwegian Institute of Public Health. This project was considered to not to require a full ethical review (reference 18/11854), because no sensitive data was included. All information was collected through Nettskjema (a web-based survey system), ascertaining a high level of data security and safety.
The population was drawn from the National registry who provided us with a CD-ROM with respondents’ addresses. This CD was destroyed once data collection was terminated.
All respondents were given information about the purpose of the study and how the results would be managed and presented. To gain access to the questionnaire, each participant had to provide written consent electronically. The questionnaire was anonymous and once submitted, the information could not be traced back to the respondent.
The Claim Evaluation Tools item bank was originally developed in English for use in Uganda. For this study, a language specialist (KFO) translated all items to Norwegian. Three researchers trained in evidence-based medicine and epidemiology reviewed this translation (AD, ADO, and Atle Fretheim). The questions were not modified other than changing some of the names of people in the scenarios to more familiar names in Norwegian. For example, “Dr. Acheng” was changed to “Dr. Anker” (a more familiar name in Norwegian). The original English version was developed using plain language and this was also an important concern in the Norwegian translation.
All MCQs in the Claim Evaluation Tools item bank include a scenario with a treatment claim, and response options, with one answer being the “best” (correct) and the remaining options considered “worse” (incorrect). An example of a multiple-choice question is shown in Box 1.
Judith wants smoother skin. The younger girls in her school have smoother skin than the older girls. Judith thinks this is because the younger girls use cream on their skin to make the skin smoother.
Question: Based on this link between using cream and smooth skin, is Judith correct?
Options:
A) It is not possible to say. It depends on how many younger and older girls there are
B) It is not possible to say. There might be other differences between the younger and older girls
C) Yes, because the younger girls use cream on their skin and they have smoother skin
D) No, Judith should try using the cream herself to see if it works for her
In addition to knowledge questions, the item bank also includes two sets of questions scored on likert scales evaluating intended behaviour and self-efficacy associated with assessing claims.
For this study we decided to include two MCQs for all of the 32 Key Concepts. We also included the question sets for intended behavior and self-efficacy. Considering the large number of questions, we decided to split the questions across four questionnaires (see Additional file 1 and Additional file 5, Extended data)33,34.
We considered four of the 32 Key Concepts to be amongst the most important Key Concepts that can be understood and used by children, as well as adults. A set of 8 MCQs addressing these four Key Concepts were therefore included in all four questionnaires:
An outcome may be associated with a treatment but not caused by it.
The results of one study considered in isolation can be misleading.
Widely used treatments or those that have been used for decades are not necessarily beneficial or safe.
Comparison groups should be as similar as possible.
In order to compare the findings of the Norwegian study with the Uganda study, the first questionnaire was designed so that it was identical to the questionnaire used in Uganda and included 27 questions23,24. The second questionnaire included the remaining items that address concepts in the first (Claims) category as well as items from the second (Comparisons) category (24 questions) (Table 1). The third questionnaire included items from the third (Choices) category. This questionnaire also included three items evaluating intended behaviours regarding assessing treatment claims and agreeing to participate in a study evaluating treatments, and four items regarding self-efficacy (21 questions in total). The fourth questionnaire included MCQs addressing Key Concepts that we judged to be more difficult, mostly belonging to the Comparisons category (25 questions).
All four questionnaires also included the same set of questions regarding the participant’s sex, level of education, health professional background, training in research methods, and prior participation in research.
Based on evidence from two systematic reviews that evaluated strategies to improve participation in research35,36, we developed attractive postcards personally addressed to each potential participant inviting people to take part in the study. The postcard is shown in Figure 1. The postcard included a short description of the purpose of the study as well as a URL (“link”) to one of the four online questionnaires. Participants received full information about the study when they accessed the website and were asked to provide informed consent. Evidence also suggests greater participation rates can be expected if people are informed how the research may benefit them. This was stated on the postcard and the website provided information about how the study results can be accessed.
In January and February 2019, we mailed postcards to a representative sample of 4500 adults (≥18 years) living in Norway. The sample was provided to us by the Norwegian National Registry, considering level of education, sex and geographical spread37. The questionnaires were administrated electronically using Nettskjema, a service provided by University of Oslo38. In addition to the first postcard, one reminder postcard was sent out to each person.
The questionnaires were evaluated for their psychometric properties as part of this study. Although some MCQs were tested in Norway as part of a previous pilot, this is the first full-scale psychometric evaluation of the complete available battery of the Claim Evaluation Tools Item bank26,27.
Rasch analysis is a dynamic and practical approach to address important measurement issues required for validating an outcome39–41. In this study we followed the fundamental steps of Rasch analysis including testing for internal construct validity (multidimensionality), invariance of the items (Item-Person Interaction), and item bias (differential item functioning), as well as testing for reliability40,41. Rasch analysis can be used for both dichotomous and polytomous data40,42,43. Raw data were exported from the electronic data collection service (Nettskjema) as Excel-files and entered into RUMM2030 for Rasch analysis. Results can be replicated using the open-access software WINSTEPS44. Each questionnaire set was evaluated separately. Questionnaires 3 included also polytomous questions assessing intended behaviour and self-efficacy questions. Therefore, for questionnaire 3 specifically, additional analysis was done for each item block separately (MCQ- item block and intended behaviour and self-efficacy questions-block respectively).
Each MCQ was scored as correct or incorrect. The Key Concepts for which there were two MCQs were scored as “understood” if a participant answered both MCQs correctly. For questions about intended behaviours and self-efficacy, we dichotomised the responses in the analysis, for example as likely (very likely or likely) vs not likely (very unlikely, unlikely, or don’t know). As previously mentioned, the Key Concepts are reviewed and revised annually. Over the duration of this study, two Key Concepts were revised into the new Key Concept 1.1a. Considering that the MCQs were no longer appropriately covering this concept, these MCQS were therefore taken out of the analysis.
Based on a previous survey we conducted32, we anticipated that women and respondents with higher education levels would be more likely to respond. To address such non-random non-response, we used iterative post-stratification to match marginal distributions of sex and educational attainment level of the sample to the Norwegian population45. We also adjusted for region of residence. Due to an error, we did not collect data on participants’ age and could not use it for post-stratification, as planned.
We obtained Eurostat data on the marginal distributions of sex and region of residence for all Norwegians, and educational attainment level for Norwegians aged 15 to 64 years46,47. Participants reported their county of residence. We mapped counties to the corresponding Nomenclature of Territorial Units for Statistics (NUTS 2) regions of Norway used by Eurostat48. Participants reported the highest level of education they attained. We mapped these to the corresponding International Standard Classification of Education (ISCED) 2011 categories used by Eurostat (levels 0–2, 3–4, and 5–8). We used multiple imputation with chained equations to account for the uncertainty introduced by missing values of the post-stratification variables sex, region of residence, and educational level49, and iteratively post-stratified each imputed data set. Based on these adjustments, we estimated the percentage of the Norwegian population that understands each key concept, and that responded positively for each question about intended behaviours and self-efficacy. We used the R packages tidyverse, mice, mitools, and survey.
We present summaries of the results for participants and the post-stratified population estimates for understanding of the 30 Key Concepts, and the intended behaviours and self-efficacy questions. We quantified the uncertainty of our estimates using 95% confidence intervals and protected the family-wise coverage probability of the confidence intervals for the four Key Concepts included in all questionnaires via Bonferroni-correction (i.e., we report a 98.75% CI for each of those concepts). For each Key Concept, we calculated the likelihood of answering both questions (or the one question for two Key Concepts) correctly if participants randomly guessed the answer. These probabilities vary between six and 25%, depending on the number of MCQs and the number of response options (between two and four) for each MCQ.
We compared Norwegian and Ugandan adults’ understanding of the Key Concepts to that of Ugandan children in the intervention arm of our randomized trial. We compared mean test scores, and probabilities of achieving passing and mastery scores, using estimates from one year after the IHC primary school intervention for the Ugandan children and their teachers50,51 and estimates from the control group for parents in the IHC podcast trial, one year after parents in the intervention group listened to the podcast51,52. We used predetermined thresholds of at least 13 of 24 questions answered correctly for a passing score and at least 20 of 24 for mastery.
We estimated mean scores and odds ratios using generalized linear mixed-effects models (GLMMs; normal errors and identity link for mean scores, binomial errors and logit link for passing and mastery) using the lme4 R package. In the trials, children and teachers were randomized in clusters (schools), while parents were individually randomized. We modelled this clustering structure as a random intercept for each randomized unit. We did not adjust for covariates such as those used for stratified sampling in the Ugandan trials because those variables are not defined for all samples (e.g., school ownership was used in the trial on Ugandan children, but there is no analogous concept for Norwegian adults). No data were missing. It is not possible to use the lme4 package to apply post-stratification weights for the Norwegians.
We conducted exploratory analyses to investigate associations between understanding each Key Concept and the demographic covariates sex, research training, research participation, education level, and health professional background. Based on findings from our previous survey32, we hypothesised that better understanding of the Key Concepts would be associated with having a research background or a higher education level; that there would be little difference between health professionals and others; and that there would be little difference between women and men. We used generalized linear models (GLMs; quasibinomial errors and logit link) as before, modelled the covariates as categorical variables, and used multiple imputation and post-stratification as in the main analysis.
We used data from the first questionnaire to perform exploratory analyses to investigate how Norwegians’ mean scores and achievement of passing and mastery scores are associated with the demographic covariates sex, research training, research participation, education, and health professional background. We used GLMs (normal errors and identity link for mean score; quasibinomial errors and logit link for passing and mastery) to model the outcomes in terms of the covariates, which were modelled as before. Multiple imputation and post-stratification were used as in the main analysis. The variables included in imputation were the post-stratification variables (sex, region of residence, and educational attainment); demographic variables that coded for whether participants had research training or a health professional background, and whether they had been a research participant; and mean score. We did not include passing and mastery in imputation, as they can be calculated from the mean score.
We performed three power analyses. The first two analysed power to estimate differences, of at least 5% from random guessing, in the proportion of the Norwegian population that understands the concepts. The first analysis focused on the four Key Concepts that would be included on all four questionnaires while protecting the 95% coverage of the four confidence intervals as a family, and the second analysis focused on a single prototypical Key Concept that would be included on only one questionnaire. We assumed that each concept would be probed by two MCQs, each having four options, such that we would expect that 6.25% of Norwegians would appear to “understand” a concept if all participants guessed at random. We further calculated confidence interval widths, aiming to estimate proportions with precision no worse than ±0.05 (i.e., ±5%) for those four Key Concepts and no worse than ±0.08 (i.e., ±8%) for the other Key Concepts. The third analysis estimated power to detect a difference of at least 10% in the mean score (proportion of correct answers) between Norwegians and Ugandan children, parents, and teachers. Finally, drawing on experience with our previous survey, we explored a range of response rate scenarios.
These analyses suggested that sending a total of 4500 postcards would be sufficient to achieve a margin of error no greater than ±5% and ±8% for the first two analyses, and approximately 97% power to detect differences greater than 10% in the mean score compared to results from the Ugandan trials.
Table 2 shows participant characteristics across the four questionnaires53. Of the 771 respondents, 21 (2.7%) did not provide data on at least one of the covariates sex, research training, research participation, education, health professional background, and county of residence.
The log describing the Rasch analysis is described in more detail in Additional file 2 (Extended data)54. Overall, the four questionnaires showed acceptable fit to the Rasch model. Three out of four questionnaires were found to be unidimensional. The questionnaires including intended behaviour and self-efficacy items suggested that this sub-scale may include more than one dimension. Separate analyses of the two item-sets in Questionnaire 3 (Key Concept MCQs and intended behaviour and self-efficacy items) suggest that the MCQ sub-test works very well, with no apparent validity issues. Overall, the analysis of the sub-test consisting of items evaluating intended behaviour and self-efficacy items also shows satisfactory fit. However, two of the intended behaviour items show disordered thresholds. We did not observe any important local dependency in Questionnaire 2, 3 and 4. However, in Questionnaire 1, twelve correlations was observed. Two questionnaires were underpowered for this analysis (Questionnaire 1 and 4). However, based on the available data, most MCQs in all questionnaires showed acceptable fit to the ICC-curve, and no important differential item functioning by gender was identified in any of the questionnaires. Five MCQs showed evidence of under discriminating with one ability class deviating: two MCQs in Questionnaire 1 (q11 and q19.1), two MCQs in Questionnaire 2 (q3 and q18.3), and one MCQ in Questionnaire 4. In the two questionnaires that were adequately powered, Cronbach’s alpha was 0.6 and 0.7, respectively.
The mean width of the confidence intervals for the four concepts common to all questionnaires was ±4.6% (range ±3.6% to ±5.9%). One concept (An outcome may be associated with a treatment but not caused by it) was not estimated to the desired precision. The mean width of the other confidence intervals was ±7.92% (range ±3.37% to ±11.9%; 15 of the 30 concepts had confidence intervals wider than ±8%).
Figure 2 shows estimates of the percentage of Norwegian adults who understand each Key Concept. The adjusted proportion of Norwegian adults who answered correctly five of the Key Concepts was > 80% and more than 50% understand 17 concepts. On the other hand, less than half answered correctly for 13 concepts, and for seven of those concepts, the proportions of correct answers were no better than if people had randomly guessed.
Intended behaviours and self-efficacy are presented in Figure 3. The results suggest that most Norwegian adults are likely to question the basis of treatment claims (75%; 95% CI 66–85%), find out if treatment claims are based on research evidence (70%; 95% CI 59–81%), and say “yes” if asked to participate in a study comparing treatments (67%; 95% CI 63–72%). However, only about one out of five Norwegian adults find it easy to assess the relevance of studies that compare treatments (21%; 95% CI 9–33%), assess whether a treatment claim is based on a study comparing treatments (18%; 95% CI 12–25%), find studies that compare treatments (18%; 95% CI 9–28%), or assess the trustworthiness of the results of studies that compare treatments (16%; 95% CI 6–26%).
The results for Norwegian and Ugandan adults are compared to results for Ugandan children in Table 3 to Table 5. The mean score for Norwegian adults who participated in the survey (Questionnaire 1) was 17% higher (95% CI 14–20%) than the mean score for Ugandan children in the intervention arm of the randomised trial of the IHC primary school resources one year after the intervention and similar to the mean score for Ugandan teachers in the intervention arm of the trial (Table 3)23. The scores for Ugandan teachers in the control group were similar to those of the children in the intervention group, and the scores of Ugandan parents in the control group of the IHC podcast trial24 were 16% lower (95% CI -19 to -13%).
Nearly all the Norwegian adults (209 out of 210) who participated in the survey and Ugandan teachers in the intervention group (77 out of 78) had a passing score (Table 4). This was 15% more than the Ugandan children in the intervention group. However, because only one Norwegian and one Ugandan teacher in the intervention group did not have a passing score, the estimated differences may be unreliable for those comparisons. The proportion of Ugandan teachers in the control group with a passing score (88%) was similar to that of the children (84%), and 47% fewer (95% CI -56% to -37%).
Three quarters (66%; 95% CI 56–74%) of the Norwegian adults who participated in the survey and 70% (95% CI 59–80%) of the Ugandan teachers in the intervention group had a mastery score (Table 5). Compared to Ugandan children in the intervention group, the Odds for achieving a mastery score by the Norwegian participants was 5.3 (95% CI 3.5–7.9%), whereas the Odds for achieving a mastery score was 0.24 (95% CI 0.15–0.4) for the Ugandan parents compared to the children.
We did not find strong associations between gender, health professional background, or having participated in research and how well participants did (Table 6). Having 1–2 years of tertiary school education (ISCED levels 5–8) was associated with a mean score that was 11% higher (95% CI 4.0–19%), and a higher likelihood of having a mastery score (OR 3.5; 95% CI 1 to 12). Norwegian men were more likely than women to state that they are likely to find out the basis of treatment claims (OR 3.0; 95% CI 1.1 to 8.2), and more likely to find it easy to assess the trustworthiness of the results of studies that compare treatments (OR 3.8; 95% CI 1 to 14). People who have participated in research may be more likely to find it easy to find research based on studies that compare treatments (OR 5.7; 95% CI 1.9 to 17). People with secondary school education (ISCED levels 3–4) may be less likely than people with no more than primary school education (ISCED levels 0-2) to find it easy to assess the trustworthiness of results of studies that compare treatments (OR 0.057; 95% CI 0.0049 to 0.68). Besides these associations, we did not find evidence suggesting that gender, education level, health professional background, or prior participation in research were associated with intended behaviours or self-efficacy. Associations between participant characteristics and their understanding of specific Key Concepts, intended behaviour and self-efficacy are reported in Additional file 3 (Extended data)55.
According to Statistics Norway, 24.7% of Norwegian adults have a primary school education, 41.7% secondary school education, 22.4% tertiary school education, 7.3% have a master’s degree, and 0.7% of the population have a PhD. As we anticipated, participants in our study had a somewhat higher educational level than the general population56. In Norway, approximately 18% of the population (15–74) have an educational background in health or welfare57. This is comparable to participants in this study.
Participants in our survey had a good understanding of the 12 Key Concepts that were addressed in our randomised trial of the IHC primary school intervention23. Their understanding was comparable to Ugandan teachers in the intervention arm of our randomized trial of a primary school intervention, and better than that of Ugandan children in the intervention arm of that trial, and teachers in the control group. It was also better than parents in the control group of our randomised trial of an educational podcast for parents of primary school children in Uganda24.
We estimate that more than 80% of Norwegian adults understand these five concepts:
Increasing the amount of a treatment does not necessarily increase its benefits and may cause harm.
Competing interests may result in misleading claims.
Personal experiences or anecdotes alone are an unreliable basis for most claims.
The people being compared should be cared for similarly apart from the treatments being studied.
Weigh the benefits and savings against the harms and costs of acting or not.
On the other hand, Norwegian adults appear to do no better than if they were to randomly guess the answers to questions about these seven Key Concepts:
Beliefs alone about how treatments work are not reliable predictors of the presence or size of effects.
Widely used treatments or those that have been used for decades are not necessarily beneficial or safe.
Comparison groups should be as similar as possible.
People’s outcomes should be counted in the group to which they were allocated.
Results for a selected group of people within a study can be misleading.
Deeming results to be “statistically significant” or “nonsignificant” can be misleading.
Considering that people who responded to our survey had a somewhat higher educational level than the general population, the ability of our respondents may be higher than in the general population56.
Based on self-report, most Norwegians are likely to find out what the basis of a treatment claim is and to find out if a treatment claim is based on research. However, they do not consider it easy to assess the relevance of a claim, whether it is based on research, or its trustworthiness. They also do not consider it is easy to find relevant studies.
In the previous Norwegian study, four Key Concepts were evaluated32. In both studies the majority responded correctly to the question “Weighing the benefits and harms”. However, in the present study, the respondents were less likely to respond correctly to “Relative effects of treatments can be misleading” (29% vs 64%) and “The use of p-values may be misleading” (19% vs 51.5%). In contrast, more responded correctly to “An outcome may be associate with a treatment but not caused by it” (64% vs 30%). It is difficult to interpret these differences, other than that these evaluations were done using different question sets. Consequently, one explanation may be difference in the questions’ difficulty level. The gender distribution in the studies were identical, however a higher percentage in the present study had at least one year of education beyond secondary school (68% vs 52%). We did not find gender or having a health professional background to be a good predictor of participants’ understanding of the Key Concepts in any of studies. This is consistent with other studies and findings that both health professionals and patients feel challenged finding, appraising, and applying relevant evidence for use in health decisions58–63.
There are few other studies to which we can compare these results6. The European health literacy survey included all domains of health literacy, responses were given as self-assessments, and results are reported as overall scores25. Consequently, comparison with our study is difficult. They concluded that a little more than one tenth (12%) of those surveyed had insufficient health literacy and almost one half (47%) had limited (insufficient or problematic) health literacy. Both our survey and the European health literacy survey examined associations with education and, unsurprisingly, found that people with higher education have higher health literacy skills. However, the European health literacy survey reported that this result varies by country.
The strength of this study is that we provide new evidence on people’s ability to assess treatment claims using multiple-choice questions. To our knowledge, few such objective surveys have been conducted in Europe. Our work complements studies in which participants assessed their own abilities.
Many conceptualisations of critical thinking exist. Not all are based on explicit criteria, and unlike the Key Concepts used in this survey, few if any are subject to annual revision64. The questionnaires we administrated were based on a framework that has been developed from methodological literature and input from multiple disciplines and people with methodological expertise65.
We validated the questionnaires using robust methods. Although the results of the Rasch analysis are promising (Additional file 2, Extended data)54, it suggests the potential for improvements. Across all four questionnaires, we found that only five MCQs warranted improvement. However, Rasch analysis of two of the questionnaires was underpowered, so the validity and reliability of these should be assessed again in future studies.
Our analysis of the questionnaire including intended behaviours and self-efficacy suggests that intended behaviour and self-efficacy items measure two different dimensions. This might be because intended behaviour is more complex and dependent on self-efficacy, values, knowledge, and other factors. Except for Questionnaire 1, we did not identify any important dependencies between items.
We did not collect information on age for this study, and thus the association between age and ability to assess treatment claims should be explored in future studies. In our previous study conducted in Norway, younger respondents had a higher proportion of correct responses and higher total scores. This may suggest that there may have been improvements over time in the ability of both health professionals and patients to assess treatment claims32.
Another possible limitation of our study is the low response rate, which we anticipated. Our sample was similar to the general population in terms of the percentage of health professionals, but people with a higher education were over-represented among respondents. We addressed this issue using prespecified post-stratification in the survey analyses but did not address this in the comparisons to Ugandans. The results of those analyses therefore cannot be assumed to apply to the general population of Norwegian adults.
The results of this study can inform the development and evaluation of educational interventions that address Key Concepts that Norwegians appear to poorly understand. Up to now, few such interventions have been evaluated. There is a need to evaluate interventions for health professionals as well as for the general public to help ensure that they can think critically about treatment claims and choices. The results also can inform the development and evaluation or strategies for improving communication of information about the effects of treatments by researchers, health professionals, and others.
Studies like this one in other countries would help to map similarities and differences in people’s abilities across different countries and settings. Such information could help to determine the extent to which interventions should be tailored to address different Key Concepts for different populations.
Health professionals and others who communicate health information should be aware that patients may not be able to think critically about treatment claims and may therefore struggle to process information necessary to making informed decisions.
Our Rasch analysis suggests that the MCQs can be used in Norway. Most MCQs performed well, and this evaluation is the first step in developing a calibrated item bank that can be used for Computer Assisted Testing. The results of our Rasch analysis can be used to improve the multiple-choice questions that did not perform well.
Norwegian adults’ understanding of Key Concepts for assessing treatment claims and making choices varies from over 80% who understand five Key Concepts to 20% or less who understand seven Key Concepts. We did not find strong evidence that gender, being a health professional, or having participated in research are associated with the ability to assess treatment claims, intended behaviours, or self-efficacy. We did find evidence that people with higher education have a better understanding of Key Concepts.
Understanding the need for systematic reviews of fair comparisons as the basis for trustworthy treatment claims has the potential to reduce waste and harm from trusting and acting on misleading claims, and from not trusting and acting on reliable claims. Future interventions should be tailored to address these essential Key Concepts as well as other Key Concepts for which there appears to be a lack of understanding.
Harvard Dataverse: Replication Data for: Raw data Norwegian Claim Study_2019. https://doi.org/10.7910/DVN/R3DHA553
This project contains the following underlying data:
With respect to the potential ability to triangulate identities, access to data is restricted. Ethical approval was granted with an informed consent including a statement where people were guaranteed to be anonymous. Prospective data users can contact the authors to access the data, which can be rejected or approved by the data owner based on certain conditions (must be for legitimate research purposes and requestors are required to sign a data use agreement).
Zenodo: multinormal/fhi.informed-health-choices-norway.2019: Version 1.0. https://doi.org/10.5281/zenodo.366996451
This project contains the following underlying data:
Figshare: Additional file 1. Claim Evaluation Tools_tests administrated in Norway_2019. https://doi.org/10.6084/m9.figshare.11439651.v133
Figshare: Additional file 2 Log Rasch analysis Norwegian Claim Study_2019.docx. https://doi.org/10.6084/m9.figshare.10109840.v154
Figshare: Additional file 3. Supplementary tables and figures_Norwegian Claim Study_2019. https://doi.org/10.6084/m9.figshare.11439915.v155
Figshare: Additional file 5. English versions of Claim Evaluation Tools_tests administrated in Norway_2019. https://doi.org/10.6084/m9.figshare.11652342.v134
Figshare: Additional file 4. STROBE_checklist_cross-sectional_Norwegian Claim Study_2019. https://doi.org/10.6084/m9.figshare.11439753.v266
Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0).
Source code available from: https://github.com/multinormal/fhi.ihc.norway.2019
Archived source code at time of publication: https://doi.org/10.5281/zenodo.360625951
License: MIT
We would like to thank Atle Fretheim for his feedback on the Norwegian translation of the Claim Evaluation Tools Item bank assuring face validity and applicability to the Norwegian context of the translated content. Professor Atle Fretheim, Centre for Informed Health Choices, Norwegian Institute of Public Health has given his permission for his name and affiliations to be included in this publication.
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
Is the work clearly and accurately presented and does it cite the current literature?
No
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
I cannot comment. A qualified statistician is required.
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
References
1. Nutbeam D, Lloyd JE: Understanding and Responding to Health Literacy as a Social Determinant of Health.Annu Rev Public Health. 2021; 42: 159-173 PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: public health and health promotion
Is the work clearly and accurately presented and does it cite the current literature?
Partly
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
I cannot comment. A qualified statistician is required.
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Partly
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Health Literacy, Value measurement in health care
Is the work clearly and accurately presented and does it cite the current literature?
Yes
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
I cannot comment. A qualified statistician is required.
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
Competing Interests: The reviewer is involved in the translation and validation of the IHC Claim evaluation tool for its use in German speaking countries.
Reviewer Expertise: evidence-based health information, informed decision-making, trainings in evidence-based medicine
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | |||
---|---|---|---|
1 | 2 | 3 | |
Version 2 (revision) 30 Jul 21 |
read | ||
Version 1 11 Mar 20 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Comments on this article Comments (0)