Skip to main content

Development and analysis of quality assessment tools for different types of patient information – websites, decision aids, question prompt lists, and videos

Abstract

Objective

Our working group has developed a set of quality assessment tools for different types of patient information material. In this paper we review and evaluate these tools and their development process over the past eight years.

Methods

We compared the content and structure of quality assessment tools for websites, patient decision aids (PDAs), question prompt lists (QPLs), and videos. Using data from their various applications, we calculated inter-rater concordance using Kendall’s W.

Results

The assessment tools differ in content, structure and length, but many core aspects remained throughout the development over time. We found a relatively large variance regarding the amount of quality aspects combined into one item, which may influence the weighting of those aspects in the final scores of evaluated material. Inter-rater concordance was good in almost all applications of the tool. Subgroups of similar expertise showed higher concordance rates than the overall agreement.

Conclusion

All four assessment tools are ready to be used by people of different expertise. However, varying expertise may lead to some differences in the resulting assessments when using the tools. The lay and patient perspective needs to be further explored and taken into close consideration when refining the instruments.

Peer Review reports

Introduction

Many patients in general and cancer patients in particular show a high need for information regarding their condition and treatment options [1,2,3]. With decisions of an often large scope and of far-reaching consequences, information and knowledge provide an essential basis for patients’ active involvement in the decision-making process [2, 4, 5]. While a personal conversation with the treating physician tends to be the preferred source of information [4, 5], evidence also suggests that patients may feel underinformed [6, 7] after consultations.

Consequently, many patients (as well as caregivers) resort to autonomous searches for additional information for which the internet has become a major source [5, 8,9,10,11]. One of the main challenges of potentially vital importance when searching for online information is to determine the quality and reliability of the search results [12], especially for people who have little experience with the subject itself or with the pitfalls of the internet in general. Most patients searching for information online are lay people regarding medical matters; when it comes to cancer patients, their average age is relatively high and, consequently, their average digital literacy often relatively low.

Clinicians, on the other hand, are often confronted with the problem of limited time opposing their patients’ need for in-depths information. Good-quality material to refer patients to is one possible way out of this dilemma. But again, it can be quite a challenge to determine how reliable the material is or how appropriate and understandable for the target audience.

The necessity for support in finding and assessing good-quality information online is evident; different sets of criteria exist in this regard, but they vary and are not commonly known to the average user. Authors and publishers of (online) health information only partly adhere to them [13,14,15,16,17,18], which may be due to a lack of awareness or to the criteria being too difficult to handle especially with regard to different types of patient information.

Websites in general provide a wide variety of types and quality of information. Authors range from established health care providers, individual practitioners, self-help groups, pharmaceutical companies to personal blogs and more. As a sub-genre of online information, videos provide a particular format which can be especially helpful for and is often preferred by people with literacy problems or simply those preferring audio-visual information to reading text on websites.

A more specific format designed to assist patients are patient decision aids (PDAs). They aim to provide the required information, which allows patients to assume an active part in the decision-making process regarding treatment, examinations or screenings.

Another tool for supporting patients’ information gathering are question prompt lists (QPLs). They are designed to aid communication between patients and physicians during consultation so that both sides receive the important information required for informed decisions and optimal treatment.

For all information formats, quality assessment to distinguish high-quality material from low-quality or even false and dangerous material is important. For written material there is a national and international consensus on quality criteria as specified in publications and instruments like afgis, DISCERN or HONcode [19,20,21,22,23]. These emphasise first and foremost the importance of up-to-date evidence-based information from transparent and unbiased sources which can be understood by lay people. For other formats, no such criteria and instruments exist, which consider the specialities of the respective format.

The aim of our working group therefore was, to develop a set of instruments for modern formats based on the consented criteria. Over the course of the past eight years (2014—2022), we have designed quality assessment tools for different types of (online available) patient information: websites [13,14,15,16], patient decision aids (PDAs) [article submitted to BMC Medical Informatics and Decision-Making], question prompt lists (QPLs) [17], and videos.

The aim of this study is to analyse the resulting instruments with respect to applicability and usefulness as well as inter-rater concordance. From there, we try to create a template of core criteria plus additional ones that can be used to evaluate the quality of different types of patient information.

Material and methods

Development and structure of the tools

To develop the respective instrument, we used a merged set of criteria developed in an expert consensus process from several renowned instruments, guidelines and assessment tools on written patient information. With the internet having become an increasingly important source of information, we started out with the development of criteria to assess the quality of websites containing patient information [13]. Based on those, the criteria for (online available) QPLs [17] and PDAs were developed next [article submitted to BMC Medical Informatics and Decision-Making]. The most recent tool eventually focussed on videos [article in progress].

For every format, we first systematically searched existing literature for instruments and criteria assessing the quality of patient information in the respective format, as well as general quality criteria for patient information. Items were then selected and adapted where necessary to fit the context – websites, PDAs, QPLs or videos. The resulting item lists were checked for completeness by at least three researchers, and context-specific items were added where necessary. For each format, a preliminary tool was developed and tested on two or three examples of the respective format by two to four different raters. Based on the results, the item lists and instructions were refined. After a second test run, the tools were finalized and then tested on a larger number of examples. Detailed information on the development process of each tool and on the incorporated sets of previously existing criteria can be retrieved from the respective publications.

Each resulting tool comprises a list of single items which can be rated on a three-point Likert scale: completely fulfilled, in part fulfilled, not fulfilled. The tools differ in the assigned values: some use values between 0 and 2, others between 1 and 3. The items are sorted in two categories: content criteria and formal criteria. The assessment tool for videos contains “user-oriented criteria” as a third category.

The tools differ in their methods of calculating the final results of the assessment. The original tool for websites only builds sums of the assigned values of each item. The tools for PDAs and QPLs add the option to use the mean values of each subcategory of items, as well as transforming the final result of an assessment into percentages. This last step allows for comparability if certain categories are not applicable to some of the publications which would otherwise leave those publications with a disadvantage in overall points.

Material

During the time from 2011 to 2022, these four assessment tools were applied to different information material. For each topic, we simulated a patient internet search and derived the top hits which then were assessed using the tool. Details of the search strategies and the information finally assessed are described in the respective publications.

  • Websites:

    • o 77 websites containing patient information on oncology, evaluated by two raters (one medical student, one physician) [13]

    • o 60 websites on cancer diets, evaluated by two raters (two students of medicine and medical informatics) [14]

    • o 50 websites on thyroid cancer, evaluated by four raters (one patient, two other lay people, one physician) [15]

    • o 96 German websites of oncologists and non-medical practitioners, evaluated by two raters (both medical students) [16]

    • o 40 websites containing information on diet and nutrition for cancer patients, evaluated by four raters (two medical students, two lay people) [article submitted to Archives of Public Health]

  • PDAs: 22 PDAs evaluated by four raters (two medical students, two physicians) [article submitted to BMC Medical Informatics and Decision-Making]

  • QPLs: 46 QPLs specifically for cancer patients. evaluated by four raters (two medical students, two physicians) [17]

  • Videos:

    • o 26 videos on complementary and alternative medicine (CAM) in oncology evaluated by four raters (two medical students, two physicians) [article in progress]

    • o 30 videos on diet and nutrition for cancer patients evaluated by four raters (two medical students, two physicians) [article in progress]

Statistics

In order to assess the agreement between different raters of the same material Kendall’s coefficient of concordance (Kendall’s W) was calculated, it being a relatively powerful alternative to the often used Cohen’s kappa, which is appropriate for nominal data but not for the ordinal data of these assessments [24]. Within each study, publications were ranked for each rater separately according to their final value. Kendall’s W was then calculated for each type of publication using IBM SPSS Statistics Version 27.0. Level of significance was set at p < 0.05. We calculated overall coefficients including all raters, as well as coefficients for subgroups where possible, differentiating between expert raters (physicians), semi-experts (medical students) and lay raters.

There is no universal interpretation of the resulting values of W, since interpretation depends on the field of application. In accordance with current literature and taking into account the complexity of most of the evaluated material, we classify concordance levels over 0.5 as acceptable, over 0.65 as reasonable, over 0.8 as strong, over 0.9 as very strong [25, 26].

Results

Comparison of the tools

Overall, the four assessment tools contained a lot of similarities with regard to the collected quality criteria. A detailed comparison of the tools is presented in Table 1. The website tool serves as basis for comparison since it was the first to be developed and the development of subsequent tools used it as a basis to varying degrees.

Table 1 Comparison of the four quality assessment tools based on the original tool for websites

A number of aspects stood out when comparing all four tools:

  • When assorting items into main groups, most tools differentiate between content criteria and formal criteria, but the tool for videos opened a third main category of user-oriented criteria.

  • Several items seem to be difficult to distinctly assign to either content or formal criteria; their assigned category differed across tools.

  • There were rather vast differences regarding the number of aspects combined into one item. While some tools contained fewer items comprising multiple questions, other tools separated those questions into multiple items. As a result, certain aspects may be weighted differently between tools regarding their impact on the final score and thus regarding their importance for a final judgement of good or bad quality.

  • The majority of items can be applied to different types of material. The tool for PDAs contained the largest number of specific items only applicable to this type of material; a few specific items were created for QPLs and videos as well.

Looking at the changes over time, the following key evolvements were noticeable:

  • Several items were altered to focus more on whether the material catered appropriately to its specific target group.

  • The original idea that patient information should not touch on areas without confirmed information was dropped to focus on the demand for open communication of missing evidence.

  • Additional criteria for web-based material were added while developing the tools for PDAs and QPLs, which would be equally applicable to and important for websites.

On the other hand, some of the original items from the website assessment tool were dropped in one or several of the later tools, even though they would have been applicable – like expertise of the authors/creators (dropped for videos), relevance of the information (dropped for PDAs), details on supplementary aids and additional information sources (dropped for PDAs and videos), focus on patient-relevant endpoints (dropped for QPLs and videos).

The apparent length of the tools differs considerably, which is due to a large extent to the different ways in which quality aspects were combined or separated into individual items, as can be seen in Table 2. Medium-specific items also contributed to this effect. A dissection of all four tools into individual quality aspects that are considered within the items is provided in additional file 1 (original German versions).

Table 2 Comparison of length

Inter-rater concordance

All tests of overall agreement between the raters were significant on a 5% level, meaning that the null hypothesis W = 0 (no agreement among raters) could be rejected with 95% certainty in all applications of the tool. When differentiating for subgroups among raters (physicians, medical students, lay people), one tests did not reach significance on a 5% level, only if allowing for a risk of error of 10%: agreement between physicians on QPLs (p = 0.097). It remains unclear whether the differences of the two ratings originated in factors pertaining to the tool, the source material, or the raters themselves.

Table 3 gives an overview of the different levels of concordance between the raters, calculated for subgroups of raters where possible.

Table 3 Inter-rater concordance of all past applications of the four assessment tools

In 7 out of 9 applications of the tool, all inter-rater concordances were of 0.6 or higher. The lowest concordance resulted from the assessment of websites on diet and nutrition for cancer patients (Kendall’s W over all 4 raters: 0.461 (total scores), 0.622 (content scores), 0.525 (formal scores), 0.578 (specific criteria of the German Nutrition Society—DGE)), where a high number of shared ranks complicates Kendall’s W as a measure for concordance.

The ratings of websites on thyroid cancer differed particularly between subgroups of raters: concordance between lay people was good at W = 0.739, while the other two raters consisting of one physician and one patient showed similarly good concordance at W = 0.725.

A general comparison of overall concordance rates to those within subgroups of raters showed that concordance within subgroups is almost always higher than the overall concordance. The exception occurred in the assessments of PDAs and QPLs, where concordance among physicians was lower than the overall concordance (physicians and medical students).

Discussion

The assessment tools for patient information which were regarded in this article, show some potential for their application not only within scientific research but also in clinical practice, possibly more so for medically trained than for lay people.

We found an overall good inter-rater concordance when tools were applied by medical students, physicians and lay people. Lowest concordance among all raters occurred for websites on thyroid cancer, websites on diet and nutrition, and oncological PDAs. For websites, these low levels are at least partly founded in differences between subgroups, as the concordance within those is substantially higher.

A pattern evolved of concordance within subgroups being usually better than the overall concordance – with the exception of concordance among physicians who evaluated PDAs and QPLs, which was relatively low (while that among medical students was very high). Each pair of physician raters was different, so the results cannot be attributed to differences in opinion of two particular people. This phenomenon needs to be studied in more detail: The physicians’ different views of the tools or the source material may provide valuable insight on how to assess the quality of patient information.

The only patient acting as rater in these studies had a higher concordance with the physician than with the other two lay raters. This result can easily be explained if one assumes that this particular patient had already gained considerable experience and knowledge about their disease. However, this is an isolated incident. At which point patients’ expertise resembles more that of a physician than that of lay people unaffected by a specific disease would need to be a subject of further research.

Another interesting result was the substantially lower concordance among the same raters (two medical students) when evaluating non-medical practitioners’ websites as opposed to oncologists’ websites. A possible explanation may be that the wide variety of topics listed and methods offered on these websites requires rather extensive knowledge in order to assess the quality of the information – a type of knowledge that is not typically taught in medical school, so the expertise of the raters on these particular subjects may have differed considerably in this regard.

Again, this is an isolated incident which calls for more in-depth research. Considering all of the described aspects, however, the question clearly arises how well people with less expertise on medical topics – which applies to the vast majority of patients – will be able use assessment tools like these or other previously published lists of criteria to adequately assess the quality of patient information. Especially PDAs may be difficult to evaluate even for medical experts, as their objective is not only to inform patients of medical topics but also to assist in an often difficult process of reaching a decision about their health. Evaluating their quality may require expertise not only in the particular medical field but also in the many intrapsychic and interpersonal processes involved in human decision-making.

Looking at the actual content of the tools, it may not be advisable to try and sort items into two large categories of content and formal criteria, since many aspects appear to touch on both areas. However, grouping items into smaller categories allows to control the weighting of certain aspects by the number of items within each category. Final values can thus either be calculated on the basis of individual items (categories with more items have a higher impact on the overall result) or on the basis of each category’s mean value (every category contributes equally to the overall result).

Other sets of criteria in international literature also tend to use main categories [20, 27,28,29]. Their number, contents and degree of detail, however, differ considerably. The difficulty of assigning items to clearly defined, selective categories appears to be a universal problem.

The differences between the tools in this review revealed some interesting developments over time. Some can be understood as a shift in focus which followed a certain ideal. For example, not dismissing topics without solid evidence allows the information material to educate about controversial or even potentially harmful topics which lack scientific evidence but may be of interest to patients. Leaving these topics out could lead to patients looking elsewhere for information and likely finding them in more problematic contexts. Including them may therefore even be beneficial; the quality criterion here is the clear communication of lack of evidence and what this means in terms of, for example, effectiveness and safety of a treatment.

Other changes were items added to the assessment tools for PDAs or QPLs which are applicable to websites as well and should be integrated into the website assessment tool for future use.

Some items of the original tool appear to have been dropped or shortened in order to streamline the tool or because these items would be difficult to assess or understand, especially for laypersons (e.g. whether field tests were carried out to evaluate the effectiveness of a PDA or QPL). Other items seem to have disappeared without externally apparent explanation.

Overall, some criteria seem to be essential when assessing information in order to ensure that no false information is provided and that patients are able to correctly understand and apply the information to their individual situation. Other criteria are of a more supplementary nature and add value to the material (such as providing further resources, feedback options for users etc.).

A final difficulty arises when applying one of the tools presented here to a single publication: No thresholds have yet been defined to differentiate between, for example, low, acceptable, and good quality. So far, transforming the final value into percentages offers the best option for an individual judgement on how well the assessed material meets the expectations. Again, this problem occurs with most other evaluation tools as well.

A 2005 review by Bernstam et al. [30] had already found a plethora of different evaluation criteria and tools but judged most of them unfit for practical use by patients or even by physicians. Since then, some lists of criteria have become more established, especially in research but also as recommendations for patients. But the usability of such tools for the latter tends to come at the cost of comprehensiveness. Restricting a tools length to make it more approachable will mean reducing the quality aspects it can cover or combining many of them into a few items, which risks reducing the discriminatory power and informative value. A possible solution to this dilemma may be the use of core criteria and supplementary once – the former constituting a strictly necessary set of criteria that determines whether a publication is trustworthy, the latter giving additional information about its value to users.

Research is needed regarding such tools’ sensitivity to a rater’s expertise. An in-depth analysis may reveal which quality aspects can be accurately evaluated independently of expertise, which items require detailed explanations or need to be broken down into several items in order to enhance inter-rater concordance across different levels of expertise.

Limitations

The result presented in this review are limited in their conclusiveness for a number of reasons. Methodically, the small number of raters in the original studies limits the conclusiveness of concordance rates. The differences in number and background (regarding expertise) of the raters in each study potentially reduce their comparability. Kendall’s coefficient of inter-rater concordance is based on ranks which ultimately results in a comparison of intra-individual benchmarks of the raters. The tools appear well reliable when used to compare several publications of patient information and ranking them. However, their reliability regarding the absolute value of an assessment’s result still needs to be confirmed, which would be the basis for an objective measure of the kind of a threshold above which a publication can be considered good quality.

Another difficult aspect are the limited applications by lay-persons. This was due mainly to the difficulty in recruiting oncological patients who are already pressed for time and energy, to partake in these often lengthy evaluations. On the other hand, researchers need additional resources when presenting lay people with such a vast amount of information as has been used in these studies, to afterwards explain the quality of the presented material and especially to point out any false information that may have been included, in case the assessment tools did not work sufficiently well for the test subjects. So far, raters of varying expertise but mostly from within the field of medicine in general and oncology in particular have used the tools. The results show potential but in order to evaluate the applicability of these tools in clinical practice, an effort needs to be made to recruit lay raters and patients as well as physicians without active involvement in scientific research, which we hope to achieve in future studies.

In this context, the raters’ original expertise on the topic should also be assessed, as well as their expertise at the end of the assessment process in order to identify potential learning effects which may influence the results.

While the current tools may be used to assess individual content without comparison by using the results in percentages of an expected optimum, another future approach should be to find an evidence-based consensus on what is considered low, acceptable, or good quality on these scales to be clearly specified with their instructions for use.

Conclusion

The assessment tools referred to in this article are ready to be used in practice, especially when comparing the quality of several publications of the same type. Researchers as well as clinicians and patients can employ them in their search for good-quality information material or to assess the quality of a given body of publications addressing patients.

They still leave room for improvement, though, regarding the optimal selection of criteria which would cover all relevant aspects without making the entire tool too complex and unwieldy. A suggestion for simplifying their use would be to find ways for integrating them into digital solutions.

The lay perspective is needed in order to assess how well these tools can really be applied in practice. To this end, a more extensive evaluation should be conducted which also assesses the raters’ original expertise on the subject of the information material. It is entirely possible, that especially for complex topics no assessment tool is sufficient to find good-quality information without first developing the user’s own expertise.

Availability of data and materials

The datasets analyzed during the current study is available from the corresponding author on reasonable request.

Abbreviations

CAM:

Complementary and alternative medicine

DGE:

Deutsche gesellschaft für Ernährung [German Nutrition Society]

PDA:

Patient decision aid

QPL:

Question prompt list

References

  1. Goerling U, Faller H, Hornemann B, Hönig K, Bergelt C, Maatouk I, Stein B, Teufel M, Erim Y, Geiser F, Niecke A, Senf B, Wickert M, Büttner-Teleaga A, Weis J. Information needs in cancer patients across the disease trajectory. A prospective study. Patient Educ Couns. 2020;103(1):120–6. https://doi.org/10.1016/j.pec.2019.08.011.

    Article  PubMed  Google Scholar 

  2. Papadakos J, Urowitz S, Olmstead C, Jusko Friedman A, Zhu J, Catton P. Informational needs of gastrointestinal oncology patients. Health Expect. 2015;18(6):3088–98. https://doi.org/10.1111/hex.12296.

    Article  PubMed  Google Scholar 

  3. Wieldraaijer T, Duineveld LAM, Bemelman WA, van Weert HCPM, Wind J. Information needs and information seeking behaviour of patients during follow-up of colorectal cancer in the Netherlands. J Cancer Surviv. 2019;13(4):603–10. https://doi.org/10.1007/s11764-019-00779-5.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  4. Heimer A, Henkel M. Bedarf an Krebsinformation in der Bevölkerung: Analyse des Informationsverhaltens von Ratsuchenden. Repräsentative Befragung im Auftrag des Deutschen Krebsforschungszentrums. Heidelberg: Akademische Verlagsgesellschaft AKA GmbH; 2012.

    Google Scholar 

  5. Weg-Remers S, Hiller B. Was Krebspatienten wissen wollen – und von wem. Krebs Foc. 2019;10:37–8.

    Google Scholar 

  6. Frenkel M, Ben-Arye E, Cohen L. Communication in cancer care: discussing complementary and alternative medicine. Integr Cancer Ther. 2010;9:177–85. https://doi.org/10.1177/1534735410363706.

    Article  PubMed  Google Scholar 

  7. Josfeld L, Keinki C, Pammer C, Zomorodbakhsch B, Hüber J. Cancer patients‘ perspective on shared decision-making and decision aids in oncology. J Cancer Res Clin Ocol. 2021;147:1725–32. https://doi.org/10.1007/s00432-021-03579-6.

    Article  Google Scholar 

  8. Graffigna G, Barello S, Bonanomi A, Riva G. Factors affecting patients’ online health information-seeking behaviours: the role of the Patient Health Engagement (PHE) Model. Patient Educ Couns. 2017;100(10):1918–27. https://doi.org/10.1016/j.pec.2017.05.033.

    Article  PubMed  Google Scholar 

  9. Muusses LD, van Weert JC, van Dulmen S, Jansen J. Chemotherapy and information-seeking behaviour: characteristics of patients using mass-media information sources. Psychooncology. 2012;21(9):993–1002. https://doi.org/10.1002/pon.1997.

    Article  PubMed  Google Scholar 

  10. Papadakos J, Bussière-Côté S, Abdelmutti N, Catton P, Friedman AJ, Massey C, Urowitz S, Ferguson SE. Informational needs of gynecologic cancer survivors. Gynecol Oncol. 2012;124(3):452–7. https://doi.org/10.1016/j.ygyno.2011.10.030.

    Article  PubMed  Google Scholar 

  11. Tustin N. The role of patient satisfaction in online health information seeking. J Health Commun. 2010;15(1):3–17. https://doi.org/10.1080/10810730903465491.

    Article  PubMed  Google Scholar 

  12. Heudel PE, Delrieu L, Dumas E, Crochet H, Hodroj K, Charrier I, Chvetzoff G, Durand T, Blay JY. Impact of limited E-health literacy on the overall survival of patients with cancer. JCO Clin Cancer Inform. 2022;6:e2100174. https://doi.org/10.1200/CCI.21.00174.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Liebl P, Seilacher E, Koester MJ, Stellamanns J, Zell J, Hübner J. What cancer patients find in the internet: the visibility of evidence-based patient information - analysis of information on German websites. Oncol Res Treat. 2015;38(5):212–8. https://doi.org/10.1159/000381739.

    Article  PubMed  Google Scholar 

  14. Herth N, Kuenzel U, Liebl P, Keinki C, Zell J, Huebner J. Internet information for patients on cancer diets - an analysis of German websites. Oncol Res Treat. 2016;39:273–81. https://doi.org/10.1159/000445861.

    Article  PubMed  Google Scholar 

  15. Kuenzel U, Monga Sindeu T, Schroth S, Huebner J, Herth N. Evaluation of the quality of online information for patients with rare cancers: thyroid cancer. J Cancer Educ. 2018;33(5):960–6. https://doi.org/10.1007/s13187-017-1173-z.

    Article  PubMed  Google Scholar 

  16. Schallock H, Bartmann B, Keinki C, Huebner J. Online information on oncologists’ and non-medical practitioners’ websites in Germany: a critical comparison. Patient Educ Couns. 2019;102(11):2038–48. https://doi.org/10.1016/j.pec.2019.05.022.

    Article  PubMed  Google Scholar 

  17. Josfeld L, Zieglowski N, Möller J, Keinki C, Hübner J. Development and application of a quality assessment tool for oncological question prompt lists. J Cancer Educ. 2023. https://doi.org/10.1007/s13187-023-02290-z.

    Article  PubMed  Google Scholar 

  18. Narwani V, Nalamada K, Lee M, Kothari P, Lakhani R. Readability and quality assessment of internet-based patient education materials related to laryngeal cancer. Head Neck. 2016;38:601–5. https://doi.org/10.1002/hed.23939.

    Article  PubMed  Google Scholar 

  19. Aktionsforum Gesundheitsinformationssystem (afgis). Afgis Transparenzkriterien. https://www.afgis.de/zertifizierung/transparenzkriterien/. Accessed 30 Aug 2022.

  20. Charnock D, Shepperd S, Needham G, Gann R. DISCERN: an instrument for judging the quality of written consumer health information on treatment choices. J Epidemiol Community Health. 1999;53(2):105–11. https://doi.org/10.1136/jech.53.2.105.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  21. Deutsches Netzwerk Evidenzbasierte Medizin. German Network for Evidence-Based Medicine, the Agency for Quality in Medicine. Gute Praxis Gesundheitsinformation. Berlin: 2015. http://www.ebm-netzwerk.de/gpgi.

  22. Health on the Net. HONcode Guidelines. https://www.hon.ch/en/guidelines-honcode.html. Accessed 1 Feb 2022.

  23. Steckelberg A, Berger B, Köpke S, Heesen C, Mühlhauser I. Kriterien für evidenzbasierte Patienteninformationen [Criteria for evidence-based patient information]. Z Arztl Fortbild Qualitaetssich. 2005;99(6):343–51.

    Google Scholar 

  24. Craigie M, Loader B, Burrows R, Muncer S. Reliability of health Information on the Internet: an examination of experts’ ratings. J Med Internet Res. 2002;4(1):e2. https://doi.org/10.2196/jmir.4.1.e2.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  25. Melinosky C, Kincaid H, Claassen J, Parikh G, Badjatia N, Morris NA. The modified fisher scale lacks interrater reliability. Neurocrit Care. 2021;35(1):72–8. https://doi.org/10.1007/s12028-020-01142-8.

    Article  PubMed  Google Scholar 

  26. Simpson E, Bissonnette R, Eichenfield LF, Guttman-Yassky E, King B, Silverberg JI, Beck LA, Bieber T, Reich K, Kabashima K, Seyger M, Siegfried E, Stingl G, Feldman SR, Menter A, van de Kerkhof P, Yosipovitch G, Paul C, Martel P, Dubost-Brama A, Armstrong J, Chavda R, Frey S, Joubert Y, Milutinovic M, Parneix A, Teixeira HD, Lin CY, Sun L, Klekotka P, Nickoloff B, Dutronc Y, Mallbris L, Janes JM, DeLozier AM, Nunes FP, Paller AS. The Validated Investigator Global Assessment for Atopic Dermatitis (vIGA-AD): the development and reliability testing of a novel clinical outcome measurement instrument for the severity of atopic dermatitis. J Am Acad Dermatol. 2020;83(3):839–46. https://doi.org/10.1016/j.jaad.2020.04.104.

    Article  PubMed  Google Scholar 

  27. Bunge M, Mühlhauser I, Steckelberg A. What constitutes evidence-based patient information? Overview of discussed criteria. Patient Educ Couns. 2010;78(3):316–28. https://doi.org/10.1016/j.pec.2009.10.029.

    Article  PubMed  Google Scholar 

  28. Elwyn G, O’Connor A, Stacey D, Volk R, Edwards A, Coulter A, Thomson R, Barratt A, Barry M, Bernstein S, Butow P, Clarke A, Entwistle V, Feldman-Stewart D, Holmes-Rovner M, Llewellyn-Thomas H, Moumjid N, Mulley A, Ruland C, Sepucha K, Sykes A, Whelan T. International Patient Decision Aids Standards (IPDAS) Collaboration. Developing a quality criteria framework for patient decision aids: online international Delphi consensus process. BMJ. 2006;333(7565):417. https://doi.org/10.1136/bmj.38926.629329.AE.

    Article  PubMed  PubMed Central  Google Scholar 

  29. LIDA. Minervation validation instrument for healthcare websites [online]. Available from: http://www.minervation.com/wp-content/uploads/2011/04/Minervation-LIDA-instrument-v1-2.pdf. Accessed 31 Aug 2022.

  30. Bernstam EV, Shelton DM, Walji M, Meric-Bernstam F. Instruments to assess the quality of health information on the World Wide Web: what can our patients actually use? Int J Med Inform. 2005;74(1):13–9. https://doi.org/10.1016/j.ijmedinf.2004.10.001.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

Open Access funding enabled and organized by Projekt DEAL. The authors declare that no funds, grants, or other support were received during the preparation of this manuscript.

Author information

Authors and Affiliations

Authors

Consortia

Contributions

Both authors worked on the conceptualization of this study and researched current literature. LJ analyzed the data and wrote the initial article draft. JH revised the texts. Both authors read and approved the final manuscript and agree to be accountable for all aspects of the work. The corresponding author has the final responsibility to submit for publication.

Corresponding author

Correspondence to Lena Josfeld.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors have no relevant financial or non-financial interests to disclose.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

Comparison of all aspects within the tools’ criteria.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Josfeld, L., Huebner, J. & On behalf of the Working Group Prevention and Integrative Oncology in the German Cancer Society. Development and analysis of quality assessment tools for different types of patient information – websites, decision aids, question prompt lists, and videos. BMC Med Inform Decis Mak 23, 111 (2023). https://doi.org/10.1186/s12911-023-02204-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12911-023-02204-5

Keywords