Public views on ethical issues in healthcare artificial intelligence: protocol for a scoping review
Systematic Reviews volume 11, Article number: 142 (2022)
In recent years, innovations in artificial intelligence (AI) have led to the development of new healthcare AI (HCAI) technologies. Whilst some of these technologies show promise for improving the patient experience, ethicists have warned that AI can introduce and exacerbate harms and wrongs in healthcare. It is important that HCAI reflects the values that are important to people. However, involving patients and publics in research about AI ethics remains challenging due to relatively limited awareness of HCAI technologies. This scoping review aims to map how the existing literature on publics’ views on HCAI addresses key issues in AI ethics and governance.
We developed a search query to conduct a comprehensive search of PubMed, Scopus, Web of Science, CINAHL, and Academic Search Complete from January 2010 onwards. We will include primary research studies which document publics’ or patients’ views on machine learning HCAI technologies. A coding framework has been designed and will be used capture qualitative and quantitative data from the articles. Two reviewers will code a proportion of the included articles and any discrepancies will be discussed amongst the team, with changes made to the coding framework accordingly. Final results will be reported quantitatively and qualitatively, examining how each AI ethics issue has been addressed by the included studies.
Consulting publics and patients about the ethics of HCAI technologies and innovations can offer important insights to those seeking to implement HCAI ethically and legitimately. This review will explore how ethical issues are addressed in literature examining publics’ and patients’ views on HCAI, with the aim of determining the extent to which publics’ views on HCAI ethics have been addressed in existing research. This has the potential to support the development of implementation processes and regulation for HCAI that incorporates publics’ values and perspectives.
Recent years have seen the development and introduction of a number of artificial intelligence (AI) enabled technologies for healthcare. AI is a term which encompasses diverse computational technologies, making it challenging to define: prominent definitions include that AI is ‘a collection of interrelated technologies used to solve problems that would otherwise require human cognition’ , or that AIs are technologies with the ability to ‘perform tasks to achieve defined objectives without explicit guidance from a human being’ . Broad in application, AI technologies arrive with optimistic promises of transforming the patient experience. Many of these modern developments are the result of innovations in machine learning (ML), a branch of AI focussed on developing algorithms which learn from examples . So far, advocates of healthcare AI (HCAI) have promised the technology will improve the accuracy of screening and diagnosis , increase the availability of care in remote regions , and free up physicians’ time so that they can engage more with patients .
Alongside innovations in HCAI, there is also a growing field of AI ethics which cautions against uncritical implementation of HCAI and raises questions about its regulation . ML technologies pose new risks and challenges to healthcare: some ML algorithms have been shown to produce biased outcomes , many ML technologies are ‘black boxes’ where the reasons behind an algorithm’s output cannot be interpreted , and questions remain about how existing liability structures in medicine will effectively manage errors made by deployed ML technologies . AI development also continues to be dominated by large private companies that have been criticised for failing to engage in meaningful conversations about the ethics of their products and research .
Publics may be both beneficiaries of new HCAI technologies and the greatest sufferers of AI-related harms . Patients and publics are important voices in developing effective and ethical AI governance, but engaging patients and publics meaningfully in research about ethical HCAI is challenging. Most people have no firsthand experience with HCAI, and some are unfamiliar with the concept of AI in general . Publics may have limited understanding of how HCAI may be implemented, and limited knowledge about the potential wrongs and harms that could arise from implementing HCAI. To ensure that HCAI has a positive impact on patients, it is crucial that AI ethics reflects the values that are important to people [12, 14], but it remains unclear how this should be achieved.
The aim of this review is to determine how common and emerging themes in HCAI ethics are addressed by the existing literature on publics’ and patients’ views on machine learning in healthcare.
Scoping reviews are an effective method for exploring the range and extent of literature on a given topic . Our work will follow the framework proposed by Arksey and O’Malley  with modifications from Levac and colleagues . Their six recommended steps include (a) identifying the research question; (b) identifying relevant studies; (c) study selection; (d) charting the data; (e) collating, summarising, and reporting the results; and (f) consultation. The following sections will address each of these steps in greater detail. In preparing this protocol, we completed a PRISMA-P checklist to ensure all necessary details have been reported (Additional file 1).
Stage 1: Identifying the research questions
Our review will address the question, to what extent, and how, are HCAI ethics issues addressed in the existing literature on publics’ views on machine learning applications of HCAI. Our objectives are (1) to explore whether and how research on public views regarding HCAI has included investigation of public views on HCAI ethics (2) to describe study participants’ perspectives on HCAI ethics issues.
Stage 2: Identifying relevant literature
We developed a search query using the Population-Intervention-Context-Outcome (PICO) format. An initial search on Google Scholar helped to identify similar terms which were used to develop a comprehensive search query for published literature (Table 1).
We will use a systematic search strategy to find relevant articles for inclusion in the study. Databases to be searched are PubMed, Scopus, Web of Science, CINAHL, and Academic Search Complete. To find relevant grey literature, we will screen the first ten pages of a Google Advanced search. We will examine the reference lists of included studies to find any publications that were missed in the initial searches. All studies collected through the search project will be imported into a Zotero library.
Stage 3: Study selection
After the search is completed, all studies will be screened for eligibility. EF will be responsible for conducting the search and managing the data. First, duplicates will be removed using the deduplication module from the Systematic Review Assistant , and then the remaining files will be exported to MS Excel for the screening process. MS Excel will allow reviewers to easily indicate a study’s inclusion or exclusion, as well as keep notes about any uncertainties for discussion.
The first stage of screening will exclude irrelevant articles based on their title and abstract. Screening will be conducted based on a set of criteria defined below. The first reviewer (EF) will screen the first 10% of articles, including all articles that are potentially relevant based on their title and abstract and excluding articles that are clearly irrelevant. Of this 10%, EF will construct a sample of approximately 40 articles marked for inclusion, and 60 articles marked for exclusion. A second reviewer (RB) will screen this sample of 100 articles and compare results with EF. Results will be discussed with the team and inclusion criteria will be modified if necessary. Once any issues have been resolved, EF will conduct the initial screening on the remainder of the studies.
After initial screening is completed, excluded articles will be removed and full article texts will be collected for the remaining studies. The two-reviewer screening process will be repeated on a random sample of 10% of the full texts. Differences will be discussed and resolved, and modifications will be made to the inclusion criteria accordingly. Inter-rater scores will be generated to quantify agreeance between reviewers. Once the inclusion criteria are finalised, EF will conduct the remainder of the full-text screening.
Articles will be screened against a set of inclusion criteria developed by the team. These criteria may be modified throughout the screening process. Initial design of the criteria was guided by the JBI guidelines for scoping reviews . This guide states that inclusion criteria should address (i) the types of participants, (ii) concept, (iii) context, and (iv) types of evidence sources.
Types of participants
Studies will be included if research participants are recruited as publics, patients (or their unpaid/familial carers), or healthcare consumers. If studies recruit professionals (e.g. physicians, nurses, policymakers, or professional carers) along with publics, they will be included so long as the data related to patients/publics can be extracted.
Studies must address publics’ or patients’ views on HCAI. In this case, we utilise the term “views” to refer to the various ways participants contribute to social research. Included studies may, for example, quantitatively measure participants’ attitudes toward HCAI, or qualitatively examine participants’ perspectives on HCAI (or an application thereof). Studies will be excluded if the participants’ contribution to the research does not involve sharing views (e.g. studies only measuring whether a particular HCAI tool has improved participant outcomes in a certain area).
Studies will be included if the research addresses machine learning in patient- or (general) public-facing health care or services. An included study may address machine learning in healthcare or services in a specific field (e.g. patients’ perspectives on AI for breast screening, or publics’ attitudes toward AI-enabled mobile phone apps for skin cancer detection).
Studies will be excluded if they only address AI technologies that are not within the machine learning branch of AI. For example, studies only examining participants’ views on care robots or expert systems would be excluded. Studies will be excluded if they only address AI in non-patient/public-facing health applications. For example, studies addressing AI used only to manage bills and claims processing in hospitals would be excluded. Studies will be excluded if they only address non-health applications of AI.
Studies from any geographical location will be included, so long as the manuscript can be assessed in English. Only studies published between 1 January 2010 and 15 September 2021 will be included. This time period has seen the introduction of modern approaches such as deep learning, convolutional neural networks, and natural language processing into HCAI research . These new approaches are the source of much of the current interest and investment in HCAI, and introduce a number of new potential challenges and harms .
Types of evidence sources
Only primary research studies will be included in this review. Studies will not be excluded based on method. There will be no restrictions on study design.
Studies will be excluded if they are only available in a language other than English, if they do not address AI in a patient-facing healthcare context or if the study participant profile does not include patients or publics.
Stage 4: Charting the data
We have designed a coding framework to capture information on whether and how studies address a series of AI ethics concerns.Footnote 1 Whilst a number of different frameworks were reviewed [19, 20], the coding framework is primarily based on Fjeld and colleagues’  analysis of a series of AI ethics guidelines. Fjeld et al. identified seven domains that were frequently addressed in AI ethics frameworks: (1) privacy, (2) accountability, (3) safety and security, (4) transparency and explainability, (5) fairness and non-discrimination, (6) human control over technology, and (7) professional responsibility. To capture more detailed data on where each of these ethical issues were addressed, we separated the concepts of ‘safety’ and ‘security’, and ‘transparency’ and ‘explainability’ into individual code categories (Table 2).
We added four additional concepts to the framework. The first, power, has become a more common point of discussion in AI ethics frameworks recently, to assess how AI development and governance structures are reinforcing existing power dynamics and failing to redistribute power to marginalised groups [14, 22]. The second, environmental wellbeing, addresses the environmental impacts of AI development including energy usage, materials, and e-waste [22, 23]. Societal wellbeing addresses whether technological development is being implemented for social good . Finally, ethical governance addresses whether existing governance structures are suitable to manage the ethical issues associated with HCAI.
Additional information about study design and methods will also be collected, including detailed notes on study design. We will use MS Excel to chart the data from the studies. The initial data extraction tool (Additional file 2) covers the key areas recommended by Arksey and O’Malley , the ethics framework, and the additional information about study design. In adhering to recommendations from Levac et al. , we will modify this tool progressively throughout the data collection process. Initially, a random 10% of the included studies will be selected and coded by two coders (EF & RB) and any differences will be resolved in consultation with the research team. We will make changes to the data extraction tool if necessary. The remainder of the charting will be conducted by EF.
Stage 5: Collating, summarising, and reporting the results
We will collate results into tabular format for analysis. Guided by Arksey and O’Malley’s  recommendations, analysis will begin with descriptive quantitative reporting where it is appropriate (e.g. the number of studies which address each HCAI ethics issue in the framework).
Our reporting will synthesise publics’ and patients’ views on each of the HCAI ethics issues in Table 2. Given the inclusion criteria for this review, we are likely to collect studies with diverse designs. In some cases, direct quantitative comparison between studies may be possible. In other cases, studies with different methodological designs may be compared with one another. Where studies do not allow for direct comparisons, our results will report narrative descriptions and comparisons, noting how a study’s framing, aims, and contexts might influence the information collected. This synthesis methodology will be refined based on the types of studies collected.
This review may have some limitations. Firstly, scoping reviews are designed to map the literature in a topic and are not designed to assess the quality of included studies . The quality of the studies included in this review will not be systematically assessed. Secondly, it is possible that relevant studies will not be captured by the search strategy defined in this protocol. We will conduct a systematic pearling process on relevant identified studies to ensure as many relevant articles are identified as possible. Finally, findings will be limited to studies published in English, which may exclude relevant articles published in other languages. We will reflect on the impact of these limitations, as well as discuss any other arising limitations, in the reporting of our results.
The more widespread use of HCAI technologies is often described as inevitable . However, implementation of HCAI may exacerbate certain harms in healthcare . Although patients and publics are likely to be the greatest sufferers of HCAI-related harms, involving patients and publics in meaningful research about AI ethics remains challenging .
To date, the extent to which patients and publics are involved in research about HCAI ethics is unclear. This review will examine where existing research has involved patients and publics in research about a series of HCAI ethics issues. In doing so, we will describe patients’ and publics’ views on each HCAI ethics issue, and highlight potential gaps, or areas of HCAI ethics where research with patients and publics is limited. The results from this review will be important to understanding where further effort is required to involve patients and publics in research about HCAI ethics. Such an effort is crucial to ensuring that HCAI is implemented safely and effectively.
Availability of data and materials
No datasets were generated or used for this paper.
AI ethics considerations inevitably overlap with legal considerations, as ethics and the law are both normative domains which share some common concerns. We do not make a strong demarcation between ethical and legal considerations in this protocol, but we do approach normative concepts from the perspective of ethics rather than the law.
Healthcare artificial intelligence
Walsh T, Levy N, Bell G, Elliott A, Maclaurin J, Mareels I, et al. The effective and ethical development of artificial intelligence: an opportunity to improve our wellbeing. The Australian Council of Learned Academies; 2019.
Hajkowicz S, Karimi S, Wark T, Chen C, Evans M, Rens N, et al. Artificial Intelligence: solving problems, growing the economy and improving our quality of life. CSIRO, Data61; 2019 p. 68.
Jordan MI, Mitchell TM. Machine learning: trends, perspectives, and prospects. Science. 2015;349(6245):255–60.
Esteva A, Topol E. Can skin cancer diagnosis be transformed by AI? Lancet. 2019;394(10211):1795.
Sechopoulos I, Mann RM. Stand-alone artificial intelligence - the future of breast cancer screening? Breast. 2020;1(49):254–60.
Topol E. Deep Medicine: How artificial intelligence can make healthcare human again. Illustrated. New York: Basic Books; 2019. p. 400.
Carter SM, Rogers W, Win KT, Frazer H, Richards B, Houssami N. The ethical, legal and social implications of using artificial intelligence systems in breast cancer care. The Breast. 2020;49:25–32.
Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–53.
Dilsizian ME, Siegel EL. Machine meets biology: a primer on artificial intelligence in cardiology and cardiac imaging. Curr Cardiol Rep. 2018;20(12):139.
Grote T, Berens P. On the ethics of algorithmic decision-making in healthcare. J Med Ethics. 2020;46(3):205–11.
Holzmeyer C. Beyond ‘AI for Social Good’ (AI4SG): social transformations—not tech-fixes—for health equity. Interdisc Sci Rev. 2021;46(1–2):94–125.
Laï MC, Brian M, Mamzer MF. Perceptions of artificial intelligence in healthcare: findings from a qualitative survey study among actors in France. J Transl Med. 2020;18(1):14.
Young AT, Amara D, Bhattacharya A, Wei ML. Patient and general public attitudes towards clinical artificial intelligence: a mixed methods systematic review. The Lancet Digital Health. 2021;3(9):e599-611.
Hickok M. Lessons learned from AI ethics principles for future actions. AI Ethics. 2021;1(1):41–7.
Levac D, Colquhoun H, O’Brien KK. Scoping studies: advancing the methodology. Implementation Sci. 2010;5(1):69.
Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19–32.
Rathbone J, Carter M, Hoffmann T, Glasziou P. Better duplicate detection for systematic reviewers: evaluation of Systematic Review Assistant-Deduplication Module. Syst Rev. 2015;14(4):6.
Peters MDJ, Godfrey CM, Khalil H, McInerney P, Parker D, Soares CB. Guidance for conducting systematic scoping reviews. JBI Evidence Implementation. 2015;13(3):141–6.
Floridi L, Cowls J. A unified framework of five principles for AI in society. Harvard Data Science Review. 2021 [cited 2022 Jun 16]; Available from: https://hdsr.mitpress.mit.edu/pub/l0jsh9d1
Jobin A, Ienca M, Vayena E. Artificial intelligence: the global landscape of ethics guidelines. 2019;42.
Fjeld J, Achten N, Hilligoss H, Nagy A, Srikumar M. Principled artificial intelligence: mapping consensus in ethical and rights-based approaches to principles for AI. SSRN J. 2020 [cited 2021 Sep 2]; Available from: https://www.ssrn.com/abstract=3518482
Crawford K. Atlas of AI: power, politics, and the planetary costs of artificial intelligence. Atlas of AI. Yale University Press; 2021 [cited 2021 Sep 15]. Available from: https://www.degruyter.com/document/doi/https://doi.org/10.12987/9780300252392/html
European Commission. Ethics guidelines for trustworthy AI | Shaping Europe’s digital future. 2019 [cited 2021 Sep 2]. Available from: https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai
Guan J. Artificial intelligence in healthcare and medicine: promises, ethical challenges and governance. Chin Med Sci J. 2019;30(34):76–83.
No funding has been sought for this study.
Ethics approval and consent to participate
Not applicable to this study.
Consent for publication
Not applicable to this study.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Frost, E.K., Bosward, R., Aquino, Y.S.J. et al. Public views on ethical issues in healthcare artificial intelligence: protocol for a scoping review. Syst Rev 11, 142 (2022). https://doi.org/10.1186/s13643-022-02012-4
- Artificial Intelligence
- AI ethics
- Healthcare AI