Skip to main content

Legal concerns in health-related artificial intelligence: a scoping review protocol

Abstract

Background

Medical innovations offer tremendous hope. Yet, similar innovations in governance (law, policy, ethics) are likely necessary if society is to realize medical innovations’ fruits and avoid their pitfalls. As innovations in artificial intelligence (AI) advance at a rapid pace, scholars across multiple disciplines are articulating concerns in health-related AI that likely require legal responses to ensure the requisite balance. These scholarly perspectives may provide critical insights into the most pressing challenges that will help shape and advance future regulatory reforms. Yet, to the best of our knowledge, there is no comprehensive summary of the literature examining legal concerns in relation to health-related AI. We thus aim to summarize and map the literature examining legal concerns in health-related AI using a scoping review approach.

Methods

The scoping review framework developed by (J Soc Res Methodol 8:19-32, 2005) and extended by (Implement Sci 5:69, 2010) and the Preferred Reporting Items for Systematic Reviews and Meta-Analysis extension for scoping reviews (PRISMA-ScR) guided our protocol development. In close consultation with trained librarians, we will develop a highly sensitive search for MEDLINE® (OVID) and adapt it for multiple databases designed to comprehensively capture texts in law, medicine, nursing, pharmacy, other healthcare professions (e.g., dentistry, nutrition), public health, computer science, and engineering. English- and French-language records will be included if they examine health-related AI, describe or prioritize a legal concern in health-related AI or propose a solution thereto, and were published in 2012 or later. Eligibility assessment will be conducted independently and in duplicate at all review stages. Coded data will be analyzed along themes and stratified across discipline-specific literatures.

Discussion

This first-of-its-kind scoping review will summarize available literature examining, documenting, or prioritizing legal concerns in health-related AI to advance law and policy reform(s). The review may also reveal discipline-specific concerns, priorities, and proposed solutions to the concerns. It will thereby identify priority areas that should be the focus of future reforms and regulatory options available to stakeholders in reform processes.

Trial registration

This protocol was submitted to the Open Science Foundation registration database. See https://osf.io/zav7w.

Peer Review reports

Background

Innovation in medicine offers tremendous hope but appropriate regulations are needed to realize this hope in healthcare and public health settings while minimizing attendant risks. The advent of health-related artificial intelligence (AI) provides an example of this need for regulation that fosters the requisite balance of potential benefits and pitfalls. Health-related AI is the topic of significant debate across multiple fields. Many portend that AI will improve healthcare systems by, e.g., increasing the accuracy of diagnoses, improving the efficiency of healthcare delivery, or mitigating human biases (e.g., [1, 2]). At the same time, there are understandable concerns that AI will undermine the patient-provider relationship, contribute to the deskilling of providers, undermine transparency, misdiagnose or inappropriately treat because of errors within AI decision-making that are hard to detect, exacerbate existing racial or societal biases, or introduce algorithmic bias that will be hard to detect (e.g., [3]). Opinions are, of course, not starkly polarized. Many simultaneously recognize AI’s promise and maintain concerns about its widespread adoption. Yet, these considerations highlight the need for regulations or other forms of governance (law, policy, ethics) that help countries leverage AI’s potential benefits while minimizing any attendant risks.

The use of health-related AI will intersect with the law in several ways. First, there are questions about whether existing laws will address identified concerns with health-related AI, such as the following:

  1. i.

    Whether medical device regulation, medical malpractice laws, product liability laws, and professional self-regulation and accreditation will adequately attend to the possibility of error on the part of AI;

  2. ii.

    Whether existing rules concerning the attribution of liability for medical error are appropriate for when AI tool recommends—or even performs—a course of harmful treatment and how liability should be attributed as between healthcare professionals and AI developers and manufacturers;

  3. iii.

    Whether existing anti-discrimination and human rights laws can attend to the problem of algorithmic bias in which AI tools inappropriately produce different outcomes for historically disadvantaged groups;

  4. iv.

    Whether existing privacy laws sufficiently protect patients given AI’s big data needs and the fact that machine learning (ML) tools, for example, will collect data in real time;

  5. v.

    Whether existing laws and policies relating to data governance are sufficient to enable AI innovators to have access to representative training data sets so as to appropriately include historically underrepresented populations;

  6. vi.

    Whether existing laws of informed consent are sufficiently robust to protect patients when clinicians choose to use AI in diagnosis and treatment.

Satisfactory responses are pressing as AI use increases in healthcare and public health settings. AI tools can already read medical images and reports more quickly and accurately than human counterparts [1, 2]. The U.S. Food and Drug Administration has approved AI tools for diagnosing strokes and brain bleeds, detecting atrial fibrillation, and interpreting brain MRIs; hospitals could also use other existing AI tools to assess patients’ risk of readmission and respond accordingly [1, 2]. Health-related AI is also used in Canada, including for triage purposes and to respond to COVID-19 [4]. The development of ML as a subset of AI that uses large data sets to “make predictions and solve problems … without being explicitly programmed” [5] has produced increasingly accurate health-related applications that can “learn” from real-world data over time and improve healthcare systems. As these technological innovations continue, there is a clear need to examine whether existing laws are up to the task of ensuring beneficial health-related AI tools can be deployed in real-world settings while minimizing legitimate concerns about, e.g., bias or privacy, that may arise.

Depending on one’s conclusions regarding the sufficiency of existing regulations to address concerns associated with the use of health-related AI, the next question is what kind of regulatory reform is required. Overly onerous regulations could stifle innovation, minimizing AI’s potential to realize health-related benefits. At the same time, a lack of regulation or inadequate regulations may lead to widespread harm resulting in a loss of trust in health-related AI (from patients themselves and healthcare professionals) and, perhaps, much more arduous regulation.

Despite the importance of law for the successful implementation of AI in health settings, there is no systematic overview of the legal concerns raised by the use of health-related AI. A preliminary search of MEDLINE®, Cochrane Library, PROSPERO, and JBI Evidence Synthesis was conducted; no current or underway systematic reviews or scoping reviews on this topic were identified. Two published scoping reviews were identified that survey ethical concerns raised by the use of health-related AI [5, 6] with one of the two taking a narrower focus on ethical issues concerning the disabled [6]. Although ethical issues can overlap with legal issues in some cases, legal concerns and legal responses are important to understand in their own right. For example, the law has an important claim to democratic legitimacy and failing to include legal issues means a failure to include concepts, such as patient autonomy, as articulated in sources of public and private law (e.g., Ontario’s Personal Health Information Protection Act). To illustrate with this one example, some may think that a consideration of ethical issues will capture concerns relating to consent to treatment, but a legal requirement for informed consent (as in Ontario’s Health Care Consent Act) can differ in its nature and impact. Further, although lawyers disagree on what the law should be, there is less room for disagreement than with respect to many ethical issues. Precedent (higher courts binding lower courts) helps create consistency in interpretation notwithstanding personal views on what the law should be.

A range of reports from different jurisdictions has identified areas where law reform may be needed to respond to problems in health-related AI ([4, 7,8,9,10,11,12,13]). Scholars have produced general overviews of the use of AI in healthcare that make claims with legal implications. For instance, Eric Topol’s well-known studies discuss concerns about the risk of AI tools causing harm to patients that implicate questions about AI safety and efficacy regulation and about both contractual and tort liability for harm [1, 2]. Other scholars have produced ethical analyses of health-related AI with possible legal implications. For instance, Alessandro Blasimme and Effy Vayena [14] raise concerns regarding AI safety, bias, and informed consent that may require legal responses of some kind. Mark Henderson Arnold [15] raises similar concerns, as well as questions about what to do when providers become too reliant on AI, which raise further liability questions. A relatively small number of works have articulated views on the most pressing legal concerns and possible best practices for approaching them (e.g., [3, 16]). Yet, there is no systematic overview of the legal concerns. As discussed above and further below, law and ethics overlap in important ways but legal and ethical issues can and should be distinguished where possible. A scoping review of paradigmatic and self-identified legal issues that identifies which concerns experts from different disciplines view as most important will enable better analysis of the adequacy of laws in various jurisdictions and what reforms are required (if any).

Objective

We aim to systematically map legal concerns that are identified in health-related AI and the extent to which they are prioritized across multiple relevant disciplines, namely law, medicine, nursing, pharmacy, other healthcare professions (dentistry, nutrition, etc.), public health, computer science, and engineering. In keeping with the central purposes of a scoping review approach, we aim to examine the extent, range, and nature of research activity across the disciplines; to summarize and disseminate research findings to relevant stakeholders; and to identify research gaps in the existing literature. The scoping review will be conducted in accordance with the framework developed by Arksey and O’Malley [17] and extended by Levac et al. [18] and aligned protocol reporting to the PRISMA-P checklist [19]. It will thus include 6 stages: (1) Identifying the Research Question(s); (2) Identifying Relevant Studies; (3) Study Selection; (4) Charting the Data; (5) Collating, Summarizing, and Reporting Results; and in this case, (6) Stakeholder Consultation.

Stage one: identifying the research question(s)

Through iterative discussions, research team members met to discuss and refine the research question(s). To ensure diverse perspectives and representation, the team included expert clinicians, AI innovators, legal researchers, ethics experts, and a member with scoping review expertise. From these discussions, we decided that our aim to map the nature, extent, and range of literature examining legal concerns in health-related AI. To this end, our research question(s) are as follows:

Primary question

  1. 1.

    What is known from the literature regarding legal concerns in health-related AI?

Secondary questions

  1. 1.

    Are the legal concerns identified explicitly prioritized?

  2. 2.

    Do different disciplines identify, represent, or prioritize legal concerns differently?

Stage two: identifying relevant studies

Guided by two expert information specialists, we will develop a highly sensitive search strategy to identify relevant records. A preliminary pilot search of MEDLINE® and HeinOnline was conducted to pilot test the draft search and its ability to identify key articles. Titles, abstracts, keywords, and Medical Subject Headings (MeSH) in MEDLINE® of the key articles were analyzed to develop the final search strategy. See Table 1 for details. The MEDLINE® search was peer-reviewed by an independent information specialist using the Peer Review of Electronic Search Strategies (PRESS) checklist [20]. This original search strategy will be adapted to search additional medical, legal, and multidisciplinary databases as applicable. Searching will be conducted in the following electronic databases: MEDLINE® (Ovid), EMBASE (Ovid), HeinOnline Law Journal Library, Index to Foreign Legal Periodicals (HeinOnline), Index to Legal Periodicals and Books (EBSCO), Web of Science, Scopus, and IEEE Xplore. Given the complexity of searching (e.g., its multidisciplinary), the strategies will be augmented by hand-searching reference lists of all relevant, full-text records to identify additional sources [21]. Search results will be exported into a proprietary review software program (Covidence) to facilitate review processes and manage each stage of the review (e.g., de-duplication, eligibility assessment).

Table 1 MEDLINE® search strategy

Searches will be conducted between 2012 and the start of the review. This decision was made through consultation with all team members and was made to account for more recent developments in the field of AI and the deployment of AI into healthcare settings. For example, the development of deep learning AI is viewed as a “paradigm shift” in the field [22]. While deep learning AI was developed prior to 2012, the use of deep learning in the ImageNet Large Scale Visual Recognition Challenge 2012-winning AlexNet [23] led to the technology’s increased public recognition. Experts accordingly view 2012 as the year deep learning was “widely accepted as a viable form of” AI [2]. Health-related AI has become more common since that date [2].

Several members of the research team were part of a previous Royal College of Physicians and Surgeons of Canada Task Force Report on Artificial Intelligence and Emerging Digital Technologies (TH, DS, CF) that contained a bibliography on the grey literature [10]. The team will use that bibliography as a starting point for identifying relevant grey literature, supplementing it by crowdsourcing information about subsequent grey literature among experts working on the research project and further hand-searching, including review of references for mentions of other reports by governments (e.g., [24]), non-governmental organizations (e.g., [4, 12]), and professional organizations (e.g., [10]). Initial examinations of such reports suggest that they are written from an interdisciplinary perspective and thus inappropriate for our task of mapping concerns across disciplines. However, we will summarize their statements on legal concerns and compare them with those in published records.

Stage three: Study selection

Eligibility criteria

Our search strategy intends to identify published literature. It will be designed to identify indexed records, including articles and book chapters. It will also include American law journals, which are academic journals that undergo a standard student review process. Publications will not be limited to works in law journals as the reviewers are interested in what people working in other fields view as key legal concerns (or if they are even discussing legal concerns). Articles in medical, social scientific, computer science or engineering journals that raise legal concerns or solutions thereto are included in this review. We will not include abstracts and conference proceedings.

This review is centrally focused on identifying records that describe, articulate, analyze, or prioritize legal concerns associated with health-related AI. It is specifically focused on identified concerns in health-related AI that may require a legal response. This could include records identifying concerns that require a clear legal solution (e.g., a legal gap), evaluation of existing laws to meet the concerns, and examining the nature, advantages, and disadvantages of various law reform proposals. We will include records addressing various legal concerns (e.g., an omnibus of concerns requiring legal responses from bias to liability) as well as those addressing a single concern (e.g., privacy). We will exclude records that simply describe existing law(s) that apply to the use of health-related AI. We recognize that “law” and “ethics” are sometimes conflated and overlap in some respects. Yet, ethical concerns are and should be distinguished from legal concerns: not all laws reflect our best moral standards and not all ethical concerns implicate the law. Further, ethical debates about health-related AI have been the subject of a scoping review [5]. To this end, records that identify an ethical concern will only be included when there is clearly an articulated nexus to a legal issue (for example, where there is an ethical concern related to privacy and then whether privacy laws are sufficient to protect patients and providers when using ML devices).

Records will be eligible for inclusion if they are English- or French-language publications that describe a legal concern pertaining to the use of AI that is health-related. Working definitions for the key terms (“legal concerns,” “AI,” and “health-related”) are summarized succinctly in Table 2. For the initial searches, we will not impose a language restriction. However, we will include only records published in English or French in the scoping review itself.

Table 2 Key terms defined for eligibility assessment

Eligibility assessment

An eligibility assessment tool (for level 1 title and abstract screening and level 2 full-text screening) will be developed and pilot tested in collaboration with the entire review team. Using a proprietary review software program to manage records, duplicate assessment of each record to determine inclusion will be conducted independently. Agreement will be assessed and reported using a Kappa statistic (e.g., [29]). Conflicted decisions will be resolved by one of the subject matter expert authors (MD, CF, DS). Eligibility assessment and flow of the review will be reported in accordance with the PRISMA flow diagram.

Stage four: Charting the data

A standardized form to extract and categorize data will be developed through a series of consultations with the team and pilot-tested using known includes. The form will cover two primary areas of information: (1) record-level demographic characteristics and (2) content specific to legal concerns, priority areas, solutions, etc., that can be inferred from the records.

Basic record-level information will include year of publication, disciplinary area of focus (e.g., law, medicine), geographic information (e.g., country, region), publication type, study design, language of publication, and area of health-related focus. For example, we will identify the medical specialty in which AI and the legal concern are being confronted. To categorize, we will design a list of specialties (using as a base the list employed by the Royal College of Physicians and Surgeons of Canada but modifying it to, e.g., add family medicine specializations). When this list is insufficient to capture the area of focus (e.g., general medicine, medical administration, electronic health record), we will extract the information verbatim as described in the original record. Given that this is a scoping review and is meant to be iterative and that information gathered can inform future stages, any additional decisions on data to be extracted that occur beyond the protocol stage will be declared transparently in the distal publication.

When extracting data pertaining to any legal concern(s), we will also extract any specific text that expressly indicates prioritization of the concerns and proposed solutions (e.g., new or modified regulations), including new interpretations or expansions of private law responsibilities (e.g., in tort law or contract law), ethical reform (e.g., on the part of AI innovators), and educational/training reforms. When looking at the prioritization of legal issues, we will rely on explicit self-identified priority claims (describing issues as “high priority,” “most important,” “most pressing,” etc.).

Stage five: Collating, summarizing, and reporting results

Once data has been extracted and accuracy verified through duplicate assessment, we will summarize all record-level characteristics and use the summaries to map publication patterns (who is publishing on the intersection of health related-AI and law) and thematically code legal concerns into categories (e.g., typology). We will categorize the legal concerns using our framework, which seeks to summarize the legal concerns, priorities, and proposed solutions to the concerns.

Once categorized, we will examine the extent to which we see different legal issues of concern to different disciplines (law, medicine, nursing, pharmacy, other health professions, public health, computer science, and engineering). For example, we may see that privacy is raised as a concern by those in the fields of computer science or engineering whereas liability may be more frequently raised as a concern by legal researchers. We will also examine whether we see differences in what legal concerns are identified by specialty, country/region, and whether or not any author on the record is identifiable as an expert from a relevant discipline other than that of the primary author. To achieve the latter, we will categorize authors using record-level information pertaining to their institutional affiliation (e.g., School of Engineering) or listed credentials. It is, of course, possible that individuals trained in one discipline work in another discipline (e.g., a lawyer could work in bioethics within a faculty of medicine). However, departmental affiliation is a useful marker of disciplinary perspective and what they represented as a co-author on the work. Any normative recommendations in the article stemming from data collected in this scoping review will be based on our multidisciplinary team’s views on any inferences that can be drawn from the descriptive findings. In accordance with the PRISMA-Scr [30], we will not formally assess methodological quality of records.

Stage six: Stakeholder consultation

Understanding the complexity of searching and breadth of the review scope, we plan to engage key consumers and stakeholders to suggest additional records and sources when an initial full-text include list is generated. Given that this review is one component of a larger Canadian Institutes of Health Research-funded project, we plan to solicit input from the broader multidisciplinary team, which includes patient partners and international experts on law, AI, and medicine [31]. If gaps in expertise or perspectives are identified, the team will employ a snowball sampling technique whereby an existing contact will be asked to provide the name(s) of other relevant experts and the process can be repeated as needed. The process of selection and consultation will be detailed sufficiently to ensure transparency.

Limitations

This scoping review is mapping records from across several disciplines and aims to identify legal concerns that are discussed in health-related AI. Given the complexity of searching, and despite all attempts to be comprehensive, it is conceivable that our search strategy may miss relevant records. This may be due to the records using idiosyncratic language or irregular key word entries or other indexing-related barriers. It is also possible that relevant legal concerns could be in records that predate 2012. This said, our team anticipates that important debates will be captured. Finally, it is possible that few works will make explicit claims about the priority of legal issues; however, we will be able to impute the importance of particular legal questions by looking at the number of times that specific legal issues are raised.

Discussion

Health-related AI offers numerous potential benefits but also several potential pitfalls. Apt regulation is necessary to maximize its potential benefits while minimizing its risks. Regulatory reforms are more likely to maintain the requisite balance when informed by an understanding of how experts prioritize the concerns and potential solutions they have identified. This scoping review accordingly aims to systematically, comprehensively, and cohesively map the extant literature on legal concerns in health-related AI. A comprehensive map of how different disciplines frame, prioritize, and analyze legal concerns in health-related AI and how they propose to address those concerns will support cross-disciplinary understanding of perceived legal barriers or gaps. It will also aid in policymaking through a holistic view of legal concerns in health-related AI and multiple perspectives on solutions. Our review will thus provide the bedrock from which to better develop domestic and international responses to legal challenges posed by health-related AI.

Availability of data and materials

Not applicable

References

  1. Topol E. Deep medicine: how artificial intelligence can make healthcare human again. New York: Basic Books; 2019.

  2. Topol E. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. 2019;25:44–56.

  3. Gerke S, et al. Ethical and legal challenges of artificially intelligence-driven healthcare. In: Bohr B, Memarzadeh K, editors. Artificial Intelligence in Healthcare. Cambridge: Academic Press; 2020.

    Google Scholar 

  4. Canadian Institute for Advanced Research (CIFAR). AI and Healthcare: A Fusion of Law and Science – An Introduction to the Issues. Toronto: CIFAR; 2021.

  5. Murphy K, et al. Artificial intelligence for good health: a scoping review of the ethics literature. BMC Medical Ethics. 2021;22:1–17.

    Article  Google Scholar 

  6. Lillywhite A, Wolbring G. Coverage of ethics within the artificial intelligence and machine learning academic literature: the case of disabled people. Assistive Technology. 2019;33(3):1–17.

  7. Academy of Medical Royal Colleges. Artificial intelligence in healthcare. London: Academy of Medical Royal Colleges; 2019.

    Google Scholar 

  8. AI Forum of New Zealand. Artificial intelligence for health in New Zealand. Northshore: AI Forum of New Zealand; 2019.

    Google Scholar 

  9. Canadian Institute for Advanced Research (CIFAR). Building a learning health system for Canadians: report of the task force on Artificial Intelligence for Health (AI4Health). Toronto: CIFAR; 2020.

    Google Scholar 

  10. Royal College of Physicians and Surgeons of Canada (Royal College). Task force report on artificial intelligence and emerging digital technologies. Ottawa: RCPSC; 2020.

    Google Scholar 

  11. Canadian Medical Association (CMA). The future of technology in health and health care: a primer. Ottawa: CMA; 2020.

    Google Scholar 

  12. Canadian Institute for Advanced Research (CIFAR). AI and Healthcare: A Fusion of Law and Science – The Regulation of Medical Devices with AI Toronto: CIFAR; 2021.

  13. World Health Organization (WHO). Ethics and governance of artificial intelligence for health. Geneva: WHO; 2021.

    Google Scholar 

  14. Blasimme A, Vayena E. The ethics of AI in biomedical research, medicine and public health. In: Dubber MD, et al., editors. The Oxford Handbook of Ethics of AI Oxford. Oxford UP; 2020.

  15. Henderson AM. Teasing out artificial intelligence in medicine: an ethical critique of artificial intelligence and machine learning in medicine. J Bioethical Inquiry. 2021;18:121–39.

    Article  Google Scholar 

  16. Kerr I, et al. Robots and artificial intelligence in health care. In: Erdman J, et al., editors. Canadian Health Law & Policy. 5th ed. Toronto: LexisNexis Canada; 2017.

    Google Scholar 

  17. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8:19–32.

    Article  Google Scholar 

  18. Levac D, et al. Scoping studies: advancing the methodology. Implementation Science. 2010;5:69.

    Article  Google Scholar 

  19. Moher D, et al. Preferred Reporting Items for Systematic Review and Meta-Analysis Protocols (PRISMA-P) 2015 Statement. Syst Rev. 2015;4:1.

    Article  Google Scholar 

  20. McGowan J, et al. PRESS peer review of electronic search strategies: 2015 Guideline Statement. J Clin Epidemiol. 2016;75:40–6.

    Article  Google Scholar 

  21. Horsley T, et al. Checking reference lists to find additional studies for systematic reviews. Cochrane Database Syst Rev. 2011;10(8):MR000026.

    Google Scholar 

  22. Bohr A, Memarzadeh K. The rise of artificial intelligence in healthcare applications. In: Bohr B, Memarzadeh K, editors. Artificial Intelligence in Healthcare. Cambridge: Academic Press; 2020.

    Google Scholar 

  23. Krizhevky A, et al. ImageNet classification with deep convolutional neural networks. NIPS’12: Proceedings of the 25th International Conference on Neural Information Processing. Systems. 2012;1:1097–105.

    Google Scholar 

  24. United States Government Accountability Office (US GAO). Artificial intelligence in health care benefits and challenges of technologies to augment patient care. Washington: US GAO; 2020.

    Google Scholar 

  25. Hage J. Foundations. In: Hage J, Akkermans B, editors. Introduction to law. New York: Springer; 2014.

  26. Hage J. Sources of Law. In: Hage J, Akkermans B, editors. Introduction to law. New York: Springer; 2014.

  27. Meriam-Webster. Health Care. 2021. Online: <https://www.merriam-webster.com/dictionary/health%20care>. Accessed 10 Dec 2021.

  28. Meriam-Webster. Public Health. 2021. Online: <https://www.merriam-webster.com/dictionary/public%20health>. Accessed 10 Dec 2021.

  29. Kundel HL, Polansky M. Measurement of observer agreement. Radiology. 2003;228(2):303–8.

    Article  Google Scholar 

  30. Tricco AC, et al. 2018. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169(7):467–73.

    Article  Google Scholar 

  31. University of Ottawa Centre for Health Law, Policy and Ethics (CHLPE). Machine M.D. 2021. Online: https://www.ottawahealthlaw.ca/projects/machine-m-d. Accessed 10 Dec 2021.

Download references

Acknowledgements

The authors thank Lindsey Sikora MISt PhD(c), University of Ottawa, for peer review of the MEDLINE® search strategy.

Funding

This protocol benefits from a Canadian Institutes of Health Research Project Grant, Machine M.D.: How Should We Regulate AI in Health Care? MD received post-doctoral research funding from the Alex Trebek Forum for Dialogue. KCF received post-doctoral research funding from Scotiabank. None of these funding bodies played a role in the design of this protocol.

Author information

Authors and Affiliations

Authors

Contributions

CF conceptualized the research program. MD, TH, DS, and CF contributed substantially to the conception and design of the review. ED and VL designed the search strategy and drafted the description thereof. MD initially drafted the manuscript. TH led the development of the review methodology and, with MD, CF, DS, and BT, contributed substantively to the secondary draft protocol. The authors read and approved the final version of the protocol.

Corresponding author

Correspondence to Michael Da Silva.

Ethics declarations

Ethics approval and consent to participate

Not applicable

Consent for publication

Not applicable

Competing interests

Dr. Devin Singh is the co-founder and CEO of Hero AI, a healthcare technology start-up company. The other authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visithttp://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Da Silva, M., Horsley, T., Singh, D. et al. Legal concerns in health-related artificial intelligence: a scoping review protocol. Syst Rev 11, 123 (2022). https://doi.org/10.1186/s13643-022-01939-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13643-022-01939-y

Keywords