This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\copyyear

2024 \startpage1

\authormark

Chinta et al. \titlemarkStatistics in Medicine

\corres

Wenbin Zhang, Knight Foundation School of Computing and Information Sciences, Florida International University, 11200 SW 8th Street, CASE 333, Miami, Florida, USA 33199

AI-Driven Healthcare: A Survey on Ensuring Fairness and Mitigating Bias

Sribala Vidyadhari Chinta    Zichong Wang    Xingyu Zhang    Thang Doan Viet    Ayesha Kashif    Monique Antoinette Smith    Wenbin Zhang \orgnameFlorida International University, \orgaddress\cityMiami, \stateFL, \postcode33199 \orgnameUniversity of Pittsburgh, \orgaddress\cityPittsburgh, \statePA, \postcode15260 \orgnameJose Marti MAST 6-12 Academy, \orgaddress\cityHialeah, \stateFL, \postcode33016 \orgnameEmory University, \orgaddress\cityAtlanta, \stateGA, \postcode30322 [email protected]    Chinta S. V    Wang Z    Zhang X    Doan T    Kashif A    Sminth M. A    Zhang W
(Date Month Year; Date Month Year; Date Month Year)
Abstract

[Abstract]Artificial intelligence (AI) is rapidly advancing in healthcare, enhancing the efficiency and effectiveness of services across various specialties, including cardiology, ophthalmology, dermatology, emergency medicine, etc. AI applications have significantly improved diagnostic accuracy, treatment personalization, and patient outcome predictions by leveraging technologies such as machine learning, neural networks, and natural language processing. However, these advancements also introduce substantial ethical and fairness challenges, particularly related to biases in data and algorithms. These biases can lead to disparities in healthcare delivery, affecting diagnostic accuracy and treatment outcomes across different demographic groups. This survey paper examines the integration of AI in healthcare, highlighting critical challenges related to bias and exploring strategies for mitigation. We emphasize the necessity of diverse datasets, fairness-aware algorithms, and regulatory frameworks to ensure equitable healthcare delivery. The paper concludes with recommendations for future research, advocating for interdisciplinary approaches, transparency in AI decision-making, and the development of innovative and inclusive AI applications.

\jnlcitation\cname

, , , , , , and . \ctitleAI-Driven Healthcare: A Survey on Ensuring Fairness and Mitigating Bias. \cjournalStatistics in medicine. \cvol2021;00(00):1–18.

keywords:
AI in Healthcare, Ethical Challenges, Bias Mitigation, Data Diversity, Fairness-aware Algorithms
articletype: Research articlejournal: Statistics in Medicinevolume: 00

1 Introduction

Artificial intelligence (AI) is revolutionizing modern healthcare, dramatically transforming the ways we diagnose, treat, and manage diseases. The integration of AI into healthcare began in the late 20th century with systems like MYCIN 1 in the 1970s, which helped diagnose infections and recommend antibiotics, and CADUCEUS 2 in the 1980s, which emulated human diagnostic reasoning. These early systems laid the groundwork for today’s advanced machine learning and deep learning techniques, which now significantly enhance diagnostic accuracy, treatment personalization, and patient outcome predictions.

As AI technologies have advanced, their impact on healthcare has grown exponentially. Modern AI applications, particularly deep learning, have enhanced image recognition, significantly improving diagnostic accuracy in fields such as radiology and pathology 3. Predictive analytics, powered by AI, are essential in patient monitoring and management, using real-time data to forecast potential patient deteriorations 4. Additionally, natural language processing (NLP) tools have revolutionized the handling of unstructured data, improving the functionality of electronic health record systems and facilitating more comprehensive patient care 5.

Several key algorithms and technologies underpin these advancements. Neural networks, particularly convolutional neural networks (CNNs), are extensively used for medical image analysis, aiding in the detection and characterization of various pathological findings 6. Decision support systems incorporate diverse data, including genetic profiles and prior health records, to optimize treatment strategies 7. Among the notable AI tools, IBM Watson stands out for its application in cancer treatment, although its widespread adoption faces challenges 8.

The integration of AI into healthcare brings substantial benefits, enhancing diagnostic and operational efficiencies while significantly reducing human error 9. AI’s ability to manage and analyze large datasets improves resource allocation and patient scheduling, directly impacting patient outcomes and satisfaction 10. In public health, AI’s predictive models are crucial for tracking disease patterns and managing epidemics, as evidenced during the COVID-19 pandemic 11. However, these advancements bring significant ethical and fairness challenges, particularly biases in training data that can lead to skewed AI models and disparities in healthcare outcomes among different demographic groups 12. For example, an algorithm used in US hospitals was biased against black patients in resource allocation 13, and dermatological AI showed lower diagnostic accuracy for conditions like melanoma in darker-skinned individuals due to training primarily on fair-skinned images  14. Similarly, AI tools for diagnosing depression have faced challenges when applied across different linguistic and cultural backgrounds because they were primarily trained on English-speaking, Western populations, leading to potential misdiagnoses in non-Western patients 15.

These examples highlight the urgent need for diverse datasets and transparent AI systems to ensure fairness and equity in healthcare delivery. Governments and regulatory bodies are responding to these challenges by crafting targeted policies and regulations. The European Union’s General Data Protection Regulation (GDPR) addresses privacy and ethical considerations in AI applications, setting a precedent for global data handling and protection standards 16. In the US, the Food and Drug Administration (FDA) is at the forefront of regulating medical AI applications, ensuring that new technologies are both safe and effective for clinical use 17.

This survey paper addresses the following questions:

  • How does bias manifest in AI systems used in healthcare?

  • What are the sources of bias?

  • What are the potential consequences of biased AI systems?

  • How do we address and mitigate these biases?

We investigate existing research on AI integration in healthcare, focusing on the emergence and origins of biases, their impact on outcomes, and strategies for detecting and mitigating these biases. The paper also delves into ethical and legal considerations related to AI in healthcare.

This paper, Section 2, discusses the applications of AI in various healthcare dimensions. The biases exhibited by those AI applications, their root causes, and other details have been discussed in Section 3. Section 4 explores various approaches to address and mitigate bias. Ethical and legal considerations are discussed in Section 5. Section 6 identifies research gaps and future directions for enhancing AI in healthcare. This paper is concluded in Section 7.

2 Applications of AI in Healthcare

AI has emerged as a transformative force in healthcare, revolutionizing various medical domains through advanced algorithms, data analytics, and machine learning techniques. AI’s ability to process vast amounts of data with precision and speed has opened new avenues for enhancing diagnostic accuracy, personalizing treatments, and improving patient outcomes. This section explores the diverse applications of AI across several key areas of healthcare, highlighting its impact and potential for future innovations.

2.1 Cardiology

AI has significantly transformed cardiology by enhancing diagnostic accuracy, personalizing treatment, and improving patient outcomes 76, 77. AI algorithms are adept at analyzing cardiovascular data to facilitate early detection and diagnosis of heart diseases, including arrhythmias, heart failure, and coronary artery disease 78, 79. For instance, machine learning models can interpret echocardiograms with high precision, often surpassing human experts in diagnostic speed and accuracy 80. AI also assists in managing cardiovascular risk by integrating diverse patient data to predict individual risk factors more effectively 81. Furthermore, AI plays a crucial role in developing personalized treatment plans by analyzing patient data and predicting responses to various treatment modalities, optimizing therapeutic decisions 82. The integration of AI in clinical practice necessitates rigorous validation and ethical alignment to ensure patient safety and data security 83. The continuous evolution of AI technologies promises further innovations in diagnostic tools and therapeutic strategies, enhancing precision and predictive capabilities in whole person care whereby patients may be prescribed lifestyle changes specific to their context as seamlessly as medications tuned to their genomics 84. Moreover, AI can contribute to health equity by making advanced cardiac care accessible in underserved areas through telemedicine platforms and remote monitoring, thereby reducing disparities in cardiovascular health outcomes 85.

2.2 Ophthalmology

AI has revolutionized ophthalmology by enhancing diagnostic accuracy, improving patient outcomes, and streamlining clinical workflows. More and more, AI algorithms, especially deep learning models, are being used to look at complicated visual data from imaging methods like fundus photography and optical coherence tomography (OCT) 86. These models are very good at diagnosing common eye diseases like diabetic retinopathy, glaucoma, and age-related macular degeneration; they can sometimes match or beat the performance of experts 87. Automated analysis using AI not only speeds up the diagnostic process but also reduces human error, facilitating earlier and more precise interventions 88. Furthermore, AI’s predictive capabilities are being harnessed to forecast disease progression, which is crucial for conditions like glaucoma, where early detection can prevent severe vision loss 6. Using AI tools in clinical practice also includes using AI-driven decision support systems that help plan treatments by guessing how different types of treatments will work 89. Despite these advancements, challenges remain, including data privacy concerns, the need for large annotated datasets for training algorithms, and the integration of AI into existing clinical workflows 90. However, ongoing research and collaboration between AI technologists and ophthalmic experts are likely to overcome these obstacles, solidifying AI’s role in modern ophthalmology 91. This further opens the frontier for expansion of access to ophthalmological care in existing deserts through technology-enabled care models and non-specialist operators at the point of service, thereby addressing disparities in eye care availability and quality.

2.3 Dermatology

AI is transforming dermatology by enhancing diagnostic accuracy, personalizing treatment plans, and streamlining patient management. AI-powered tools, particularly CNNs, are highly effective in diagnosing skin cancer, matching or surpassing dermatologists in identifying melanomas 92, 93. Moreover, AI applications in dermatology extend beyond cancer detection. They include the assessment and management of chronic conditions such as psoriasis and atopic dermatitis, where AI algorithms help in monitoring disease progression and response to treatment 94. AI systems are also utilized in cosmetic dermatology, optimizing treatment recommendations based on individual facial analysis 95. Ensuring the efficacy and safety of AI tools requires rigorous validation processes with diverse datasets to mitigate bias and enhance generalizability 14. Integration into clinical workflows and user training are crucial to maximizing the benefits of AI in dermatological practice 96. As AI continues to evolve, continuous collaboration between technologists and clinicians is essential to address ethical considerations and improve patient outcomes in dermatology. Importantly, AI can democratize access to high-quality dermatological care by enabling remote consultations and diagnostics, thus reducing barriers for individuals in remote or underserved regions.

2.4 Neurology

AI has significantly advanced the field of neurology by enhancing diagnostic accuracy, personalizing treatment plans, and facilitating neurological research. AI techniques, particularly machine learning and deep learning, are pivotal in diagnosing neurological disorders such as Alzheimer’s disease and epilepsy by analyzing imaging data with high precision 97, 5. These technologies improve the interpretation of MRI and CT scans, often achieving higher accuracy than traditional methods 6. In treatment, AI algorithms personalize therapies based on patient data, optimizing treatment outcomes for diseases like Parkinson’s 98. Additionally, AI has revolutionized neuroprosthetics by creating adaptive interfaces, significantly improving the quality of life for patients with motor disabilities 99. Predictive analytics also play a crucial role in anticipating disease progression for conditions such as multiple sclerosis 100. AI’s application extends into research, facilitating the understanding of complex neurological phenomena and the development of innovative treatments 101. Overall, AI’s integration into neurology not only improves clinical practices but also opens new avenues to address rare diseases as well as more finely tune medications to the individual 102. Furthermore, AI can enhance health equity in neurology by enabling remote diagnosis and monitoring, ensuring that patients in underserved areas receive timely and accurate neurological 103.

2.5 Radiology and Cancer Treatments

AI is revolutionizing radiology and cancer treatment by enhancing imaging accuracy, enabling personalized therapy, and improving diagnostic workflows. AI algorithms, particularly deep learning models, have significantly improved the precision of image interpretation in radiology, aiding in the early detection and diagnosis of various cancers, such as breast and lung cancer 104. These models analyze imaging data much more rapidly than human radiologists, reducing diagnostic times and increasing efficiency 105. In cancer treatment, AI assists in formulating personalized treatment plans based on patient data and predictive analytics, which can predict treatment outcomes and suggest optimal therapies 106. AI applications in radiology extend to prognostic evaluations, predicting disease progression and survival rates, refining treatment protocols and developing follow-up strategies  107. These advancements are supported by large datasets and strict validation processes, ensuring reliability and clinical applicability 108.

2.6 Emergency Medicine and Critical Care

AI stands to significantly transform emergency medicine and critical care by enhancing patient triage, treatment efficacy, and time to diagnosis. As AI-powered tools like diagnostic algorithms have been incredibly good at the early identification of conditions like sepsis and its progression, there is the potential to improve patient outcomes 109, 110. The prediction of patient outcomes has also positioned AI to facilitate the optimization of critical resource allocation. During the COVID-19 pandemic, AI-based self-triage tools had a critical role to play in predicting cases and hospitalizations at the population level 111. AI-powered chatbots and virtual assistants are streamlining patient intake and symptom assessment 112, 113. Furthermore, AI’s role in triaging patients based on urgency has revolutionized care prioritization, improving ED operations and patient flow 114, 115. These advancements underscore AI’s potential to not only enhance emergency medical services but also pave the way for a more efficient, accurate, patient-centered healthcare system that extends care beyond the traditional brick-and-mortar facilities.

3 Fairness Concerns in Healthcare

This section delves into various fairness concerns that arise during the application of AI in healthcare. Initially, it examines the sources of bias in AI healthcare systems, including data bias, algorithm bias, explicit bias, implicit bias, and selection bias. Furthermore, it will discuss the potential consequences of these biases, such as misdiagnosis, inequitable healthcare outcomes, loss of trust in healthcare systems, legal and ethical implications, resource misallocation, and stifling innovation. By delving into these topics, this section aims to underscore the critical need to address biases to ensure AI technologies contribute to fair and equitable healthcare for all.

3.1 Sources of bias

Data Bias: Data bias in AI healthcare systems can severely compromise their effectiveness and fairness. For instance, a study by Obermeyer et al. 12 found that a healthcare management algorithm disproportionately favored white patients over black patients with similar health conditions due to reliance on historical healthcare cost data, which reflected existing socioeconomic disparities. Further studies reinforce the pervasive nature of data bias in healthcare AI. For example, Rajkomar and his team found that the models trained on patient datasets from one geographical region often failed to generalize to populations outside that region, indicating a geographic data bias that limits the practical utility of AI systems in diverse settings 4. Additionally, the AI system used for diagnosing diabetic retinopathy performed well in tests but showed decreased accuracy in clinical settings, especially among minority populations not well represented in the training data, underscoring the impact of demographic data biases 116.

Algorithm Bias: Algorithmic bias in AI healthcare occurs when AI systems develop systematic errors that disadvantage certain patient groups. For example, AI models used to diagnose pneumonia displayed biases against specific racial groups, failing to capture important socio-economic and demographic factors despite extensive training on hospital data. This led to inappropriate treatment recommendations 4, 13, 117. Another study in dermatology faced the same issue. The CNNs used for detecting skin cancer were less effective for individuals with darker skin. This discrepancy arose because the training data predominantly consisted of images featuring lighter skin tones, which does not reflect the diversity of global skin types. Consequently, this training limitation can lead to poorer health outcomes for underrepresented groups 14.

Explicit Bias: Explicit bias in AI systems within healthcare refers to observable prejudices that directly influence these systems’ decision-making processes. These biases are often a result of human prejudices being encoded in to AI through biased data or algorithms, leading to discriminatory practices and unequal treatment of patients 118. One example of explicit bias in healthcare AI is the underdiagnosis of certain demographic groups by AI models. A study investigating the fairness of vision language foundation models in medical imaging found that these AI systems consistently underdiagnosed marginalized groups, such as black female patients, across a range of pathologies 119. This explicit bias in diagnostic AI tools can lead to delayed or inadequate medical care for specific patient subgroups, exacerbating healthcare disparities. Underlying Clinical Decision Support systems, explicit bias may be due to the selection of data and algorithm design, leading to biased outcomes. For example, the decreased likelihood of receiving diagnostic imaging for nonwhite groups compared to their white counterparts when controlling for patient and facility level factors is embedded in clinical data 120. Accordingly, models predicting imaging use may reflect the human prejudices around who deserves and receives diagnostics 121. Moreover, if an AI system is trained predominantly on data from one ethnic group or gender, it may not perform accurately for other ethnicities or gender. The use of biased AI in healthcare can result in systematic discrimination, where certain groups received suboptimal care due to the AI’s prejudiced programming. This can manifest in various ways, such as through biased diagnostic tools, treatment recommendations, or predictive analytics that unfairly disadvantage certain groups.

Implicit Bias: Implicit bias in AI systems in healthcare refers to the unconscious prejudices that can influence decision-making processes without explicit acknowledgment. These biases often reflect societal stereotypes and may not align with the conscious values or beliefs of the developers or users of the AI system. Implicit biases can alter perceptions and judgments, leading to disparities in healthcare delivery and outcomes 119. In healthcare settings, implicit biases can manifest in various ways. For example, a study on obstetric care providers found that implicit biases favored French women over African migrant women, potentially influencing medical decisions and leading to healthcare disparities 122. Similarly, healthcare professionals have been shown to exhibit implicit biases based on ethnicity, which could affect the likelihood of patients being given another opportunity to attend an outpatient clinic after missing an appointment 123. Another example is the use of AI in diagnostic tools, where implicit biases can result in differential accuracy rates across racial or ethnic groups. This can lead to underdiagnosis or misdiagnosis for certain populations, contributing to unequal treatment and outcomes 124. Implicit biases can also affect the development of clinical decision support systems. This is because the data used to train AI might not fully reflect the diversity of the patient population, which could lead to recommendations and treatment plans that aren’t based on facts 125, 126. Furthermore, implicit biases can impact the patient-provider relationship, where unconscious prejudices may influence the level of care and communication provided to patients of different racial or ethnic backgrounds, potentially affecting their health outcomes 127.

Selection Bias: Selection bias occurs when the training data for an AI model is not representative of the broader population or the context in which the model will be applied. This leads to AI systems performing well on training data but poorly in real-world settings, particularly for underrepresented groups 128. The development of AI models for disease prediction or diagnosis, primarily trained on data from specific demographic groups like a particular race, gender, or age group, is one example of selection bias in healthcare settings. If an AI model for diagnosing skin cancer is trained primarily on images of lighter-skinned individuals, it may perform poorly when diagnosing conditions on darker skin, leading to higher misdiagnosis rates for individuals with darker skin tones 125. AI models that predict patient outcomes or treatment efficacy also exhibit selection bias. If the historical data used to train these models predominantly come from patients who received a certain level of care or who have specific socioeconomic backgrounds, the AI’s predictions may not accurately reflect outcomes for patients outside of these groups 128. This can lead to disparities in healthcare recommendations and interventions offered to patients from different socioeconomic or racial backgrounds. Furthermore, selection bias can also manifest in AI-driven tools for patient triage and resource allocation. If the data used to train these models does not adequately account for the diversity of patient presentations and conditions, the AI system may prioritize certain patient groups over others, inadvertently exacerbating existing healthcare disparities.

3.2 Potential consequences of biases

Biases in AI systems, particularly in healthcare, can have profound and far-reaching consequences. These biases usually stem from the data on which the AI systems are trained, the design of the algorithms themselves, and the contexts in which they are deployed (as discussed in Section 3.1). Here are several potential consequences of biases in AI systems in healthcare:

Misdiagnosis and Inequitable Healthcare Outcomes: AI integration in healthcare brings significant advancements but also risks of misdiagnosis and inequitable outcomes due to biased algorithms. AI systems often rely on non-representative data, leading to biased decision-making. For instance, Obermeyer highlighted that an algorithm used in healthcare disproportionately favored white patients over black patients because it used healthcare costs as a proxy for healthcare needs, indirectly embedding racial biases in its predictions 12. Moreover, these biases in AI can exacerbate existing healthcare disparities. Rajkomar and others discussed how AI applications, if not carefully designed and monitored, could inherit and amplify socioeconomic and racial disparities. This is particularly problematic in diagnostics, where AI systems are trained predominantly on data from specific demographic groups, potentially leading to poorer diagnostic accuracy for underrepresented groups 4. This was evident in a study by Adamson and Smith, which found that dermatology AI systems demonstrated lower accuracy rates in skin lesion diagnosis for dark-skinned individuals compared to those with lighter skin 14. Such biased AI not only risks misdiagnosis but also contributes to inequitable healthcare outcomes by potentially steering healthcare resources away from those who may need them most. Vayena and his team emphasized the ethical imperative to ensure that AI tools in healthcare are developed with consideration for fairness and equity, advocating for diverse and inclusive data sets and algorithmic transparency to mitigate these biases 129.

Loss of trust in Healthcare Systems: Recent studies have echoed the concern that the deployment of biased AI in healthcare could significantly undermine public trust in medical systems. Trust is foundational to the patient-provider relationship and is crucial for the effective delivery of healthcare services 130. When AI tools exhibit bias, whether in diagnosis, treatment recommendations, or patient management, they can lead to misdirected care, fostering distrust among patients, particularly in marginalized communities 4. Kherbache and his team pointed out that when patients perceive AI-driven processes as opaque or unfair, their trust in the overall healthcare system may decline 131. Incidents of AI failures that receive public attention can exacerbate this erosion of trust, leading patients to question the reliability and ethics of using AI in medical decision-making. Researchers like Kerasidou argue that trust is not only about the accuracy of AI but also its alignment with ethical principles that govern healthcare, such as beneficence and non-maleficence 132. Furthermore, the lack of transparency in how AI models make decisions can be a major barrier to trust. Blease’s study 133 suggests that without clear communication about how AI tools contribute to healthcare decisions, patients may become skeptical of diagnoses and treatments, fearing that their personal healthcare data could be misused or misunderstood. This potential trust deficit could have severe implications, not just for individual health outcomes but also for public health at large, as mistrust in healthcare systems can lead to lower rates of healthcare utilization, vaccine hesitancy, and poor adherence to medical advice 134. Veinot highlights the need for healthcare systems to maintain high standards of accountability and transparency as they integrate AI technologies to mitigate such risks 135.

Legal and Ethical Implications: The use of biased AI in healthcare not only poses clinical risks but also entails significant legal and ethical implications. Deploying biased AI systems could lead to legal breaches of anti-discrimination laws. In the United States, for example, the Civil Rights Act and the Americans with Disabilities Act set legal standards that could be violated by biased AI algorithms, which fail to provide equitable care across different patient demographics 136. Ethically, biased AI conflicts with the fundamental medical ethics principles of justice and non-maleficence, demanding fairness and avoidance of harm, respectively 137. Moreover, biased algorithms in healthcare could potentially expose medical practices and institutions to litigation related to malpractice or negligence, especially if these algorithms contribute to substandard care outcomes 138. For instance, if an AI system were to consistently provide inferior diagnostic support for certain racial groups, this could be viewed as a form of systemic negligence or malpractice 129. Additionally, the ethical implications extend to the breach of patient trust and the compromise of patient autonomy. Ethical medical practice relies heavily on the principles of informed consent and respect for patients’ autonomy—principles that are challenged by opaque AI systems that do not make their decision-making processes or inherent biases clear to patients or practitioners 133. These potential legal and ethical failures underscore the necessity for rigorous oversight and transparent development processes for AI in healthcare, aiming to ensure that these technologies adhere to both existing legal frameworks and ethical standards of practice.

Resource Misallocation: The deployment of biased AI in healthcare settings can lead to resource misallocation, a critical issue that impacts both the efficiency and fairness of medical services. Biased algorithms may misdirect resources by prioritizing certain groups over others based on flawed data inputs or biased training procedures. For example, a study by Obermeyer demonstrated how an algorithm used for managing healthcare resources inadvertently favored healthier white patients over sicker black patients due to biased data inputs that did not accurately reflect patient needs 12. This misallocation can exacerbate existing healthcare disparities by diverting necessary medical attention and resources away from those who are most in need. As Chen pointed out, such disparities are not just a matter of clinical outcomes but are deeply tied to social and economic inequalities that AI tools can inadvertently perpetuate 13. Furthermore, biased AI can influence the allocation of resources within healthcare facilities, potentially resulting in inefficiencies that strain healthcare systems. This includes misallocating medical staff, diagnostic tools, and hospital beds, which can degrade the quality of care delivered while increasing wait times and healthcare costs 4. Resource misallocation also raises ethical questions about fairness and equity in healthcare provisioning. It challenges the ethical principle of justice, which demands that healthcare resources be distributed based on need rather than biased algorithms 137. This ethical breach can lead to further mistrust and reluctance among underserved populations to engage with healthcare systems, perpetuating a cycle of disadvantage.

Stifling Innovation: The presence of bias in AI systems used in healthcare not only affects the accuracy and fairness of medical services but can also stifle innovation. When AI models are made using biased data, they might not accurately reflect the different needs of the population. This could make these innovations less useful and applicable to a wider range of demographic groups. Liang talks about how biased training datasets in AI development make it harder for these systems to work with different types of data. This could stymie innovation by preventing it from solving larger or more complex health problems that affect many people. Moreover, reliance on biased AI can deter investment in developing new technologies that are inclusive and equitable. Potential investors may be cautious about funding projects that might not meet regulatory standards for fairness or could lead to public backlash — a concern highlighted by Corbett-Davies and Goel, who note the legal and social implications of deploying biased AI 139. In addition, the perpetuation of bias in AI could solidify existing disparities in healthcare innovation. As Benjamin notes, if innovation is directed predominantly at solving problems specific to well-represented groups, less common but equally pressing issues in underrepresented groups are neglected, thereby limiting the scope and impact of technological advancements in healthcare 140. The studies of Rajkomar 4, Hardt 141, and Howell 142 made things even more complicated by pointing out that biased algorithms can make it harder to find new treatments that work for everyone. This feeds a cycle of innovation that helps groups that are already overrepresented in data sets more than others. This restricted focus on innovation not only affects the equity of healthcare delivery but also limits the development of truly innovative, comprehensive healthcare solutions. These factors combined suggest that biased AI not only hinders the progression of medical technologies but also potentially locks the healthcare sector into a cycle of uneven innovation where only the needs of the majority are systematically addressed.

4 Addressing and Mitigating Unfairness in AI

As AI technologies increasingly influence healthcare delivery, it becomes essential to scrutinize and refine these systems to prevent disparities in care and outcomes. This section explores various bias detecting and mitigating strategies in AI systems in healthcare.

4.1 Bias Detection Methods

In healthcare, detecting bias in AI systems is crucial, as it can result in disparities in healthcare delivery and outcomes. This section provides an overview of bias detection strategies in healthcare AI system, including: statistical analysis, auditing tools, and feedback from end-users.

4.1.1 Statistical Analysis:

In the realm of AI healthcare, detecting bias through statistical analysis is a critical step towards ensuring equitable healthcare outcomes across diverse populations. Regular statistical reviews serve as a foundational approach to identifying and addressing disparities in AI outputs, thereby improving the fairness and effectiveness of AI-driven healthcare solutions 123. The systematic review on bias detection and mitigation strategies in electronic health record-based models shows how important it is to find different types of bias that can affect AI models created using EHR data. This comprehensive analysis highlights the necessity of employing fairness metrics such as statistical parity 143, equal opportunity 144, and predictive equity 145 to detect implicit and algorithmic biases effectively. Using chi-squared statistical analysis, differences in preconception health indicators have been found based on age, race or ethnicity, education level, living in a city, and income. This shows how important it is to do regular statistical reviews to find healthcare inequalities 146. Similarly, a study of the clinical predictive values for radiographic pneumonia in children showed the sensitivity and specificity of commonly assessed signs and symptoms. This gave us information about how well AI models can be used for diagnosis in pediatric healthcare 147, 148. The epidemiological study of autism spectrum disorders in Greece used statistical analysis to find differences in the rates of ASD across different regions. This shows how useful statistical methods are for finding biases in healthcare datasets 149. Additionally, a prospective diagnostic evaluation comparing self-taken and healthcare worker-taken swabs for rapid COVID-19 testing highlighted the accuracy of AI-driven diagnostic tools, further illustrating the importance of statistical analysis in evaluating AI model performance 150. Furthermore, the exploration of AI in enhancing diagnosis, treatment, and healthcare systems in India emphasizes the potential of AI to revolutionize healthcare while also acknowledging the challenges posed by biases and ethical considerations 151. This underscores the need for rigorous statistical analysis to ensure that AI-driven healthcare solutions are both effective and equitable.

4.1.2 Auditing Tools

The Disparities Impact Statement (DIS): An innovative tool, the DIS 152, evaluates healthcare policies and practices for potential biases. It assesses how different population groups might be differently affected by a healthcare policy or practice, aiming to preemptively identify disparitiesṪhe tool involves qualitative and quantitative analyses to ensure a comprehensive review of the policy’s implications across various demographics 153. By implementing DIS, healthcare institutions can address disparities at the policy formulation stage, promoting equity in healthcare delivery 154.

PROGRESS Framework: The PROGRESS framework 155, which stands for Place of Residence, Race, Occupation, Gender, Religion, Education, Socioeconomic Status, and Social Capital, is a tool used to identify and evaluate the factors contributing to health inequalities. This framework encourages researchers and practitioners to systematically consider these dimensions when developing and implementing studies or interventions in healthcare settings 156. It is particularly useful in designing public health interventions and assessing the socio-economic and demographic factors influencing health outcomes 157.

Implicit Association Test (IAT): The IAT 158 is employed in healthcare to detect unconscious biases among healthcare providers that may affect patient care and treatment outcomes. Originally developed in psychological research to explore unconscious biases, IAT has been adapted for healthcare to assess biases in race, age, and obesity, among other aspects 159. Studies utilizing IAT have demonstrated that even well-intentioned staff can harbor biases that influence their decisions and interactions with patients 160.

Healthcare Disparities Dashboard: The Healthcare Disparities Dashboard 161 is a data-driven tool that aggregates patient data to highlight discrepancies in healthcare outcomes, utilization, and satisfaction across different demographic groups. It allows healthcare organizations to visualize and monitor disparities, thus aiding in targeted interventions 162. By presenting data in an accessible format, the Dashboard supports ongoing monitoring and the proactive management of healthcare equity 163.

Equity Quality Improvement (EQI) Tool: Healthcare systems design the EQI tool 164 to integrate equity into their quality improvement frameworks. This tool helps identify areas where healthcare disparities exist and suggests practical interventions to mitigate these disparities. The EQI tool facilitates the inclusion of equity as a central quality dimension, ensuring that improvements in healthcare delivery uniformly benefit all patient groups 165.

Cultural Competence Assessment (CCA): The CCA tool 166 helps healthcare institutions assess and enhance their effectiveness in providing culturally sensitive care. This tool evaluates healthcare providers’ awareness of and sensitivity to the cultural needs of diverse patient populations. The CCA can lead to improved patient satisfaction by fostering better communication and understanding between healthcare providers and patients from diverse backgrounds 167.

4.1.3 Feedback from End-Users

Collecting and analyzing feedback from healthcare providers who use AI systems is essential for identifying and mitigating biases that may affect clinical decisions and patient outcomes. Healthcare professionals are at the forefront of witnessing the operational aspects of AI, and their firsthand insights can reveal how these systems function across diverse clinical scenarios 5. Studies show that user feedback is invaluable for improving AI models, especially to ensure that these tools are equally effective and free from bias across different populations 168. By engaging healthcare providers in the evaluation process, developers can gather critical data on AI’s practical implications and areas where it may unintentionally perpetuate disparities 135. Moreover, incorporating such feedback facilitates the development of more robust, equitable, and contextually appropriate AI solutions, ultimately leading to better patient care and outcomes 4. As a result, systematic feedback mechanisms are not only beneficial but also necessary for the iterative refinement and ethical deployment of healthcare AI systems 137.

4.2 Mitigation Strategies

Mitigating biases in AI systems within healthcare is crucial for ensuring equitable treatment across all patient demographics. This section outlines key strategies to address and reduce biases, including the use of diverse and representative data and fairness-aware approaches, thereby enhancing the fairness and accuracy of AI models.

4.2.1 Diverse and Representative Data

Mitigating biases in AI systems within healthcare is crucial for ensuring equitable treatment across all patient demographics. A fundamental strategy involves using diverse and representative datasets in the training phase of these systems. Research indicates that AI models can only be as unbiased as the data they are trained on 169. Thus, inclusion of comprehensive data from a wide array of patient groups, particularly those historically underrepresented in medical research, is essential 13.

The significance of representative data extends beyond just including diverse groups; it also involves the depth and quality of the data collected from these groups. Data must capture a breadth of variables that influence health outcomes, such as socioeconomic factors, environmental conditions, and genetic differences 4. This approach helps in creating models that are better tuned to the nuances of various patient needs and conditions 14.

Additionally, involving stakeholders from diverse backgrounds in the data collection and AI development process can enhance the relevancy and sensitivity of the datasets 170. This inclusion helps in identifying and addressing potential blind spots in AI training datasets, which, if overlooked, could perpetuate biases and inequalities in healthcare delivery 171. Employing these strategies not only improves the fairness and effectiveness of AI tools but also builds trust in these technologies among all user groups, thereby fostering a more inclusive healthcare ecosystem 137.

4.2.2 Fairness-aware approaches

Incorporating various fairness-aware approaches is crucial for developing AI systems that are equitable and unbiased. These approaches can be categorized into three main strategies: pre-processing, in-processing, and post-processing. Each strategy targets different stages of the machine learning pipeline to address potential biases and ensure fair treatment across diverse demographic groups.

Pre-Processing:

In healthcare, pre-processing involves techniques to adjust the input data for AI models to ensure it accurately represents diverse patient demographics before model training begins. This process aims to eliminate any inherent biases that may skew AI predictions and outcomes, thus fostering equity in healthcare treatments and diagnostics 172. By refining the dataset upfront, pre-processing helps in building AI systems that perform fairly across all patient groups 13.

  • Re-Sampling: Re-sampling is a common pre-processing method that manipulates data samples to create a more balanced dataset. This can be done through either over-sampling minority classes or under-sampling majority classes 173. An example of this in healthcare is the use of the Synthetic Minority Oversampling Technique (SMOTE) 174 in medical datasets where certain conditions or outcomes are rare. Chawla applied SMOTE to a dataset of imbalanced classes to improve the prediction of minority-class outcomes in medical diagnostics, resulting in more reliable predictive models for underrepresented conditions. This technique helps to ensure that the AI does not become biased towards the more frequent classes.

  • Re-Weighting: To reduce bias, instances in the training data are reweighted. We assign a weight to each instance in a dataset based on its representation, thereby increasing the influence of underrepresented groups in the model training process. This technique was demonstrated by Calmon in 175, who developed a data pre-processing method that modifies features and outcomes to improve fairness. In healthcare, this could mean adjusting weights so that data from minority ethnic groups has greater influence on the model, helping to prevent the AI from developing biases that might affect diagnosis or treatment recommendations 172.

In-Processing:

In-processing methods involve integrating fairness directly into the learning algorithm itself. This technique adjusts the algorithm during the training phase to minimize bias. Zemel in his study introduced a method where they developed a representation for data that is invariant to protected attributes like race or gender while maintaining the informative characteristics necessary for prediction 176. In healthcare, this can be crucial for ensuring that diagnostic or treatment recommendations are not adversely skewed by underlying biases in the training data 177.

  • Adversarial Debiasing: Adversarial debiasing involves the simultaneous training of a predictor and an adversary. The predictor learns to make accurate predictions, and the adversary learns to determine whether the predictions are biased toward certain groups. Deldjoo et al. 177 explored this in the context of healthcare, aiming to create clinical decision support systems that are unbiased towards patients’ demographic attributes. This technique helps ensure that the AI’s treatment recommendations do not reflect discriminatory patterns that might exist in the training data 178.

  • Constraint-Based Optimization: Constraint-based optimization involves modifying the learning algorithm to satisfy fairness constraints while minimizing prediction error. You can implement this method by adding fairness constraints to the objective function the algorithm is optimizing. Elzayn et al. 179 applied this method to healthcare datasets, ensuring that treatment recommendations do not disproportionately favor one group over another by enforcing demographic parity or equality of opportunity as constraints during model training 180.

Post-Processing:

After training a model, post-processing techniques are applied to modify the outputs of AI systems, ensuring fair treatment across different demographic groups. These adjustments address any residual biases to promote equitable outcomes. For example, Hardt et al. 144 applied different decision thresholds for various groups to equalize treatment effects. This method effectively adjusted risk scores in a healthcare dataset, ensuring that treatment recommendations were equitable 181.

  • Threshold Adjustment: Threshold adjustment in AI healthcare involves setting an optimal cutoff point for predictive models to classify outcomes effectively, such as predicting disease presence or patient risk levels. A significant example of this is in cardiovascular disease prediction. In a study by Siontis 182, the researchers examined the application of different thresholds for predicting cardiovascular events using a logistic regression model. They aimed to optimize the sensitivity and specificity of predictions by adjusting the threshold to better serve clinical decision-making. This adjustment proved crucial in identifying higher-risk patients who might benefit from preventive treatments. Another notable case is in breast cancer screening, where Fenton 183 applied threshold adjustment to mammography interpretation models. By altering the decision threshold, they were able to reduce false positives without substantially missing cases of cancer, thereby balancing the need for early detection with the risk of overdiagnosis and unnecessary anxiety for patients.

  • Output Recalibration: Output recalibration in AI healthcare is critical for adapting predictive models to local contexts and ensuring that the predicted probabilities match actual clinical outcomes in a new patient population. A notable case involves the recalibration of a sepsis prediction model at Johns Hopkins Hospital, as detailed by Henry et al. 184. The model was originally developed with a dataset from one patient demographic, but recalibration using local demographic data was required to maintain accuracy across diverse populations within the hospital. This recalibration helped align the model’s predictions with the actual rates of sepsis observed, improving clinical decision-making and patient care. Another example is from the University of California, San Francisco (UCSF), where Rajkomar 4 recalibrated Google’s deep learning model used for predicting medical outcomes. The model, initially trained on a national dataset, was recalibrated with UCSF-specific patient data to ensure its predictions accurately reflected the local patient population’s risk factors and outcomes. This was particularly important for diseases with variable manifestations across different demographics.

5 Ethical and legal considerations

This section discusses the various ethical and legal considerations that are necessary to reduce bias in AI systems in healthcare. It will cover ethical issues such as privacy and data protection, bias in AI algorithms, transparency and explainability of AI decisions, potential displacement of healthcare professionals, and accountability for AI-driven decisions. Additionally, it will explore legal aspects including intellectual property rights, liability for AI-induced errors, compliance with healthcare regulations, and informed consent for AI involvement in patient care.

5.1 Ethical Considerations

The integration of AI in healthcare presents significant ethical considerations that necessitate careful scrutiny. One primary concern is privacy and data protection, as AI systems typically require large datasets, including sensitive patient information, to train and operate effectively. Ensuring the confidentiality and security of this data against breaches is paramount 185. Additionally, the issue of bias in AI algorithms poses a serious ethical challenge. AI systems can perpetuate and even exacerbate existing biases if they are trained on non-representative data, leading to unequal treatment outcomes among different demographic groups 4. There is also the ethical issue of transparency and explainability. Practitioners and patients may find it challenging to comprehend the reasoning behind AI decision-making processes due to their opaque nature. This lack of transparency can undermine trust in AI systems and complicate informed consent processes 186. Furthermore, the deployment of AI in healthcare must take into account the potential displacement of healthcare professionals, raising concerns about job security and the loss of valuable human expertise and empathy in patient care 187. Lastly, accountability for AI-driven decisions in healthcare is a critical ethical issue. It is critical to define who is responsible—the developers, the users, or the AI itself—when an AI system’s decision results in patient harm 188.

5.2 Legal Considerations

The deployment of AI in healthcare brings forth numerous legal considerations that are crucial for ensuring compliance and safeguarding patient rights. Intellectual property rights stand out prominently, as there is ongoing debate over whether AI-generated medical inventions can be patented and who owns the rights to AI-generated data and algorithms 138. Furthermore, liability issues arise when AI systems make erroneous decisions that could harm patients. Determining whether the healthcare provider, the software developer, or another party is liable requires intricate legal analysis and possibly new legal frameworks 189. Compliance with medical standards and regulatory requirements is another critical legal area. AI applications in healthcare must adhere to existing healthcare regulations, such as the Health Insurance Portability and Accountability Act (HIPAA) in the U.S., which governs the privacy and security of patient data, and the General Data Protection Regulation (GDPR) in the EU, which sets stringent guidelines for data protection and privacy 190. As AI technologies evolve, there may also be a need for specific regulations that address AI’s unique aspects in healthcare to ensure that these innovations are safely and effectively integrated into medical practice 191. Lastly, issues of informed consent are magnified with AI, as patients must understand the role of AI in their care, including how their data will be used and the implications of AI-driven decisions 188.

6 Research gaps and Future directions

This section delves into the research gaps in implementing fair AI in healthcare and future directions to enhance its efficacy. Ensuring fairness in AI systems within healthcare is a complex and multifaceted challenge that requires addressing various deficiencies and promoting interdisciplinary collaboration.

6.1 Diversity in training datasets

The lack of diversity in training datasets is a significant research gap that impacts AI fairness in healthcare. Models trained on non-representative data can perpetuate or even exacerbate existing biases against underrepresented groups 4. To address this, techniques for synthetically augmenting underrepresented classes in datasets or incentivizing the collection of more comprehensive and inclusive data must be developed 117. Additionally, implementing algorithmic fairness approaches during model training, such as fairness constraints or adversarial debiasing, can help mitigate biases introduced by skewed datasets 192. Future directions also include promoting policy changes that enforce transparency and fairness in data collection and algorithmic processing in healthcare applications 137.

6.2 Longitudinal Effects of AI Systems

The longitudinal effects of AI systems on healthcare fairness remain underexplored, with existing studies often focusing on short-term outcomes and immediate biases in algorithmic decision-making 193. Understanding how these technologies affect health disparities over time is crucial, as initial biases can amplify if not properly addressed 4. Future research should involve continuous monitoring of AI systems post-deployment to assess their impact on various demographic groups across different time intervals 194. This approach will help identify evolving biases and enable timely modifications to algorithms, thereby ensuring more equitable healthcare outcomes 12. Additionally, incorporating feedback loops that allow healthcare providers to report disparities can further refine AI applications in real-world settings.

6.3 Interdisciplinary Approaches to Fair AI

The integration of AI in healthcare requires interdisciplinary approaches that combine insights from medical ethics, data science, social sciences, and clinical practice to comprehensively address fairness 195. However, current AI applications often lack this multidimensional perspective, resulting in solutions that are technically adequate but socially or ethically inappropriate 137. Future research should focus on frameworks that incorporate not only technical accuracy but also ethical, legal, and social implications from the outset, ensuring that AI systems are developed with a holistic view of fairness 135. These collaborative frameworks can help prevent the exacerbation of existing disparities and promote a more equitable distribution of healthcare resources 13.

6.4 Transparency in AI Algorithms

Transparency in AI algorithms is a significant research gap that limits broader adoption and trust in AI-driven healthcare systems. The inherent “black box” nature of many advanced machine learning models, such as deep neural networks, makes their decision-making processes difficult to understand, leading to issues in fairness and accountability 196. To address this, future research should focus on developing “glass-box” approaches that are inherently interpretable, such as decision trees or linear models, fostering greater trust among healthcare providers and patients 197. Additionally, implementing model documentation standards like Model Cards can provide essential transparency by detailing the capabilities and limitations of AI models to all stakeholders 198.

6.5 Impact of AI on Healthcare Access

The impact of AI on healthcare access represents a critical research gap, as disparities in AI-driven healthcare solutions can exacerbate existing inequalities. Studies suggest that geospatial AI models, when applied to geographic information system (GIS), can uncover regions with inadequate healthcare services, thus guiding interventions to those most in need 135. However, predictive analytics must be carefully designed to ensure they do not perpetuate biases present in underlying data 12. Integrating feedback mechanisms from healthcare providers into AI systems offers a pathway to refine these models, ensuring they adapt to serve diverse populations effectively. The future direction in bridging this gap involves developing AI technologies that are not only predictive but also prescriptive, providing actionable insights to improve universal healthcare accessibility 4.

6.6 Interpretability of AI Decisions

The interpretability of AI decisions is a significant research gap in ensuring fairness in healthcare applications. Complex AI models, such as deep neural networks, often act as “black boxes”,making it difficult for healthcare providers to understand and trust their outputs 197. Enhanced interpretability is critical for trust and verifying the fairness of AI outcomes across diverse patient groups 196. Future research should focus on adopting model-agnostic explanation frameworks like LIME 196 or SHAP 199, which provide insights into the decision-making processes of any AI model, regardless of its underlying architecture. Developing these tools will enable clinicians to better understand and oversee AI-driven decisions, potentially leading to more equitable healthcare outcomes.

7 Conclusion

In conclusion, AI in healthcare holds transformative potential for improving diagnosis, treatment, and patient management. However, its integration also brings significant ethical, legal, and operational challenges, especially regarding fairness and equity. To balance the benefits of AI with these challenges, it is essential to address biases, ensure ethical practices, and foster interdisciplinary collaboration. Addressing biases involves implementing more inclusive data collection practices to ensure diverse representation and developing bias-aware AI models to mitigate disparities in healthcare delivery. Ensuring ethical practices requires establishing robust regulatory frameworks to oversee AI implementation, guaranteeing compliance with ethical standards and legal requirements. Additionally, promoting transparency in AI decision-making processes is essential to building trust among healthcare providers and patients. To achieve this goals, fostering interdisciplinary collaboration is vital. Encouraging ongoing cooperation between technologists, clinicians, legal experts, and ethicists will help address the multifaceted challenges of AI in healthcare. Interdisciplinary research should focus on developing innovative and inclusive AI applications that effectively serve diverse populations without perpetuating existing disparities. Future research should aim to refine AI tools to ensure their effectiveness across diverse populations, continuously monitoring and adjusting to evolving biases. Investigating the long-term impacts of AI systems on healthcare equity is crucial for making necessary adjustments over time. By addressing these critical areas, we can harness the capabilities of AI to create a more equitable and effective healthcare system. As we advance, it is imperative to remain vigilant about the ethical implications of AI, ensuring that these technologies benefit all segments of society equally. This requires a committed effort to integrate fairness, transparency, and inclusivity into every stage of AI development and deployment in healthcare.

References

  • 1 Shortliffe EH, Davis R, Axline SG, Buchanan BG, Green CC, Cohen SN. Computer-based consultations in clinical therapeutics: explanation and rule acquisition capabilities of the MYCIN system. Computers and Biomedical Research. 1975;8(4):303–320.
  • 2 Miller R. INTERNIST-1/CADUCEUS: problems facing expert consultant programs. Methods of information in medicine. 1984;23(01):9–14.
  • 3 Esteva A, Robicquet A, Ramsundar B, et al. A guide to deep learning in healthcare. Nature medicine. 2019;25(1):24–29.
  • 4 Rajkomar A, Dean J, Kohane I. Machine learning in medicine. New England Journal of Medicine. 2019;380(14):1347–1358.
  • 5 Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke and vascular neurology. 2017;2(4).
  • 6 Litjens G, Kooi T, Bejnordi BE, et al. A survey on deep learning in medical image analysis. Medical image analysis. 2017;42:60–88.
  • 7 Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Computational and structural biotechnology journal. 2015;13:8–17.
  • 8 Strickland E. IBM Watson, heal thyself: How IBM overpromised and underdelivered on AI health care. IEEE Spectrum. 2019;56(4):24–31.
  • 9 Jha S, Topol EJ. Adapting to artificial intelligence: radiologists and pathologists as information specialists. Jama. 2016;316(22):2353–2354.
  • 10 Feldman B, Martin E. Big Data in Healthcare Hype and Hope. In: 2012.
  • 11 Bachtiger P, Peters NS, Walsh SL. Machine learning for COVID-19—asking the right questions. The Lancet Digital Health. 2020;2(8):e391–e392.
  • 12 Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–453.
  • 13 Chen IY, Szolovits P, Ghassemi M. Can AI help reduce disparities in general medical and mental health care?. AMA journal of ethics. 2019;21(2):167–179.
  • 14 Adamson AS, Smith A. Machine learning and health care disparities in dermatology. JAMA dermatology. 2018;154(11):1247–1248.
  • 15 Parikh RB, Teeple S, Navathe AS. Addressing bias in artificial intelligence in health care. Jama. 2019;322(24):2377–2378.
  • 16 Voigt P, Bussche V. dA. The eu general data protection regulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing. 2017;10(3152676):10–5555.
  • 17 Food , Administration D, others . Proposed regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD). 2019.
  • 18 Wang Z, Zhou Y, Qiu M, et al. Towards fair machine learning software: Understanding and addressing model bias through counterfactual thinking. arXiv preprint arXiv:2302.08018. 2023.
  • 19 Yin Z, Wang Z, Zhang W. Improving Fairness in Machine Learning Software via Counterfactual Fairness Thinking. In: 2024:420–421.
  • 20 Dzuong J, Wang Z, Zhang W. Uncertain Boundaries: Multidisciplinary Approaches to Copyright Issues in Generative AI. arXiv preprint arXiv:2404.08221. 2024.
  • 21 Chinta SV, Fernandes K, Cheng N, et al. Optimization and Improvement of Fake News Detection using Voting Technique for Societal Benefit. In: 2023:1565–1574.
  • 22 Wang Z, Narasimhan G, Yao X, Zhang W. Mitigating multisource biases in graph neural networks via real counterfactual samples. In: 2023:638–647.
  • 23 Yazdani S, Saxena N, Wang Z, Wu Y, Zhang W. A Comprehensive Survey of Image and Video Generative AI: Recent Advances, Variants, and Applications. 2024.
  • 24 Wang Z, Wallace C, Bifet A, Yao X, Zhang W. FG2AN: Fairness-Aware Graph Generative Adversarial Networks. In: 2023:259–275.
  • 25 Zhang W, Wang Z, Kim J, et al. Individual Fairness under Uncertainty. In: 2023:3042–3049.
  • 26 Wang Z, Saxena N, Yu T, et al. Preventing Discriminatory Decision-making in Evolving Data Streams. In: 2023.
  • 27 Wang Z, Ulloa D, Yu T, Rangaswami R, Yap R, Zhang W. Individual Fairness with Group Constraints in Graph Neural Networks. In: , , , 2024.
  • 28 Wang Z, Chu Z, Blanco R, Chen Z, Chen SC, Zhang W. Advancing Graph Counterfactual Fairness through Fair Representation Learning. In: 2024.
  • 29 Wang Z, Qiu M, Chen M, Salem MB, Yao X, Zhang W. Toward fair graph neural networks via real counterfactual samples. Knowledge and Information Systems. 2024:1–25.
  • 30 Wang Z, Dzuong J, Yuan X, et al. Individual Fairness with Group Awareness under Uncertainty. In: 2024.
  • 31 Doan TV, Wang Z, Nguyen MN, Zhang W. Fairness in Large Language Models in three hours. In: 2024.
  • 32 Wang Z, Zhang W. Group Fairness with Individual and Censorship Constraints. In: 2024.
  • 33 Chinta SV, Wang Z, Yin Z, et al. FairAIED: Navigating Fairness, Bias, and Ethics in Educational AI Applications. In: 2024.
  • 34 Chinta SV, Wang Z, Zhang X, et al. AI-Driven Healthcare: A Survey on Ensuring Fairness and Mitigating Bias. 2024.
  • 35 Chu Z, Wang Z, Zhang W. Fairness in Large Language Models: A Taxonomic Survey. ACM SIGKDD Explorations Newsletter. 2024.
  • 36 Doan TV, Chu Z, Wang Z, Zhang W. Fairness Definitions in Language Models Explained. 2024.
  • 37 Chinta SV, Wang Z, Zhang X, et al. AI-Driven Healthcare: A Survey on Ensuring Fairness and Mitigating Bias. Journal of Healthcare Informatics. 2024.
  • 38 Saxena NA, Zhang W, Shahabi C. Unveiling and mitigating bias in ride-hailing pricing for equitable policy making. AI and Ethics. 2024:1–12.
  • 39 Zhang W, Weiss JC. Fairness with censorship and group constraints. Knowledge and Information Systems. 2023:1–24.
  • 40 Quy TL, Roy A, Iosifidis V, Zhang W, Ntoutsi E. A survey on datasets for fairness-aware machine learning. Data Mining and Knowledge Discovery. 2022.
  • 41 Zhang W, Pan S, Zhou S, Walsh T, Weiss JC. Fairness Amidst Non-IID Graph Data: Current Achievements and Future Directions. arXiv preprint arXiv:2202.07170. 2022.
  • 42 Zhang W, Hernandez-Boussard T, Weiss J. Censored fairness through awareness. In: . 37. 2023:14611–14619.
  • 43 Zhang W, Weiss JC. Longitudinal fairness with censorship. In: . 36. 2022:12235–12243.
  • 44 Zhang W, Weiss J. Fair Decision-making Under Uncertainty. In: 2021.
  • 45 Zhang W, Bifet A, Zhang X, Weiss JC, Nejdl W. FARF: A Fair and Adaptive Random Forests Classifier. In: 2021:245–256.
  • 46 Zhang W. Learning fairness and graph deep generation in dynamic environments. 2020.
  • 47 Zhang W, Zhao L. Online decision trees with fairness. arXiv preprint arXiv:2010.08146. 2020.
  • 48 Zhang W, others . Flexible and adaptive fairness-aware learning in non-stationary data streams. In: 2020:399–406.
  • 49 Zhang W, Bifet A. Feat: A fairness-enhancing and concept-adapting decision tree classifier. In: 2020:175–189.
  • 50 Zhang W, Tang X, Wang J. On fairness-aware learning for non-discriminative decision-making. In: 2019:1072–1079.
  • 51 Zhang W, Ntoutsi E. FAHT: an adaptive fairness-aware decision tree classifier. In: 2019:1480–1486.
  • 52 Saxena NA, Zhang W, Shahabi C. Missed Opportunities in Fair AI. In: 2023:961–964.
  • 53 Zhang W. Fairness with Censorship: Bridging the Gap between Fairness Research and Real-World Deployment. In: . 38. 2024:22685–22685.
  • 54 Zhang W. AI Fairness in Practice: Paradigm, Challenges, and Prospects. Ai Magazine. 2024.
  • 55 Zhang W, Tang J, Wang N. Using the machine learning approach to predict patient survival from high-dimensional survival data. In: 2016.
  • 56 Zhang W, Wang J. Content-bootstrapped collaborative filtering for medical article recommendations. In: 2018.
  • 57 Tang X, Zhang L, others . Using machine learning to automate mammogram images analysis. In: 2020:757–764.
  • 58 Zhang L, others . A comparison of different pattern recognition methods with entropy based feature reduction in early breast cancer classification. European Scientific Journal. 2014;3:303–312.
  • 59 Zhang M, Zhao X, others . Deep discriminative learning for autism spectrum disorder classification. In: 2020:435–443.
  • 60 Wang X, Zhang W, Jadhav A, Weiss J. Harmonic-Mean Cox Models: A Ruler for Equal Attention to Risk. In: 2021:171–183.
  • 61 Liu Z, Wang R, Zhang W. Improving the generalization of unsupervised feature learning by using data from different sources on gene expression data for cancer diagnosis. Medical & Biological Engineering & Computing. 2022:1–19.
  • 62 Wang J, Huang Z, others . Wearable sensor based human posture recognition. In: 2016.
  • 63 Zhang W, Wang J. A hybrid learning framework for imbalanced stream classification. In: 2017:480–487.
  • 64 Zhang W. Phd forum: Recognizing human posture from time-changing wearable sensor data streams. In: 2017.
  • 65 Zhang M, Zhang F, Zhang J, et al. Autoencoder for neuroimage. In: 2021:84–90.
  • 66 Liu Z, Wang R, Zhang W, Tang D. An unsupervised feature learning method for enhancing the generalization of cancer diagnosis. In: 2021:252–257.
  • 67 Zhang W, Wang J, Jin D, Oreopoulos L, Zhang Z. A deterministic self-organizing map approach and its application on satellite data based cloud type classification. In: 2018.
  • 68 Tang X, Qiu J, others . The internet of responsibilities-connecting human responsibilities using big data and blockchain. In: 2019.
  • 69 Tang X, others . A data-driven human responsibility management system. In: 2020.
  • 70 Tang X, Huang X, others . Cognitive Visual Commonsense Reasoning Using Dynamic Working Memory. In: 2021.
  • 71 Tang X, Zhang W, Yu Y, et al. Interpretable Visual Understanding with Cognitive Attention Network. In: 2021:555–568.
  • 72 Liu Z, Wang R, Japkowicz N, Tang D, Zhang W, Zhao J. Research on unsupervised feature learning for Android malware detection based on Restricted Boltzmann Machines. Future Generation Computer Systems. 2021;120:91–108.
  • 73 Liu Z, Wang R, Japkowicz N, Gomes HM, Peng B, Zhang W. SeGDroid: An Android malware detection method based on sensitive function call graph learning. Expert Systems with Applications. 2023:121125.
  • 74 Guyet T, Zhang W, Bifet A. Incremental Mining of Frequent Serial Episodes Considering Multiple Occurrences. In: 2022:460–472.
  • 75 Cai Y, Youngstrom D, Zhang W. Exploring Approaches for Teaching Cybersecurity and AI for K-12. In: 2023:1559–1564.
  • 76 Johnson KW, Torres Soto J, Glicksberg BS, et al. Artificial intelligence in cardiology. Journal of the American College of Cardiology. 2018;71(23):2668–2679.
  • 77 Krittanawong C, Zhang H, Wang Z, Aydar M, Kitai T. Artificial intelligence in precision cardiovascular medicine. Journal of the American College of Cardiology. 2017;69(21):2657–2664.
  • 78 Attia ZI, Kapa S, Lopez-Jimenez F, et al. Screening for cardiac contractile dysfunction using an artificial intelligence–enabled electrocardiogram. Nature medicine. 2019;25(1):70–74.
  • 79 Hannun AY, Rajpurkar P, Haghpanahi M, et al. Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nature medicine. 2019;25(1):65–69.
  • 80 Madani A, Arnaout R, Mofrad M, Arnaout R. Fast and accurate view classification of echocardiograms using deep learning. NPJ digital medicine. 2018;1(1):6.
  • 81 Ambale-Venkatesh B, Lima JA. Cardiac MRI: a central prognostic tool in myocardial fibrosis. Nature Reviews Cardiology. 2015;12(1):18–29.
  • 82 Krittanawong C, Johnson KW, Rosenson RS, et al. Deep learning for cardiovascular medicine: a practical primer. European heart journal. 2019;40(25):2058–2073.
  • 83 Farhud DD, Zokaei S. Ethical issues of artificial intelligence in medicine and healthcare. Iranian journal of public health. 2021;50(11):i.
  • 84 Fuster V, Kelly BB, Vedanthan R. Global cardiovascular health: urgent need for an intersectoral approach. Journal of the American College of Cardiology. 2011;58(12):1208–1210.
  • 85 Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nature medicine. 2019;25(1):44–56.
  • 86 Ting DSW, Cheung CYL, Lim G, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. Jama. 2017;318(22):2211–2223.
  • 87 Gulshan V, Peng L, Coram M, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. jama. 2016;316(22):2402–2410.
  • 88 De Fauw J, Ledsam JR, Romera-Paredes B, et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nature medicine. 2018;24(9):1342–1350.
  • 89 Balyen L, Peto T. Promising artificial intelligence-machine learning-deep learning algorithms in ophthalmology. The Asia-Pacific Journal of Ophthalmology. 2019;8(3):264–272.
  • 90 Kermany DS, Goldbaum M, Cai W, et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. cell. 2018;172(5):1122–1131.
  • 91 Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ digital medicine. 2018;1(1):39.
  • 92 Esteva A, Kuprel B, Novoa RA, et al. Dermatologist-level classification of skin cancer with deep neural networks. nature. 2017;542(7639):115–118.
  • 93 Han SS, Kim MS, Lim W, Park GH, Park I, Chang SE. Classification of the clinical images for benign and malignant cutaneous tumors using a deep learning algorithm. Journal of Investigative Dermatology. 2018;138(7):1529–1538.
  • 94 Tschandl P, Rinner C, Apalla Z, et al. Human–computer collaboration for skin cancer recognition. Nature Medicine. 2020;26(8):1229–1234.
  • 95 Gomolin A, Netchiporouk E, Gniadecki R, Litvinov IV. Artificial intelligence applications in dermatology: where do we stand?. Frontiers in medicine. 2020;7:100.
  • 96 Liu Y, Jain A, Eng C, et al. A deep learning system for differential diagnosis of skin diseases. Nature medicine. 2020;26(6):900–908.
  • 97 LeCun Y, Bengio Y, Hinton G. Deep learning. nature. 2015;521(7553):436–444.
  • 98 Dorsey ER, Bloem BR. The Parkinson pandemic—a call to action. JAMA neurology. 2018;75(1):9–10.
  • 99 Ramos-Murguialday A, Broetz D, Rea M, et al. Brain–machine interface in chronic stroke rehabilitation: a controlled study. Annals of neurology. 2013;74(1):100–108.
  • 100 Van Calster B, Wynants L, Timmerman D, Steyerberg EW, Collins GS. Predictive analytics in health care: how can we know it works?. Journal of the American Medical Informatics Association. 2019;26(12):1651–1654.
  • 101 Bzdok D, Meyer-Lindenberg A. Machine learning for precision psychiatry: opportunities and challenges. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging. 2018;3(3):223–230.
  • 102 Fornito A, Zalesky A, Bullmore E. Fundamentals of brain network analysis. London: Academic press, 2016.
  • 103 Burdick H, Lam C, Mataraso S, et al. Prediction of respiratory decompensation in Covid-19 patients using machine learning: The READY trial. Computers in biology and medicine. 2020;124:103949.
  • 104 Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts HJ. Artificial intelligence in radiology. Nature Reviews Cancer. 2018;18(8):500–510.
  • 105 Wang S, Summers RM. Machine learning and radiology. Medical image analysis. 2012;16(5):933–951.
  • 106 Yu KH, Beam AL, Kohane IS. Artificial intelligence in healthcare. Nature biomedical engineering. 2018;2(10):719–731.
  • 107 Bi WL, Hosny A, Schabath MB, et al. Artificial intelligence in cancer imaging: clinical challenges and applications. CA: a cancer journal for clinicians. 2019;69(2):127–157.
  • 108 McKinney SM, Sieniek M, Godbole V, et al. International evaluation of an AI system for breast cancer screening. Nature. 2020;577(7788):89–94.
  • 109 Wardi G, Carlile M, Holder A, Shashikumar S, Hayden SR, Nemati S. Predicting progression to septic shock in the emergency department using an externally generalizable machine-learning algorithm. Annals of emergency medicine. 2021;77(4):395–406.
  • 110 Brann F, Sterling NW, Frisch SO, Schrager JD, others . Sepsis Prediction at Emergency Department Triage Using Natural Language Processing: Retrospective Cohort Study. JMIR AI. 2024;3(1):e49784.
  • 111 Lee YTH, Di M, Schrager JD, Buckareff Z, Patzer RE, Yaffee AQ. The Use of a Self-triage Tool to Predict COVID-19 Cases and Hospitalizations in the State of Georgia. Western Journal of Emergency Medicine. 2022;23(4):532.
  • 112 Miner AS, Laranjo L, Kocaballi AB. Chatbots in the fight against the COVID-19 pandemic. NPJ digital medicine. 2020;3(1):65.
  • 113 Semigran HL, Linder JA, Gidengil C, Mehrotra A. Evaluation of symptom checkers for self diagnosis and triage: audit study. bmj. 2015;351.
  • 114 Dong E, Du H, Gardner L. An interactive web-based dashboard to track COVID-19 in real time. The Lancet infectious diseases. 2020;20(5):533–534.
  • 115 Sterling NW, Brann F, Patzer RE, et al. Prediction of emergency department resource requirements during triage: An application of current natural language processing techniques. Journal of the American College of Emergency Physicians Open. 2020;1(6):1676–1683.
  • 116 Chen IY, Pierson E, Rose S, Joshi S, Ferryman K, Ghassemi M. Ethical machine learning in healthcare. Annual review of biomedical data science. 2021;4:123–144.
  • 117 Chen I, Johansson FD, Sontag D. Why is my classifier discriminatory?. Advances in neural information processing systems. 2018;31.
  • 118 Ferrara E. Fairness and bias in artificial intelligence: A brief survey of sources, impacts, and mitigation strategies. Sci. 2023;6(1):3.
  • 119 Yang Y, Liu Y, Liu X, et al. Demographic Bias of Expert-Level Vision-Language Foundation Models in Medical Imaging. arXiv preprint arXiv:2402.14815. 2024.
  • 120 Schrager JD, Patzer RE, Kim JJ, et al. Racial and ethnic differences in diagnostic imaging utilization during adult emergency department visits in the United States, 2005 to 2014. Journal of the American College of Radiology. 2019;16(8):1036–1045.
  • 121 Zhang X, Kim J, Patzer RE, Pitts SR, Chokshi FH, Schrager JD. Advanced diagnostic imaging utilization during emergency department visits in the United States: A predictive modeling study for emergency department triage. PloS one. 2019;14(4):e0214905.
  • 122 Arif SA, Schlotfeldt J. Gaps in measuring and mitigating implicit bias in healthcare. Frontiers in Pharmacology. 2021;12:633565.
  • 123 Chen F, Wang L, Hong J, Jiang J, Zhou L. Unmasking bias in artificial intelligence: a systematic review of bias detection and mitigation strategies in electronic health record-based models. Journal of the American Medical Informatics Association. 2024:ocae060.
  • 124 Moghadasi N, Piran M, Baek S, et al. Systems Analysis of Bias and Risk in AI-Enabled Medical Diagnosis. In: 2023:1800–1807.
  • 125 Mittermaier M, Raza MM, Kvedar JC. Bias in AI-based models for medical applications: challenges and mitigation strategies. NPJ Digital Medicine. 2023;6(1):113.
  • 126 Chase AC. Ethics of AI: Perpetuating racial inequalities in healthcare delivery and patient outcomes. Voices in Bioethics. 2020;6.
  • 127 Nguyen-Truong CK, Rakha S, Eti DU, Angelesco L. Deconstructing racialized experiences in healthcare: what a missed opportunity for healing looks like and healthcare resources for children and their families. Asian/Pacific Island nursing journal. 2021;5(4):227.
  • 128 Tidjon LN, Khomh F. Never trust, always verify: a roadmap for trustworthy ai?. arXiv preprint arXiv:2206.11981. 2022.
  • 129 Vayena E, Blasimme A, Cohen IG. Machine learning in medicine: addressing ethical challenges. PLoS medicine. 2018;15(11):e1002689.
  • 130 Mittelstadt BD, Allo P, Taddeo M, Wachter S, Floridi L. The ethics of algorithms: Mapping the debate. Big Data & Society. 2016;3(2):2053951716679679.
  • 131 Kherbache A, Mertens E, Denier Y. Moral distress in medicine: an ethical analysis. Journal of health psychology. 2022;27(8):1971–1990.
  • 132 Kerasidou A. Trust me, I’ma researcher!: the role of trust in biomedical research. Medicine, Health Care and Philosophy. 2017;20(1):43–50.
  • 133 Blease C, Kaptchuk TJ, Bernstein MH, Mandl KD, Halamka JD, DesRoches CM. Artificial intelligence and the future of primary care: exploratory qualitative study of UK general practitioners’ views. Journal of medical Internet research. 2019;21(3):e12802.
  • 134 Larson HJ, Jarrett C, Eckersberger E, Smith DM, Paterson P. Understanding vaccine hesitancy around vaccines and vaccination from a global perspective: a systematic review of published literature, 2007–2012. Vaccine. 2014;32(19):2150–2159.
  • 135 Veinot TC, Mitchell H, Ancker JS. Good intentions are not enough: how informatics interventions can worsen inequality. Journal of the American Medical Informatics Association. 2018;25(8):1080–1088.
  • 136 Carter SM, Rogers W, Win KT, Frazer H, Richards B, Houssami N. The ethical, legal and social implications of using artificial intelligence systems in breast cancer care. The Breast. 2020;49:25–32.
  • 137 Char DS, Shah NH, Magnus D. Implementing machine learning in health care—addressing ethical challenges. The New England journal of medicine. 2018;378(11):981.
  • 138 Price WN, Cohen IG. Privacy in the age of medical big data. Nature medicine. 2019;25(1):37–43.
  • 139 Corbett-Davies S, Gaebler JD, Nilforoshan H, Shroff R, Goel S. The measure and mismeasure of fairness. The Journal of Machine Learning Research. 2023;24(1):14730–14846.
  • 140 Benjamin R. Race after technology: Abolitionist tools for the new Jim code. Hoboken, New Jersey: John Wiley & Sons, 2019.
  • 141 Hardt M, Cila N, Desmet P. Love. The forgotten dimension for just and democratic AI Futures. 2024.
  • 142 Howell MD, Corrado GS, DeSalvo KB. Three epochs of artificial intelligence in health care. Jama. 2024;331(3):242–244.
  • 143 Dwork C, Hardt M, Pitassi T, Reingold O, Zemel R. Fairness through awareness. In: 2012:214–226.
  • 144 Hardt M, Price E, Srebro N. Equality of opportunity in supervised learning. Advances in neural information processing systems. 2016;29.
  • 145 Jiang W, Pardos ZA. Towards equity and algorithmic fairness in student grade prediction. In: 2021:608–617.
  • 146 Terry R, Gatewood A, Elenwo C, et al. Disparities in preconception health indicators in US women: a cross-sectional analysis of the behavioral risk factor surveillance system 2019. Journal of Perinatal Medicine. 2024;52(2):192–201.
  • 147 Rees CA, Basnet S, Gentile A, et al. An analysis of clinical predictive values for radiographic pneumonia in children. BMJ global health. 2020;5(8):e002708.
  • 148 Basnet S, Gentile A, Gessner BD, et al. An analysis of clinical predictive values for radiographic pneumonia in children.. 2020.
  • 149 Kouznetsov R, Angelopoulos P, Moulinos S, Dimakos I, Gourzis P, Jelastopulu E. Epidemiological study of autism spectrum disorders in Greece for 2021: nationwide prevalence in 2–17-year-old children and regional disparities. Journal of Clinical Medicine. 2023;12(7):2510.
  • 150 Savage HR, Finch L, Body R, et al. A prospective diagnostic evaluation of accuracy of self-taken and healthcare worker-taken swabs for rapid COVID-19 testing. PloS one. 2022;17(6):e0270715.
  • 151 Neyigapula* BS. AI in Healthcare: Enhancing Diagnosis, Treatment, and Healthcare Systems for a Smarter Future in India. Advances in Bioengineering and Biomedical Science Research. 2023.
  • 152 Kumagai AK, Lypson ML. Beyond cultural competence: critical consciousness, social justice, and multicultural education. Academic medicine. 2009;84(6):782–787.
  • 153 Sequist TD, Adams A, Zhang F, Ross-Degnan D, Ayanian JZ. Effect of quality improvement on racial disparities in diabetes care. Archives of Internal Medicine. 2006;166(6):675–681.
  • 154 Van Ryn M, Burgess DJ, Dovidio JF, et al. The impact of racism on clinician cognition, behavior, and clinical decision making. Du Bois review: social science research on race. 2011;8(1):199–218.
  • 155 O’Neill J, Tabish H, Welch V, et al. Applying an equity lens to interventions: using PROGRESS ensures consideration of socially stratifying factors to illuminate inequities in health. Journal of clinical epidemiology. 2014;67(1):56–64.
  • 156 Solar O, Irwin A. A conceptual framework for action on the social determinants of health. tech. rep., WHO Document Production Services; Vancouver, Canada: 2010.
  • 157 Krieger N. Methods for the scientific study of discrimination and health: an ecosocial approach. American journal of public health. 2012;102(5):936–944.
  • 158 FitzGerald C, Hurst S. Implicit bias in healthcare professionals: a systematic review. BMC medical ethics. 2017;18:1–18.
  • 159 Green AR, Carney DR, Pallin DJ, et al. Implicit bias among physicians and its prediction of thrombolysis decisions for black and white patients. Journal of general internal medicine. 2007;22:1231–1238.
  • 160 Sabin DJA, Nosek DBA, Greenwald DAG, Rivara DFP. Physicians’ implicit and explicit attitudes about race by MD race, ethnicity, and gender. Journal of health care for the poor and underserved. 2009;20(3):896.
  • 161 Chin MH, Walters AE, Cook SC, Huang ES. Interventions to reduce racial and ethnic disparities in health care. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2655819/; 2007.
  • 162 Gallifant J, Kistler EA, Nakayama LF, et al. Disparity dashboards: an evaluation of the literature and framework for health equity improvement. The Lancet Digital Health. 2023;5(11):e831–e839.
  • 163 Kilbourne AM, Switzer G, Hyman K, Crowley-Matoka M, Fine MJ. Advancing health disparities research within the health care system: a conceptual framework. American journal of public health. 2006;96(12):2113–2121.
  • 164 Cook BL, McGuire TG, Zaslavsky AM. Measuring racial/ethnic disparities in health care: methods and practical issues. Health services research. 2012;47(3pt2):1232–1254.
  • 165 Grimshaw J, McAuley L, Bero L, et al. Systematic reviews of the effectiveness of quality improvement strategies and programmes. BMJ Quality & Safety. 2003;12(4):298–303.
  • 166 Betancourt JR, Green AR, Carrillo JE, Owusu Ananeh-Firempong I. Defining cultural competence: a practical framework for addressing racial/ethnic disparities in health and health care. Public health reports. 2003.
  • 167 Weech-Maldonado R, Dreachslin JL, Epané JP, Gail J, Gupta S, Wainio JA. Hospital cultural competency as a systematic organizational intervention: Key findings from the national center for healthcare leadership diversity demonstration project. Health care management review. 2018;43(1):30–41.
  • 168 Gianfrancesco MA, Tamang S, Yazdany J, Schmajuk G. Potential biases in machine learning algorithms using electronic health record data. JAMA internal medicine. 2018;178(11):1544–1547.
  • 169 Obermeyer Z, Emanuel EJ. Predicting the future—big data, machine learning, and clinical medicine. The New England journal of medicine. 2016;375(13):1216.
  • 170 Vyas DA, Eisenstein LG, Jones DS. Hidden in plain sight—reconsidering the use of race correction in clinical algorithms. https://dl.acm.org/doi/10.1145/3351095.3372873; 2020.
  • 171 Gebru T, Morgenstern J, Vecchione B, et al. Datasheets for datasets. Communications of the ACM. 2021;64(12):86–92.
  • 172 Kamiran F, Calders T. Data preprocessing techniques for classification without discrimination. Knowledge and information systems. 2012;33(1):1–33.
  • 173 Lum K, Johndrow J. A statistical framework for fair predictive algorithms. arXiv preprint arXiv:1610.08077. 2016.
  • 174 Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. Journal of artificial intelligence research. 2002;16:321–357.
  • 175 Calmon F, Wei D, Vinzamuri B, Natesan Ramamurthy K, Varshney KR. Optimized pre-processing for discrimination prevention. Advances in neural information processing systems. 2017;30.
  • 176 Zemel R, Wu Y, Swersky K, Pitassi T, Dwork C. Learning fair representations. In: 2013:325–333.
  • 177 Deldjoo Y, Noia TD, Merra FA. A survey on adversarial recommender systems: from attack/defense strategies to generative adversarial networks. ACM Computing Surveys (CSUR). 2021;54(2):1–38.
  • 178 Zafar MB, Valera I, Rogriguez MG, Gummadi KP. Fairness constraints: Mechanisms for fair classification. In: 2017:962–970.
  • 179 Elzayn H, Jabbari S, Jung C, et al. Fair algorithms for learning in allocation problems. In: 2019:170–179.
  • 180 Donini M, Oneto L, Ben-David S, Shawe-Taylor JS, Pontil M. Empirical risk minimization under fairness constraints. Advances in neural information processing systems. 2018;31.
  • 181 Pleiss G, Raghavan M, Wu F, Kleinberg J, Weinberger KQ. On fairness and calibration. Advances in neural information processing systems. 2017;30.
  • 182 Siontis KC, Zhang X, Eckard A, et al. Outcomes associated with apixaban use in patients with end-stage kidney disease and atrial fibrillation in the United States. Circulation. 2018;138(15):1519–1529.
  • 183 Fenton JJ, Xing G, Elmore JG, et al. Short-term outcomes of screening mammography using computer-aided detection: a population-based study of medicare enrollees. Annals of internal medicine. 2013;158(8):580–587.
  • 184 Henry C. Hospital closures: the sociospatial restructuring of labor and health care. Annals of the Association of American Geographers. 2015;105(5):1094–1110.
  • 185 Mittelstadt B. Principles alone cannot guarantee ethical AI. Nature machine intelligence. 2019;1(11):501–507.
  • 186 Goodman KW. Ethics in health informatics. Yearbook of medical informatics. 2020;29(01):026–031.
  • 187 Davenport T, Kalakota R. The potential for artificial intelligence in healthcare. Future healthcare journal. 2019;6(2):94.
  • 188 Luxton DD. Artificial intelligence in psychological practice: Current and future applications and implications.. Professional Psychology: Research and Practice. 2014;45(5):332.
  • 189 Hall MA, Orentlicher D, Bobinski MA, Bagley N, Cohen IG. Medical liability and treatment relationships. Aspen Publishing, 2018.
  • 190 Terry NP. Regulatory disruption and arbitrage in health-care data protection. Yale J. Health Pol’y L. & Ethics. 2017;17:143.
  • 191 Gerke S, Minssen T, Cohen G. Ethical and legal challenges of artificial intelligence-driven healthcare. In: , , Amsterdam, Netherlands: Elsevier, 2020:295–336.
  • 192 Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A. A survey on bias and fairness in machine learning. ACM computing surveys (CSUR). 2021;54(6):1–35.
  • 193 Ghassemi M, Naumann T, Schulam P, Beam AL, Chen IY, Ranganath R. Practical guidance on artificial intelligence for health-care data. The Lancet Digital Health. 2019;1(4):e157–e159.
  • 194 Beam AL, Kohane IS. Big data and machine learning in health care. Jama. 2018;319(13):1317–1318.
  • 195 Goodman B, Flaxman S. European Union regulations on algorithmic decision-making and a “right to explanation”. AI magazine. 2017;38(3):50–57.
  • 196 Ribeiro MT, Singh S, Guestrin C. " Why should i trust you?" Explaining the predictions of any classifier. In: 2016:1135–1144.
  • 197 Holzinger A, Biemann C, Pattichis CS, Kell DB. What do we need to build explainable AI systems for the medical domain?. arXiv preprint arXiv:1712.09923. 2017.
  • 198 Mitchell M, Wu S, Zaldivar A, et al. Model cards for model reporting. In: 2019:220–229.
  • 199 Lundberg SM, Lee SI. A unified approach to interpreting model predictions. Advances in neural information processing systems. 2017;30.