Explainable, trustworthy, and ethical machine learning for healthcare: A survey
With the advent of machine learning (ML) and deep learning (DL) empowered applications for critical applications like healthcare, the questions about liability, trust, and interpretability of their outputs are raising. The black-box nature of various DL models is a roadblock to clinical utilization. Therefore, to gain the trust of clinicians and patients, we need to provide explanations about the decisions of models. With the promise of enhancing the trust and transparency of black-box models, researchers are in the phase of maturing the field of eXplainable ML (XML). In this paper, we provided a comprehensive review of explainable and interpretable ML techniques for various healthcare applications. Along with highlighting security, safety, and robustness challenges that hinder the trustworthiness of ML, we also discussed the ethical issues arising because of the use of ML/DL for healthcare. We also describe how explainable and trustworthy ML can resolve all these ethical problems. Finally, we elaborate on the limitations of existing approaches and highlight various open research problems that require further development.
Other Information
Published in: Computers in Biology and Medicine
License: http://creativecommons.org/licenses/by/4.0/
See article on publisher's website: https://dx.doi.org/10.1016/j.compbiomed.2022.106043
Funding
Open Access funding provided by the Qatar National Library.
History
Language
- English
Publisher
ElsevierPublication Year
- 2022
License statement
This Item is licensed under the Creative Commons Attribution 4.0 International License.Institution affiliated with
- Qatar University
- College of Engineering - QU
- Hamad Bin Khalifa University
- College of Islamic Studies - HBKU
- Research Center for Islamic Legislation and Ethics - CIS
- College of Science and Engineering - HBKU