Mohd Khalid, Nur Hidayah and Ismail, Amelia Ritahani and Abdul Aziz, Normaziah (2022) Interpretation of machine learning model using medical record visual analytics. In: 8th International Conference on Computational Science and Technology ICCST 2021, 28-29 August 2021, Labuan, Malaysia.
PDF
- Published Version
Restricted to Registered users only Download (529kB) | Request a copy |
||
|
PDF (SCOPUS)
- Supplemental Material
Download (512kB) | Preview |
Abstract
The state of the art of medical application that being implemented are mostly based on common machine learning model. Nevertheless, one of the draw- backs of the practice of medical diagnosis is the lack of explanation on the proposed solution, which is also known as a black box, without knowing the internal decision process between the input and output. It will lead to untrustworthiness and difficult to understand by the medical expert. They are questioning how the complexity of machine learning methods decide on the output without clear and understandable explanations. Moreover, in machine learning field the characteristic of a black box model may lead to biased data analysis and incorrect output decisions. There is work that uses visual analytics techniques to interpret the machine learning output to ease the understanding of medical experts. However, the functionality of existed and combined visual analytics techniques is not sufficient to visualized and interpreted the output of machine learning operation. Other visual analytic techniques faced the same problem, unreliability to produce strong reason on the output when working with complex machine learning models. This paper analyzed several visual analytics approaches instantiated in machine learning algorithm for medical record analytics. The motivation of this paper is to allow medical experts to understand the interpreta- tion of a black box machine learning model in predicting medical outcome. This paper studied on the effectiveness of visual analytics techniques to identify the appropriate technique to be instantiated to the machine learning algorithm to further elaborate the results obtained by demonstrating transparency, interpretability and explainability of the machine learning algorithm. The visual analytics that are been studied are Local Interpretable Model-agnostic Explanations (LIME) and Shapley Additive exPlana- tions (SHAP). Based on the comparison of LIME and SHAP methods, this paper found that SHAP has consistent interpretability as compared to LIME.
Item Type: | Conference or Workshop Item (Plenary Papers) |
---|---|
Uncontrolled Keywords: | Machine learning · Interpretability · Visual analytics |
Subjects: | Q Science > QA Mathematics > QA75 Electronic computers. Computer science |
Kulliyyahs/Centres/Divisions/Institutes (Can select more than one option. Press CONTROL button): | Kulliyyah of Information and Communication Technology > Department of Computer Science Kulliyyah of Information and Communication Technology > Department of Computer Science |
Depositing User: | Amelia Ritahani Ismail |
Date Deposited: | 25 Jul 2022 08:40 |
Last Modified: | 25 Jul 2022 08:40 |
URI: | http://irep.iium.edu.my/id/eprint/98904 |
Actions (login required)
View Item |