Statistical Learning in Medical Research with Decision Threshold and Accuracy Evaluation
Volume 19, Issue 4 (2021), pp. 634–657
Pub. online: 23 September 2021
Type: Data Science Reviews
Received
10 May 2021
10 May 2021
Accepted
18 August 2021
18 August 2021
Published
23 September 2021
23 September 2021
Abstract
Machine learning methods are increasingly applied for medical data analysis to reduce human efforts and improve our understanding of disease propagation. When the data is complicated and unstructured, shallow learning methods may not be suitable or feasible. Deep learning neural networks like multilayer perceptron (MLP) and convolutional neural network (CNN), have been incorporated in medical diagnosis and prognosis for better health care practice. For a binary outcome, these learning methods directly output predicted probabilities for patient’s health condition. Investigators still need to consider appropriate decision threshold to split the predicted probabilities into positive and negative regions. We review methods to select the cut-off values, including the relatively automatic methods based on optimization of the ROC curve criteria and also the utility-based methods with a net benefit curve. In particular, decision curve analysis (DCA) is now acknowledged in medical studies as a good complement to the ROC analysis for the purpose of decision making. In this paper, we provide the R code to illustrate how to perform the statistical learning methods, select decision threshold to yield the binary prediction and evaluate the accuracy of the resulting classification. This article will help medical decision makers to understand different classification methods and use them in real world scenario.
Supplementary material
Supplementary MaterialsSupplementary material online include: The review of different smoothers used in Generalized additive models, Installation details for R interface for Keras and Tensorflow, data and R code needed to reproduce the results.
References
Andrychowicz M, Denil M, Colmenarejo SG, Hoffman MW, Pfau D, Schaul T, et al. (2016). Learning to learn by gradient descent by gradient descent. CoRR, arXiv preprint: https://arxiv.org/abs/1606.04474.
Chollet F, Allaire J, et al. (2017). R interface to keras, https://github.com/rstudio/keras.
Kingma DP, Ba J (2014). Adam: A method for stochastic optimization. CoRR, arXiv preprint: https://arxiv.org/abs/1412.6980.
Reddi SJ, Kale S, Kumar S (2019). On the convergence of adam and beyond. CoRR, arXiv preprint: https://arxiv.org/abs/1904.09237.
Ruder S (2016). An overview of gradient descent optimization algorithms. CoRR, arXiv preprint: https://arxiv.org/abs/1609.04747.
Zeiler MD (2012). ADADELTA: An adaptive learning rate method. CoRR, arXiv preprint: https://arxiv.org/abs/1212.5701.