Date of Award
4-17-2019
Document Type
Masters Project
Abstract
This paper explores various techniques to estimate a confidence interval on accuracy for machine learning algorithms. Confidence intervals on accuracy may be used to rank machine learning algorithms. We investigate bootstrapping, leave one out cross validation, and conformal prediction. These techniques are applied to the following machine learning algorithms: support vector machines, bagging AdaBoost, and random forests. Confidence intervals are produced on a total of nine datasets, three real and six simulated. We found in general not any technique was particular successful at always capturing the accuracy. However leave one out cross validation had the most consistency amongst all techniques for all datasets.
Recommended Citation
Zhang, Jesse, "Estimating confidence intervals on accuracy in classification in machine learning" (2019). Mathematics and Statistics . 41.
https://ualaska.researchcommons.org/uaf_grad_math_stats/41
Handle
http://hdl.handle.net/11122/10958