TY - GEN
T1 - Fast cross-validation
AU - Liu, Yong
AU - Lin, Hailun
AU - Ding, Lizhong
AU - Wang, Weiping
AU - Liao, Shizhong
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledgements: This work is supported in part by the National Key Research and Development Program of China (2016YFB1000604), the National Natural Science Foundation of China (No.6173396, No.61673293, No.61602467) and the Excellent Talent Introduction of Institute of Information Engineering of CAS (Y7Z0111107).
PY - 2018/7/5
Y1 - 2018/7/5
N2 - Cross-validation (CV) is the most widely adopted approach for selecting the optimal model. However, the computation of CV has high complexity due to multiple times of learner training, making it disabled for large scale model selection. In this paper, we present an approximate approach to CV based on the theoretical notion of Bouligand influence function (BIF) and the Nyström method for kernel methods. We first establish the relationship between the theoretical notion of BIF and CV, and propose a method to approximate the CV via the Taylor expansion of BIF. Then, we provide a novel computing method to calculate the BIF for general distribution, and evaluate BIF for sample distribution. Finally, we use the Nyström method to accelerate the computation of the BIF matrix for giving the finally approximate CV criterion. The proposed approximate CV requires training only once and is suitable for a wide variety of kernel methods. Experimental results on lots of datasets show that our approximate CV has no statistical discrepancy with the original CV, but can significantly improve the efficiency.
AB - Cross-validation (CV) is the most widely adopted approach for selecting the optimal model. However, the computation of CV has high complexity due to multiple times of learner training, making it disabled for large scale model selection. In this paper, we present an approximate approach to CV based on the theoretical notion of Bouligand influence function (BIF) and the Nyström method for kernel methods. We first establish the relationship between the theoretical notion of BIF and CV, and propose a method to approximate the CV via the Taylor expansion of BIF. Then, we provide a novel computing method to calculate the BIF for general distribution, and evaluate BIF for sample distribution. Finally, we use the Nyström method to accelerate the computation of the BIF matrix for giving the finally approximate CV criterion. The proposed approximate CV requires training only once and is suitable for a wide variety of kernel methods. Experimental results on lots of datasets show that our approximate CV has no statistical discrepancy with the original CV, but can significantly improve the efficiency.
UR - http://hdl.handle.net/10754/665273
UR - https://www.ijcai.org/proceedings/2018/346
UR - http://www.scopus.com/inward/record.url?scp=85055724794&partnerID=8YFLogxK
U2 - 10.24963/ijcai.2018/346
DO - 10.24963/ijcai.2018/346
M3 - Conference contribution
SN - 9780999241127
SP - 2497
EP - 2503
BT - Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
PB - International Joint Conferences on Artificial Intelligence Organization
ER -