DISSERTATION

Kernel-based expectile regression

Muhammad Shoaib Farooq

Year: 2017 University:   OPUS Publication Server of the University of Stuttgart (University of Stuttgart)   Publisher: University of Stuttgart

Abstract

Conditional expectiles are becoming an increasingly important tool in finance as well as in other areas of application such as demography when the goal is to explore the conditional distribution beyond the conditional mean. In this thesis, we consider a support vector machine (SVM) type approach with the asymmetric least squares loss for estimating conditional expectiles. Firstly, we establish learning rates for this approach that are minimax optimal modulo a logarithmic factor if Gaussian RBF kernels are used and the desired expectile is smooth in a Besov sense. It turns out that our learning rates, as a special case, improve the best known rates for kernel-based least squares regression in aforementioned scenario. As key ingredients of our statistical analysis, we establish a general calibration inequality for the asymmetric least squares loss, a corresponding variance bound as well as an improved entropy number bound for Gaussian RBF kernels. Furthermore, we establish optimal learning rates in the case of a generic kernel under the assumption that the target function is in a real interpolation space. Secondly, we complement the theoretical results of our SVM approach with the empirical findings. For this purpose we use a sequential minimal optimization method and design an SVM-like solver for expectile regression considering Gaussian RBF kernels. We conduct various experiments in order to investigate the behavior of the designed solver with respect to different combinations of initialization strategies, working set selection strategies, stopping criteria and number of nearest neighbors, and then look for the best combination of them. We further compare the results of our solver to the recent R-package ER-Boost and find that our solver exhibits a better test performance. In terms of training time, our solver is found to be more sensitive to the training set size and less sensitive to the dimensions of the data set, whereas, ER-Boost behaves the other way around. In addition, our solver is found to be faster than a similarly implemented solver for the quantile regression. Finally, we show the convergence of our designed solver.

Keywords:
Statistics Regression Mathematics Kernel (algebra) Econometrics

Metrics

1
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Gaussian Processes and Bayesian Inference
Physical Sciences →  Computer Science →  Artificial Intelligence
Neural Networks and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Learning rates for kernel-based expectile regression

Muhammad Shoaib FarooqIngo Steinwart

Journal:   Machine Learning Year: 2018 Vol: 108 (2)Pages: 203-227
JOURNAL ARTICLE

KLERC: kernel Lagrangian expectile regression calculator

Songfeng Zheng

Journal:   Computational Statistics Year: 2020 Vol: 36 (1)Pages: 283-311
JOURNAL ARTICLE

Flexible Expectile Regression in Reproducing Kernel Hilbert Spaces

Yi YangTeng ZhangHui Zou

Journal:   Technometrics Year: 2017 Vol: 60 (1)Pages: 26-35
© 2026 ScienceGate Book Chapters — All rights reserved.