Variance function estimation with LS-SVM for replicated data

  • Shim, Joo-Yong (Department of Applied Statistics, Catholic University of Daegu) ;
  • Park, Hye-Jung (Division of General Education and Teacher's Certification, Daegu University) ;
  • Seok, Kyung-Ha (Department of Data Science, Institute of Statistical Information, Inje University)
  • Published : 2009.09.30

Abstract

In this paper we propose a variance function estimation method for replicated data based on averages of squared residuals obtained from estimated mean function by the least squares support vector machine. Newton-Raphson method is used to obtain associated parameter vector for the variance function estimation. Furthermore, the cross validation functions are introduced to select the hyper-parameters which affect the performance of the proposed estimation method. Experimental results are then presented which illustrate the performance of the proposed procedure.

Keywords

References

  1. Anderson, T. G. and Lund, J. (1997). Estimating continuous-time stochastic volatility models of short-term Interest Rate. Journal of Econometrics, 77, 343-377. https://doi.org/10.1016/S0304-4076(96)01819-2
  2. Craven, P. and Wahba, G. (1979). Smoothing noisy data with spline functions: Estimating the correct degree of smoothing by the method of generalized cross-validation. Numerical Mathematics, 31, 377-403. https://doi.org/10.1007/BF01404567
  3. Green, P. J. and Silverman, B. W. (1995). Nonparametric regression and generalized linear models, Chapman & Hall, London.
  4. Kimeldorf, G. S. and Wahba, G. (1971). Some results on Tchebycheffian spline functions. Journal of Mathematical Analysis and its Applications, 33, 82-95. https://doi.org/10.1016/0022-247X(71)90184-3
  5. Liu, A., Tong, T. and Wang, Y. (2007). Smoothing spline estimation of variance functions. Journal of Computational and Graphical Statistics, 16, 312-329. https://doi.org/10.1198/106186007X204528
  6. Mercer, J. (1909). Functions of positive and negative type and their connection with theory of integral equations. Philosophical Transactions of Royal Society, A: 415-446.
  7. Shim, J. and Seok, K. H. (2008). Kernel ridge regression with randomly right censored data. Journal of Korean Data & Information Science Society, 19, 1353 -1360.
  8. Shim, J., Kim, T. Y., Lee, S. Y. and Hwang, C. (2009). Credibility estimation via Kernel mixed effects model. Journal of Korean Data & Information Science Society, 20, 445-452.
  9. Suykens, J. A. K. and Vanderwalle, J. (1999). Least square support vector machine classifier. Neural Processing Letters, 9, 293-300. https://doi.org/10.1023/A:1018628609742
  10. Suykens, J. A. K., De Brabanter, J., Lukas, L. and Vandewalle, J. (2002). Weighted least squares support vector machines: Robustness and sparse approximation. Neurocomputing, Special issue on fundamental and information processing aspects of neurocomputing, 48, 85-105.
  11. Suyken, J. A. K., Lukas, L. and Vandewalle, J. (2000). Sparse approximation using least squares support vector machines. IEEE International Symposium on Circuits and Systems (ISCAS 2000), 757-760, Geneva, Switzerland.
  12. Vapnik, V. N. (1998). Statistical learning theory, John Wiley & Sons.
  13. Wei, Y., Pere, A., Koenker, R. and He, X. (2006). Quantile regression for reference growth charts. Statistics in Medicine, 25, 1369-1382. https://doi.org/10.1002/sim.2271
  14. Xiang, D. and Wahba, G. (1996). A generalized approximate cross validation for smoothing splines with non-gaussian data. Statistian Sinica, 6, 675-692.