Incremental learning for ν-Support Vector Regression

Bin Gu, Victor S. Sheng, Zhijie Wang, Derek Ho, Said Osman, Shuo Li

Research output: Contribution to journalArticlepeer-review

443 Scopus citations

Abstract

The ν-Support Vector Regression (ν-SVR) is an effective regression learning algorithm, which has the advantage of using a parameter ν on controlling the number of support vectors and adjusting the width of the tube automatically. However, compared to ν-Support Vector Classification (ν-SVC) (Schölkopf etal., 2000), ν-SVR introduces an additional linear term into its objective function. Thus, directly applying the accurate on-line ν-SVC algorithm (AONSVM) to ν-SVR will not generate an effective initial solution. It is the main challenge to design an incremental ν-SVR learning algorithm. To overcome this challenge, we propose a special procedure called initial adjustments in this paper. This procedure adjusts the weights of ν-SVC based on the Karush-Kuhn-Tucker (KKT) conditions to prepare an initial solution for the incremental learning. Combining the initial adjustments with the two steps of AONSVM produces an exact and effective incremental ν-SVR learning algorithm (INSVR). Theoretical analysis has proven the existence of the three key inverse matrices, which are the cornerstones of the three steps of INSVR (including the initial adjustments), respectively. The experiments on benchmark datasets demonstrate that INSVR can avoid the infeasible updating paths as far as possible, and successfully converges to the optimal solution. The results also show that INSVR is faster than batch ν-SVR algorithms with both cold and warm starts.

Original languageEnglish
Pages (from-to)140-150
Number of pages11
JournalNeural Networks
Volume67
DOIs
StatePublished - Jul 1 2015

Keywords

  • Incremental learning
  • Online learning
  • Support vector machine
  • ν-Support Vector Regression

Fingerprint

Dive into the research topics of 'Incremental learning for ν-Support Vector Regression'. Together they form a unique fingerprint.

Cite this