+ All Categories
Home > Documents > Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial...

Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial...

Date post: 31-Dec-2020
Category:
Upload: others
View: 6 times
Download: 0 times
Share this document with a friend
8
Research Article Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs Yongming Cai, 1 Lei Song, 1 Tingwei Wang, 1 and Qing Chang 2 1 School of Management, University of Jinan, Jinan 250002, China 2 School of Management, Inner Mongolia University of Technology, Huhhot 010050, China Correspondence should be addressed to Yongming Cai; [email protected] Received 21 February 2014; Accepted 2 April 2014; Published 24 April 2014 Academic Editor: Wei Chen Copyright © 2014 Yongming Cai et al. is is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Support vector machines (SVMs) are a promising alternative to traditional regression estimation approaches. But, when dealing with massive-scale data set, there exist many problems, such as the long training time and excessive demand of memory space. So, the SVMs algorithm is not suitable to deal with financial time series data. In order to solve these problems, directed-weighted chunking SVMs algorithm is proposed. In this algorithm, the whole training data set is split into several chunks, and then the support vectors are obtained on each subset. Furthermore, the weighted support vector regressions are calculated to obtain the forecast model on the new working data set. Our directed-weighted chunking algorithm provides a new method of support vectors decomposing and combining according to the importance of chunks, which can improve the operation speed without reducing prediction accuracy. Finally, IBM stock daily close prices data are used to verify the validity of the proposed algorithm. 1. Introduction Financial time series forecasting is an important aspect of financial decisions. Financial practitioners and academic researchers have proposed a lot of methods and techniques to improve the accuracy of predictions. Because it is inher- ently noisy, nonstationary, and deterministically chaotic [1], financial time series forecasting is regarded as one of the most challenging applications of modern time series forecasting. Time series analysis of the early years study is focused on the regression model, such as autoregression model like AR and ARMA and volatility model, like ARCH and GARCH. In recent years, studies focused on application of artificial intelligence algorithms (AI), such as artificial neural networks (ANN) [24], reasoning neural networks (RNN) [5], genetic algorithms (GA) [6], particle swarm optimization (PSO) [7, 8], and support vector machines (SVMs) [9, 10]. Among these artificial intelligence algorithms, SVMs are an elegant tool for solving pattern recognition and regression problems. According to the research of Vapnik [11], SVMs implement the structural risk minimization principle which seeks to minimize an upper bound of the generalization error rather than minimize the training error. e regression model of SVMs, called support vector regression (SVR), has also been receiving increasing attention to solve linear or nonlinear estimation problems. For instance, Tay and Cao [9] studied the five real future contracts in Chicago Mercantile Market, Cao and Tay [12] studied the S&P 500 daily price index, and Kim [13] studied the daily Korea composite stock price index (KOSPI). Based on the criteria of normalized mean square error (NMSE), mean absolute error (MAE), directional symmetry (DS), and weighted directional symmetry (WDS), the above researches indicate that the performance of SVMs is better than ARMA, GARCH (ARCH), and ANN. According to the statistical learning theory (SLT), support vector machine regression is a convex quadratic program- ming (QP) optimization with linear constraint. However, there is an obvious disadvantage of SVMs that the training time scales somewhere between quadratic and cubic depend on the number of training samples. In order to deal with mas- sive data set and improve the training speed, many improved Hindawi Publishing Corporation Mathematical Problems in Engineering Volume 2014, Article ID 170424, 7 pages http://dx.doi.org/10.1155/2014/170424
Transcript
Page 1: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

Research ArticleFinancial Time Series Forecasting UsingDirected-Weighted Chunking SVMs

Yongming Cai1 Lei Song1 Tingwei Wang1 and Qing Chang2

1 School of Management University of Jinan Jinan 250002 China2 School of Management Inner Mongolia University of Technology Huhhot 010050 China

Correspondence should be addressed to Yongming Cai cym2001099163com

Received 21 February 2014 Accepted 2 April 2014 Published 24 April 2014

Academic Editor Wei Chen

Copyright copy 2014 Yongming Cai et al This is an open access article distributed under the Creative Commons Attribution Licensewhich permits unrestricted use distribution and reproduction in any medium provided the original work is properly cited

Support vector machines (SVMs) are a promising alternative to traditional regression estimation approaches But when dealingwith massive-scale data set there exist many problems such as the long training time and excessive demand of memory spaceSo the SVMs algorithm is not suitable to deal with financial time series data In order to solve these problems directed-weightedchunking SVMs algorithm is proposed In this algorithm the whole training data set is split into several chunks and then thesupport vectors are obtained on each subset Furthermore the weighted support vector regressions are calculated to obtain theforecast model on the new working data set Our directed-weighted chunking algorithm provides a newmethod of support vectorsdecomposing and combining according to the importance of chunks which can improve the operation speed without reducingprediction accuracy Finally IBM stock daily close prices data are used to verify the validity of the proposed algorithm

1 Introduction

Financial time series forecasting is an important aspect offinancial decisions Financial practitioners and academicresearchers have proposed a lot of methods and techniquesto improve the accuracy of predictions Because it is inher-ently noisy nonstationary and deterministically chaotic [1]financial time series forecasting is regarded as one of themostchallenging applications of modern time series forecasting

Time series analysis of the early years study is focusedon the regression model such as autoregression modellike AR and ARMA and volatility model like ARCH andGARCH In recent years studies focused on application ofartificial intelligence algorithms (AI) such as artificial neuralnetworks (ANN) [2ndash4] reasoning neural networks (RNN)[5] genetic algorithms (GA) [6] particle swarm optimization(PSO) [7 8] and support vector machines (SVMs) [9 10]

Among these artificial intelligence algorithms SVMs arean elegant tool for solving pattern recognition and regressionproblems According to the research of Vapnik [11] SVMsimplement the structural risk minimization principle which

seeks to minimize an upper bound of the generalizationerror rather than minimize the training error The regressionmodel of SVMs called support vector regression (SVR)has also been receiving increasing attention to solve linearor nonlinear estimation problems For instance Tay andCao [9] studied the five real future contracts in ChicagoMercantile Market Cao and Tay [12] studied the SampP 500daily price index and Kim [13] studied the daily Koreacomposite stock price index (KOSPI) Based on the criteriaof normalized mean square error (NMSE) mean absoluteerror (MAE) directional symmetry (DS) and weighteddirectional symmetry (WDS) the above researches indicatethat the performance of SVMs is better thanARMAGARCH(ARCH) and ANN

According to the statistical learning theory (SLT) supportvector machine regression is a convex quadratic program-ming (QP) optimization with linear constraint Howeverthere is an obvious disadvantage of SVMs that the trainingtime scales somewhere between quadratic and cubic dependon the number of training samples In order to deal withmas-sive data set and improve the training speed many improved

Hindawi Publishing CorporationMathematical Problems in EngineeringVolume 2014 Article ID 170424 7 pageshttpdxdoiorg1011552014170424

2 Mathematical Problems in Engineering

support vector machine methods are proposed One wayis to combine SVMs with some other methods such asactive learning [14 15] multitask learning [16 17] multiviewlearning [18] and semisupervised learning [19 20] Anotherway is to develop some optimization techniques of trainingalgorithms in SVMs such as sequential updating methodslike kernel-Adatron algorithm [21] successive over relaxationalgorithm [22] working set methods like chunking SVMs[23 24] reduced support vector machine (RSVM) [25] andsequential minimal optimization algorithm (SMO) [26]

Chunking SVMs reduced support vector machine(RSVM) and SVMs with sequential minimal optimization(SMO) are outstanding methods in dealing with massivedata set For example Lee and Mangasarian [25] designed areduced support vector machine algorithm (RSVM) whichcan greatly reduce the size of the quadratic program to besolved by reducing the data set volume so the memory usageis much smaller than that of a conventional SVM using theentire data set Osuna et al [27] designed a decompositionalgorithm that is guaranteed to solve the QP problem andthat does not make assumptions on the expected number ofsupport vectors Platt [26] put forward a sequential minimaloptimization algorithm (SMO) which breaks the large QPproblem into a series of smallest possible QP problems whichare analytically solvable in order to speed up the trainingtime Tay and Cao [28] proposed to combine support vectormachines (SVMs) with self-organizing feature map (SOM)for financial time series forecasting where SOM is used asa clustering algorithm to partition the whole input spaceinto several disjoint regions Tay and Cao [29] also putforward C-ascending support vector machines to amend the120576-insensitive errors in the nonstationary financial time series

Most of the improved support vector machine methodsdo well in memory requirement and CPU time but theforecast accuracy has declinedmore or less For financial timeseries prediction the typical massive-scale data we shouldpay more attention to the prediction accuracy while reducingthe computational complexity In this paper we proposeda directed-weighted chunking SVMs algorithm which canimprove the operation speed without reducing predictionaccuracy

This paper consists of five sections Section 1 introducesthe basic algorithm of SVM Section 2 proposes a directed-weighted chunking SVMs algorithm Section 3 designs aseries of experiments and empirical results are summarizedand discussed Section 4 presents the conclusions and limita-tions of this study

2 SVMs Regression Theory

In this section we will briefly introduce the support vectorsregression (SVR) theory Suppose there are a given set of datapoints 119866 = (119909

119894 119889119894)119897

119894(119909119894is the input vector 119889

119894is the desired

value) SVMs approximate the function in the following form

119910 =

119897

sum

119894=1

119908119894120601119894 (119909) + 119887 (1)

where 120601(119909)119897119894=1

are the features of inputs and 119908119894119897

119894=1 119887 are

coefficientsAccording to the structural risk minimization principle

the support vectors regression problem can be expressed as

Minimize 119877 (119908 120585(lowast)) =

1

2

119908119879119908 + 119862

lowast

119897

sum

119894=1

(120585119894+ 120585lowast

119894)

Subject to 119908120601 (119909119894) + 119887 minus 119889

119894le 120576 + 120585

119894

119889119894minus 119908120601 (119909

119894) minus 119887 le 120576 + 120585

lowast

119894

120585119894 120585lowast

119894ge 0

119894 = 1 2 119897

(2)

where 119909119894is mapped to a higher dimensional space by the

function 120601 120585119894and 120585

119894

lowast are slack variables (120585119894is the upper

training error 120585lowast119894is the lower) which are subject to the 120576-

insensitive tube |119889119894minus(119908120601(119909

119894)+119887)| le 120576 The parameters which

control the regression quality are the cost of error119862 thewidthof the tube 120576 and the mapping function 120601

Thus (1) becomes the following explicit form

119891 (119909 119886119894 119886lowast

119894) =

119897

sum

119894=1

(119886119894minus 119886lowast

119894)119870 (119909 119909

119894) + 119887 (3)

Then we can obtain the following form by maximizingthe dual form of function (3)

120601 (119886119894 119886lowast

119894) =

119897

sum

119894=1

119889119894(119886119894minus 119886lowast

119894) minus 120576

119897

sum

119894=1

(119886119894+ 119886lowast

119894)

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

(4)

with the following constraints

119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862 119894 = 1 2 119897

0 le 119886lowast

119894le 119862 119894 = 1 2 119897

(5)

where the 119886119894 119886lowast119894are the Lagrange multipliers introduced

and 119870(119909 119909119894) is named the kernel function The value is

equal to the inner product of two vectors 119909119894and 119909

119895in the

feature spaces 120601(119909119894) and 120601(119909

119895) There are many choices of the

kernel function common examples are the polynomial kernel119870(119909119894 119909119895) = (119909

119894119909119895+ 1)119889 and Gaussian kernel 119870(119909

119894 119909119895) =

exp(minus1205752(119909119894minus 119909119895)2)

Training SVMs is equivalent to optimizing the Lagrangemultipliers 119886

119894 119886lowast119894with constraints based on (4) Good fitting

function can be obtained by choosing appropriate functionalspace

There are many different researches using the straight-forward approaches to construct and implement SVMs

Mathematical Problems in Engineering 3

for financial time series analysis (see eg [9 13 30])other methods like chunking SVMs reduced support vec-tor machine (RSVM) and SVMs with sequential minimaloptimization (SMO) are used to deal with massive-scaledata sets Among these methods SVM chunking providesan alternative method to running a typical SVM on a dataset by breaking up the training data and running the SVMon smaller chunks of data In previous literature manydecomposition and combination methods are proposed

We would like to mention that for financial time seriesprediction problems the data set from different periodswill have different effects on current forecasts and thechange of direction of past stock price will also affect thecurrent forecasts Therefore we proposed directed-weightedchunking SVMs which can improve the operation speedwithout reducing prediction accuracy

3 The Directed-Weighted Chunking SVMs

31 Chunking Model in Support Vector Regression TrainingSVMs is equivalent to solving a linearly constrained QPproblem Training SVMs depends on QP optimization tech-niques Standard QP techniques cannot be directly applied toSVMsproblemswithmassive-scale data set In training stagesof directed-weighted chunking SVMs the whole trainingset is decomposed into several chunks and support vectorsare calculated respectively in their working subset In theprediction stage all these support vectors are combinedinto new working data set to get the model in accordancewith their importance as illustrated in Figure 1 and theprogress of directed-weighted chunking SVMs algorithm canbe described as follows

Step 1 Decompose the whole training set 119866 into 119898 subsets1198661 1198662 119866

119898 and 119866 = 119866

1cup 1198662 cup sdot sdot sdot cup119866

119898 119866119894cap 119866119895= Φ

Step 2 Calculate the support vector regression SVG119894for each

subset 119866119894

Step 3 Calculate the weight and direction of each subset

Step 4 Combine all support vectors 119866SVM119894 for each subsetinto new working data set 119866SVM

Step 5 Calculate the weighted support vector regression onthe new working data set 119866SVM and get the model

32 Directed-Weighted Chunking SVMs In the time seriesforecasting such as stock market the effect of past stockprices on the future stock prices will be different Usuallythe more recent the period of time the greater the weightcoefficient According to a certain time interval the total timeseries of stock priceswill be divided into different chunks andsupport vectors in each chunk are calculated respectivelyFurthermore the weighted support vector regressions arecalculated to obtain the forecast model

The stock market is one of the complex systems [31 32]We can treat chunks as the nodes and relationship betweenchunks as edges Thus these nodes and edges will be acomplex networkThe entire time series of stock price can beregarded as a directed-weighted network with a large numberof nodes and edges The mutual influence between chunkshas its direction For example in chunks G1 and G2 if thereexists a nonzero correlation coefficient from chunk G1 to G2we will draw a directed edge between G1 and G2 Becausethe strength of mutual influence between chunks is differentthe edge weight which is called correlation intensity isalso different So simply chunking SVMs cannot reflect theinfluence of respective chunk on final model Here we willintroduce directed-weighted chunking algorithm into SVMs

In traditional SVM regression optimization problem theparameter 119862lowast is a constant In order to show the differentinfluence on the prediction results we introduce a function119876() and modify the SVMs regression function as follows

119877 (119908 120585(lowast)) =

1

2

119908119879119908 + 119862

lowast

119897

sum

119894=1

119876 (119866119894 Δ119905) (120585

119894+ 120585lowast

119894) (6)

In (6) 119876(119866119894 Δ119905) can be understood as the weight of

each chunk and the values of 119876(119866119894 Δ119905) can be positive or

negative The positive or negative values of 119876(119866119894 Δ119905) will

change parameter 119862lowast in SVMs which is very similar tothat of the positive or negative information impact on thefuture stock price According to the definition of networkcorrelation coefficient introduced by Bonanno et al [33] wecan define the correlation intensity 119876(119866

119894 Δ119905) (influence of

chunk 119866119894by the time interval Δ119905) as follows

119876 (119866119894 Δ119905) =

⟨119877 (119866119894 119905) 119877 (119866

119894 119905 minus Δ119905)⟩ minus ⟨119877 (119866

119894 119905)⟩ ⟨119877 (119866

119894 119905 minus Δ119905)⟩

radic⟨[119877 (119866119894 119905) minus ⟨119877 (119866

119894 119905)⟩]2⟩ ⟨[119877 (119866

119894 119905 minus Δ119905) minus ⟨119877 (119866

119894 119905 minus Δ119905)⟩

2]⟩

(7)

where ⟨⟩ is a temporal average always performed over theinvestigated time period 119905 represents a time Δ119905 represents atime interval119877(119866

119894 119905) is the stock returns in a time intervalΔ119905

and119877(119866119894 119905) = ln[119901(119866

119894 119905)]minusln[119901(119866

119894 119905minusΔ119905)] is the logarithmic

difference of current stock price and the stock price before thetime interval Δ119905

According to the data point and time interval Δ119905 originaldata can be decomposed into several chunks The boundaryof 119876(119866

119894 Δ119905) is minus1 le 119876(119866

119894 Δ119905) le 1 If the 119876(119866

119894 Δ119905) is

positive the influence of chunk 119866119894on the time interval Δ119905 is

positive and vice versa We can calculate all the correlationintensity119876(119866

119894 Δ119905) according to (7) and get a119898times119898matrix of

4 Mathematical Problems in Engineering

Input data

SVR Mapping

Output model

CombiningDecomposingG1

G2

Gm

middot middot middot middot middot middot middot middot middot

SVG1

SVG2

SVGm

GSVM1

GSVM2

GSVM119898

GSVM

Figure 1 Model of directed-weighted chunking SVMs

IBM close value [1999-12-312013-12-31]Last 18757

Volume (millions)3619700

Moving average convergence divergence (12 26 9)

MACD 0957

Signal 0250

50

100

150

200

0

10

20

30

40

minus5

0

5

Dec 31 1999 Jun 01 2001 Dec 02 2002 Jun 01 2004 Dec 01 2005 Jun 01 2007 Dec 01 2008 Jun 01 2010 Dec 01 2011 Jun 03 2013

Figure 2 The IBM stock daily close prices (from December 31 1999 up to December 31 2013) Source Yahoo Finance httpfinanceyahoocom

correlation intensity Using these correlation intensity valueswe can get a relationship matrix with direction and weightNow dual form of the original optimization problem can bededuced as

max

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

+

119897

sum

119894=1

119886119894(119889119894minus 120576) minus 119886

lowast

119894(119889119894+ 120576)

Subject to 119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

0 le 119886lowast

119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

(8)

Now the solution of this equation is the overall optimalsolution of the original optimization problem

4 Experiments

In order to make a fair and thorough comparative betweendirected-weighted chunking SVMs and ordinary SVMs theIBM stock daily close prices are selected as shown in Figure 2The data points cover the time period from December 311999 up to December 31 2013 which has 21132 data points

Data points starting from December 31 1999 up toDecember 31 2007 (12075 data points) are used for trainingand data points starting from January 1 2008 up to theDecember 31 2013 (9057 data points) are used for testingNow we decomposed the training data into 1208 chunks bythe time intervals of 10 days and calculated the correlationintensity 119876(119866

119894 Δ119905) according to the function (7) Finally we

obtain a 1208 times 1208matrix of correlation intensity

41 Forecast Accuracy Assessment of Directed-Weighted ChunkSVMs The prediction performance is evaluated by using thefollowing statistical metrics namely the normalized meansquared error (NMSE) the mean absolute error (MAE) andthe directional symmetry (DS) These criteria are calculatedas (9) NMSE and MAE are the measures of the deviationbetween the actual and predicted valuesThe values of NMSE

Mathematical Problems in Engineering 5IB

M cl

ose v

alue

Real valuePredicted values in ordinary SVMPredicted values in D-W chunking SVM

100

150

200

Nov 2008 Nov 2009 Nov 2010 Nov 2011 Nov 2012 Oct 2013

Figure 3The predicted value of IBM stock daily close prices in testdata set (from January 01 2008 up to December 31 2013)

and MAE denote the accuracy of prediction A detaileddescription of performance metrics in financial forecastingcan be referred to in Abecasis [34]

NMSE = 1

1205902119873

119873

sum

119894=1

(119910119894minus 119910119894)2

where 1205902 = 1

119873 minus 1

119873

sum

119894=1

(119910119894minus 119910119894)2 119910119894=

1

119873

119873

sum

119894=1

119910119894

MAE = 1

119873

119873

sum

119894=1

1003816100381610038161003816119910119894minus 119910119894

1003816100381610038161003816

DS = 100

119873

119873

sum

119894=1

119889119894 where 119889

119894=

1 (119910119894minus 119910119894minus1) (119910119894minus 119910119894minus1)

0 otherwise(9)

The program of directed-weighted chunking SVMs algo-rithm is developed using R language In this paper theGaussian function is used as the kernel function of the SVMsThe experiments show that a width value of the Gaussianfunction of 002 is found to produce the best possible results119862 and 120576 are arbitrarily chosen to be 85 and 10minus3 respectively

We calculate the SVMs on the training data sets (fromDecember 31 1999 up to December 31 2007) and then obtainthe trained model Finally we obtain the predicted result byapplying the trained model on the test data set (from January1 2008 up to December 31 2013) In order to comparethe differences of various algorithms real value predictedvalue in ordinary SVMs and the predicted value in directed-weighted chunking SVMs are plotted in Figure 3

In Figure 3 we can clearly see that these two forecastingmethods are very precise but it is hard to tell which one ismore excellent So we calculated the performance criteriarespectively as shown in Table 1 By comparing these datawe find that the NMSE MAE and DS of directed-weightedchunking SVMs are 03760 01325 and 3829 on the trainingset and 10121 02846 and 4378 on the test set It is evidentthat these values are much smaller than those of ordinarySVMs which indicates that there is a smaller deviation

Table 1 Results of accuracy performance criteria

D-W chunking SVMs Ordinary SVMsNMSE MAE DS NMSE MAE DS

Training set 03760 01325 3829 05625 01778 4167Test set 10121 02846 4378 13574 02564 4897

Table 2 Performance comparison between the traditional SVMsand directed-weighted chunk SVMs

SVMmethod Sensitivity Specificity CPU time(ms)

Traditional SVMs 079 081 458231Directed-weightedchunk SVMs 083 085 65367

Dataset IBM stock daily close prices in training data set (12075 data pointsfrom December 31 1999 up to December 31 2007)SVMs parameters kernel is Gaussian function with 120590 = 002 119862 = 85 120576 =0001 and tolerance = 0001Chunking methods now we decomposed the training data into 1208 chunksby the time intervals of 10 daysThe CPU time covered the execution of entire algorithm excluding the fileIO time

between the actual values and predicted values with directed-weighted chunking SVMs

42 Calculation Performance of Directed-Weighted ChunkSVMs As is well known the performance of SVM dependson the parameters But it is difficult to choose suitableparameters for different problems Chunk algorithm reusedthe Hessian matrix elements from one to the next which canimprove the performance sharply

The calculation performance of all algorithms is mea-sured on an unloaded AMD E-350 16GHz processor run-ning Windows 7 and R 301 The same experiment will bedone on the data set of IBM stock daily close prices Theresults of the experiments are shown in Table 2

The primary purpose of these experiments is to examinedifferences of training times between two methods An over-all comparison of the SVMmethods can be found in Table 2Compared to the traditional SVMs directed-weighted chunkSVMs can improve the accuracy and decrease run timessharply Additionally the directed-weighted chunk SVMsmethod allows users to add machines to make the algorithmtraining even faster

43 Analysis of Optimal Number of Chunks in Directed-Weighted Chunk SVMs In the experiment described abovewe decomposed the training data into 1208 chunks by thetime intervals of 10 days arbitrarily and got a satisfactorypredictionOriginal training data set can be decomposed into500 chunks or 5000 chunks also Doing the same experimentson the same training sets by different chunks we will geta series of performance data Plotting the curve based onchunks number and NMSE values (in Figure 4) we canintuitively discover relationships between chunks numberand errors and get the optimal number of chunks

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 2: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

2 Mathematical Problems in Engineering

support vector machine methods are proposed One wayis to combine SVMs with some other methods such asactive learning [14 15] multitask learning [16 17] multiviewlearning [18] and semisupervised learning [19 20] Anotherway is to develop some optimization techniques of trainingalgorithms in SVMs such as sequential updating methodslike kernel-Adatron algorithm [21] successive over relaxationalgorithm [22] working set methods like chunking SVMs[23 24] reduced support vector machine (RSVM) [25] andsequential minimal optimization algorithm (SMO) [26]

Chunking SVMs reduced support vector machine(RSVM) and SVMs with sequential minimal optimization(SMO) are outstanding methods in dealing with massivedata set For example Lee and Mangasarian [25] designed areduced support vector machine algorithm (RSVM) whichcan greatly reduce the size of the quadratic program to besolved by reducing the data set volume so the memory usageis much smaller than that of a conventional SVM using theentire data set Osuna et al [27] designed a decompositionalgorithm that is guaranteed to solve the QP problem andthat does not make assumptions on the expected number ofsupport vectors Platt [26] put forward a sequential minimaloptimization algorithm (SMO) which breaks the large QPproblem into a series of smallest possible QP problems whichare analytically solvable in order to speed up the trainingtime Tay and Cao [28] proposed to combine support vectormachines (SVMs) with self-organizing feature map (SOM)for financial time series forecasting where SOM is used asa clustering algorithm to partition the whole input spaceinto several disjoint regions Tay and Cao [29] also putforward C-ascending support vector machines to amend the120576-insensitive errors in the nonstationary financial time series

Most of the improved support vector machine methodsdo well in memory requirement and CPU time but theforecast accuracy has declinedmore or less For financial timeseries prediction the typical massive-scale data we shouldpay more attention to the prediction accuracy while reducingthe computational complexity In this paper we proposeda directed-weighted chunking SVMs algorithm which canimprove the operation speed without reducing predictionaccuracy

This paper consists of five sections Section 1 introducesthe basic algorithm of SVM Section 2 proposes a directed-weighted chunking SVMs algorithm Section 3 designs aseries of experiments and empirical results are summarizedand discussed Section 4 presents the conclusions and limita-tions of this study

2 SVMs Regression Theory

In this section we will briefly introduce the support vectorsregression (SVR) theory Suppose there are a given set of datapoints 119866 = (119909

119894 119889119894)119897

119894(119909119894is the input vector 119889

119894is the desired

value) SVMs approximate the function in the following form

119910 =

119897

sum

119894=1

119908119894120601119894 (119909) + 119887 (1)

where 120601(119909)119897119894=1

are the features of inputs and 119908119894119897

119894=1 119887 are

coefficientsAccording to the structural risk minimization principle

the support vectors regression problem can be expressed as

Minimize 119877 (119908 120585(lowast)) =

1

2

119908119879119908 + 119862

lowast

119897

sum

119894=1

(120585119894+ 120585lowast

119894)

Subject to 119908120601 (119909119894) + 119887 minus 119889

119894le 120576 + 120585

119894

119889119894minus 119908120601 (119909

119894) minus 119887 le 120576 + 120585

lowast

119894

120585119894 120585lowast

119894ge 0

119894 = 1 2 119897

(2)

where 119909119894is mapped to a higher dimensional space by the

function 120601 120585119894and 120585

119894

lowast are slack variables (120585119894is the upper

training error 120585lowast119894is the lower) which are subject to the 120576-

insensitive tube |119889119894minus(119908120601(119909

119894)+119887)| le 120576 The parameters which

control the regression quality are the cost of error119862 thewidthof the tube 120576 and the mapping function 120601

Thus (1) becomes the following explicit form

119891 (119909 119886119894 119886lowast

119894) =

119897

sum

119894=1

(119886119894minus 119886lowast

119894)119870 (119909 119909

119894) + 119887 (3)

Then we can obtain the following form by maximizingthe dual form of function (3)

120601 (119886119894 119886lowast

119894) =

119897

sum

119894=1

119889119894(119886119894minus 119886lowast

119894) minus 120576

119897

sum

119894=1

(119886119894+ 119886lowast

119894)

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

(4)

with the following constraints

119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862 119894 = 1 2 119897

0 le 119886lowast

119894le 119862 119894 = 1 2 119897

(5)

where the 119886119894 119886lowast119894are the Lagrange multipliers introduced

and 119870(119909 119909119894) is named the kernel function The value is

equal to the inner product of two vectors 119909119894and 119909

119895in the

feature spaces 120601(119909119894) and 120601(119909

119895) There are many choices of the

kernel function common examples are the polynomial kernel119870(119909119894 119909119895) = (119909

119894119909119895+ 1)119889 and Gaussian kernel 119870(119909

119894 119909119895) =

exp(minus1205752(119909119894minus 119909119895)2)

Training SVMs is equivalent to optimizing the Lagrangemultipliers 119886

119894 119886lowast119894with constraints based on (4) Good fitting

function can be obtained by choosing appropriate functionalspace

There are many different researches using the straight-forward approaches to construct and implement SVMs

Mathematical Problems in Engineering 3

for financial time series analysis (see eg [9 13 30])other methods like chunking SVMs reduced support vec-tor machine (RSVM) and SVMs with sequential minimaloptimization (SMO) are used to deal with massive-scaledata sets Among these methods SVM chunking providesan alternative method to running a typical SVM on a dataset by breaking up the training data and running the SVMon smaller chunks of data In previous literature manydecomposition and combination methods are proposed

We would like to mention that for financial time seriesprediction problems the data set from different periodswill have different effects on current forecasts and thechange of direction of past stock price will also affect thecurrent forecasts Therefore we proposed directed-weightedchunking SVMs which can improve the operation speedwithout reducing prediction accuracy

3 The Directed-Weighted Chunking SVMs

31 Chunking Model in Support Vector Regression TrainingSVMs is equivalent to solving a linearly constrained QPproblem Training SVMs depends on QP optimization tech-niques Standard QP techniques cannot be directly applied toSVMsproblemswithmassive-scale data set In training stagesof directed-weighted chunking SVMs the whole trainingset is decomposed into several chunks and support vectorsare calculated respectively in their working subset In theprediction stage all these support vectors are combinedinto new working data set to get the model in accordancewith their importance as illustrated in Figure 1 and theprogress of directed-weighted chunking SVMs algorithm canbe described as follows

Step 1 Decompose the whole training set 119866 into 119898 subsets1198661 1198662 119866

119898 and 119866 = 119866

1cup 1198662 cup sdot sdot sdot cup119866

119898 119866119894cap 119866119895= Φ

Step 2 Calculate the support vector regression SVG119894for each

subset 119866119894

Step 3 Calculate the weight and direction of each subset

Step 4 Combine all support vectors 119866SVM119894 for each subsetinto new working data set 119866SVM

Step 5 Calculate the weighted support vector regression onthe new working data set 119866SVM and get the model

32 Directed-Weighted Chunking SVMs In the time seriesforecasting such as stock market the effect of past stockprices on the future stock prices will be different Usuallythe more recent the period of time the greater the weightcoefficient According to a certain time interval the total timeseries of stock priceswill be divided into different chunks andsupport vectors in each chunk are calculated respectivelyFurthermore the weighted support vector regressions arecalculated to obtain the forecast model

The stock market is one of the complex systems [31 32]We can treat chunks as the nodes and relationship betweenchunks as edges Thus these nodes and edges will be acomplex networkThe entire time series of stock price can beregarded as a directed-weighted network with a large numberof nodes and edges The mutual influence between chunkshas its direction For example in chunks G1 and G2 if thereexists a nonzero correlation coefficient from chunk G1 to G2we will draw a directed edge between G1 and G2 Becausethe strength of mutual influence between chunks is differentthe edge weight which is called correlation intensity isalso different So simply chunking SVMs cannot reflect theinfluence of respective chunk on final model Here we willintroduce directed-weighted chunking algorithm into SVMs

In traditional SVM regression optimization problem theparameter 119862lowast is a constant In order to show the differentinfluence on the prediction results we introduce a function119876() and modify the SVMs regression function as follows

119877 (119908 120585(lowast)) =

1

2

119908119879119908 + 119862

lowast

119897

sum

119894=1

119876 (119866119894 Δ119905) (120585

119894+ 120585lowast

119894) (6)

In (6) 119876(119866119894 Δ119905) can be understood as the weight of

each chunk and the values of 119876(119866119894 Δ119905) can be positive or

negative The positive or negative values of 119876(119866119894 Δ119905) will

change parameter 119862lowast in SVMs which is very similar tothat of the positive or negative information impact on thefuture stock price According to the definition of networkcorrelation coefficient introduced by Bonanno et al [33] wecan define the correlation intensity 119876(119866

119894 Δ119905) (influence of

chunk 119866119894by the time interval Δ119905) as follows

119876 (119866119894 Δ119905) =

⟨119877 (119866119894 119905) 119877 (119866

119894 119905 minus Δ119905)⟩ minus ⟨119877 (119866

119894 119905)⟩ ⟨119877 (119866

119894 119905 minus Δ119905)⟩

radic⟨[119877 (119866119894 119905) minus ⟨119877 (119866

119894 119905)⟩]2⟩ ⟨[119877 (119866

119894 119905 minus Δ119905) minus ⟨119877 (119866

119894 119905 minus Δ119905)⟩

2]⟩

(7)

where ⟨⟩ is a temporal average always performed over theinvestigated time period 119905 represents a time Δ119905 represents atime interval119877(119866

119894 119905) is the stock returns in a time intervalΔ119905

and119877(119866119894 119905) = ln[119901(119866

119894 119905)]minusln[119901(119866

119894 119905minusΔ119905)] is the logarithmic

difference of current stock price and the stock price before thetime interval Δ119905

According to the data point and time interval Δ119905 originaldata can be decomposed into several chunks The boundaryof 119876(119866

119894 Δ119905) is minus1 le 119876(119866

119894 Δ119905) le 1 If the 119876(119866

119894 Δ119905) is

positive the influence of chunk 119866119894on the time interval Δ119905 is

positive and vice versa We can calculate all the correlationintensity119876(119866

119894 Δ119905) according to (7) and get a119898times119898matrix of

4 Mathematical Problems in Engineering

Input data

SVR Mapping

Output model

CombiningDecomposingG1

G2

Gm

middot middot middot middot middot middot middot middot middot

SVG1

SVG2

SVGm

GSVM1

GSVM2

GSVM119898

GSVM

Figure 1 Model of directed-weighted chunking SVMs

IBM close value [1999-12-312013-12-31]Last 18757

Volume (millions)3619700

Moving average convergence divergence (12 26 9)

MACD 0957

Signal 0250

50

100

150

200

0

10

20

30

40

minus5

0

5

Dec 31 1999 Jun 01 2001 Dec 02 2002 Jun 01 2004 Dec 01 2005 Jun 01 2007 Dec 01 2008 Jun 01 2010 Dec 01 2011 Jun 03 2013

Figure 2 The IBM stock daily close prices (from December 31 1999 up to December 31 2013) Source Yahoo Finance httpfinanceyahoocom

correlation intensity Using these correlation intensity valueswe can get a relationship matrix with direction and weightNow dual form of the original optimization problem can bededuced as

max

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

+

119897

sum

119894=1

119886119894(119889119894minus 120576) minus 119886

lowast

119894(119889119894+ 120576)

Subject to 119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

0 le 119886lowast

119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

(8)

Now the solution of this equation is the overall optimalsolution of the original optimization problem

4 Experiments

In order to make a fair and thorough comparative betweendirected-weighted chunking SVMs and ordinary SVMs theIBM stock daily close prices are selected as shown in Figure 2The data points cover the time period from December 311999 up to December 31 2013 which has 21132 data points

Data points starting from December 31 1999 up toDecember 31 2007 (12075 data points) are used for trainingand data points starting from January 1 2008 up to theDecember 31 2013 (9057 data points) are used for testingNow we decomposed the training data into 1208 chunks bythe time intervals of 10 days and calculated the correlationintensity 119876(119866

119894 Δ119905) according to the function (7) Finally we

obtain a 1208 times 1208matrix of correlation intensity

41 Forecast Accuracy Assessment of Directed-Weighted ChunkSVMs The prediction performance is evaluated by using thefollowing statistical metrics namely the normalized meansquared error (NMSE) the mean absolute error (MAE) andthe directional symmetry (DS) These criteria are calculatedas (9) NMSE and MAE are the measures of the deviationbetween the actual and predicted valuesThe values of NMSE

Mathematical Problems in Engineering 5IB

M cl

ose v

alue

Real valuePredicted values in ordinary SVMPredicted values in D-W chunking SVM

100

150

200

Nov 2008 Nov 2009 Nov 2010 Nov 2011 Nov 2012 Oct 2013

Figure 3The predicted value of IBM stock daily close prices in testdata set (from January 01 2008 up to December 31 2013)

and MAE denote the accuracy of prediction A detaileddescription of performance metrics in financial forecastingcan be referred to in Abecasis [34]

NMSE = 1

1205902119873

119873

sum

119894=1

(119910119894minus 119910119894)2

where 1205902 = 1

119873 minus 1

119873

sum

119894=1

(119910119894minus 119910119894)2 119910119894=

1

119873

119873

sum

119894=1

119910119894

MAE = 1

119873

119873

sum

119894=1

1003816100381610038161003816119910119894minus 119910119894

1003816100381610038161003816

DS = 100

119873

119873

sum

119894=1

119889119894 where 119889

119894=

1 (119910119894minus 119910119894minus1) (119910119894minus 119910119894minus1)

0 otherwise(9)

The program of directed-weighted chunking SVMs algo-rithm is developed using R language In this paper theGaussian function is used as the kernel function of the SVMsThe experiments show that a width value of the Gaussianfunction of 002 is found to produce the best possible results119862 and 120576 are arbitrarily chosen to be 85 and 10minus3 respectively

We calculate the SVMs on the training data sets (fromDecember 31 1999 up to December 31 2007) and then obtainthe trained model Finally we obtain the predicted result byapplying the trained model on the test data set (from January1 2008 up to December 31 2013) In order to comparethe differences of various algorithms real value predictedvalue in ordinary SVMs and the predicted value in directed-weighted chunking SVMs are plotted in Figure 3

In Figure 3 we can clearly see that these two forecastingmethods are very precise but it is hard to tell which one ismore excellent So we calculated the performance criteriarespectively as shown in Table 1 By comparing these datawe find that the NMSE MAE and DS of directed-weightedchunking SVMs are 03760 01325 and 3829 on the trainingset and 10121 02846 and 4378 on the test set It is evidentthat these values are much smaller than those of ordinarySVMs which indicates that there is a smaller deviation

Table 1 Results of accuracy performance criteria

D-W chunking SVMs Ordinary SVMsNMSE MAE DS NMSE MAE DS

Training set 03760 01325 3829 05625 01778 4167Test set 10121 02846 4378 13574 02564 4897

Table 2 Performance comparison between the traditional SVMsand directed-weighted chunk SVMs

SVMmethod Sensitivity Specificity CPU time(ms)

Traditional SVMs 079 081 458231Directed-weightedchunk SVMs 083 085 65367

Dataset IBM stock daily close prices in training data set (12075 data pointsfrom December 31 1999 up to December 31 2007)SVMs parameters kernel is Gaussian function with 120590 = 002 119862 = 85 120576 =0001 and tolerance = 0001Chunking methods now we decomposed the training data into 1208 chunksby the time intervals of 10 daysThe CPU time covered the execution of entire algorithm excluding the fileIO time

between the actual values and predicted values with directed-weighted chunking SVMs

42 Calculation Performance of Directed-Weighted ChunkSVMs As is well known the performance of SVM dependson the parameters But it is difficult to choose suitableparameters for different problems Chunk algorithm reusedthe Hessian matrix elements from one to the next which canimprove the performance sharply

The calculation performance of all algorithms is mea-sured on an unloaded AMD E-350 16GHz processor run-ning Windows 7 and R 301 The same experiment will bedone on the data set of IBM stock daily close prices Theresults of the experiments are shown in Table 2

The primary purpose of these experiments is to examinedifferences of training times between two methods An over-all comparison of the SVMmethods can be found in Table 2Compared to the traditional SVMs directed-weighted chunkSVMs can improve the accuracy and decrease run timessharply Additionally the directed-weighted chunk SVMsmethod allows users to add machines to make the algorithmtraining even faster

43 Analysis of Optimal Number of Chunks in Directed-Weighted Chunk SVMs In the experiment described abovewe decomposed the training data into 1208 chunks by thetime intervals of 10 days arbitrarily and got a satisfactorypredictionOriginal training data set can be decomposed into500 chunks or 5000 chunks also Doing the same experimentson the same training sets by different chunks we will geta series of performance data Plotting the curve based onchunks number and NMSE values (in Figure 4) we canintuitively discover relationships between chunks numberand errors and get the optimal number of chunks

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 3: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

Mathematical Problems in Engineering 3

for financial time series analysis (see eg [9 13 30])other methods like chunking SVMs reduced support vec-tor machine (RSVM) and SVMs with sequential minimaloptimization (SMO) are used to deal with massive-scaledata sets Among these methods SVM chunking providesan alternative method to running a typical SVM on a dataset by breaking up the training data and running the SVMon smaller chunks of data In previous literature manydecomposition and combination methods are proposed

We would like to mention that for financial time seriesprediction problems the data set from different periodswill have different effects on current forecasts and thechange of direction of past stock price will also affect thecurrent forecasts Therefore we proposed directed-weightedchunking SVMs which can improve the operation speedwithout reducing prediction accuracy

3 The Directed-Weighted Chunking SVMs

31 Chunking Model in Support Vector Regression TrainingSVMs is equivalent to solving a linearly constrained QPproblem Training SVMs depends on QP optimization tech-niques Standard QP techniques cannot be directly applied toSVMsproblemswithmassive-scale data set In training stagesof directed-weighted chunking SVMs the whole trainingset is decomposed into several chunks and support vectorsare calculated respectively in their working subset In theprediction stage all these support vectors are combinedinto new working data set to get the model in accordancewith their importance as illustrated in Figure 1 and theprogress of directed-weighted chunking SVMs algorithm canbe described as follows

Step 1 Decompose the whole training set 119866 into 119898 subsets1198661 1198662 119866

119898 and 119866 = 119866

1cup 1198662 cup sdot sdot sdot cup119866

119898 119866119894cap 119866119895= Φ

Step 2 Calculate the support vector regression SVG119894for each

subset 119866119894

Step 3 Calculate the weight and direction of each subset

Step 4 Combine all support vectors 119866SVM119894 for each subsetinto new working data set 119866SVM

Step 5 Calculate the weighted support vector regression onthe new working data set 119866SVM and get the model

32 Directed-Weighted Chunking SVMs In the time seriesforecasting such as stock market the effect of past stockprices on the future stock prices will be different Usuallythe more recent the period of time the greater the weightcoefficient According to a certain time interval the total timeseries of stock priceswill be divided into different chunks andsupport vectors in each chunk are calculated respectivelyFurthermore the weighted support vector regressions arecalculated to obtain the forecast model

The stock market is one of the complex systems [31 32]We can treat chunks as the nodes and relationship betweenchunks as edges Thus these nodes and edges will be acomplex networkThe entire time series of stock price can beregarded as a directed-weighted network with a large numberof nodes and edges The mutual influence between chunkshas its direction For example in chunks G1 and G2 if thereexists a nonzero correlation coefficient from chunk G1 to G2we will draw a directed edge between G1 and G2 Becausethe strength of mutual influence between chunks is differentthe edge weight which is called correlation intensity isalso different So simply chunking SVMs cannot reflect theinfluence of respective chunk on final model Here we willintroduce directed-weighted chunking algorithm into SVMs

In traditional SVM regression optimization problem theparameter 119862lowast is a constant In order to show the differentinfluence on the prediction results we introduce a function119876() and modify the SVMs regression function as follows

119877 (119908 120585(lowast)) =

1

2

119908119879119908 + 119862

lowast

119897

sum

119894=1

119876 (119866119894 Δ119905) (120585

119894+ 120585lowast

119894) (6)

In (6) 119876(119866119894 Δ119905) can be understood as the weight of

each chunk and the values of 119876(119866119894 Δ119905) can be positive or

negative The positive or negative values of 119876(119866119894 Δ119905) will

change parameter 119862lowast in SVMs which is very similar tothat of the positive or negative information impact on thefuture stock price According to the definition of networkcorrelation coefficient introduced by Bonanno et al [33] wecan define the correlation intensity 119876(119866

119894 Δ119905) (influence of

chunk 119866119894by the time interval Δ119905) as follows

119876 (119866119894 Δ119905) =

⟨119877 (119866119894 119905) 119877 (119866

119894 119905 minus Δ119905)⟩ minus ⟨119877 (119866

119894 119905)⟩ ⟨119877 (119866

119894 119905 minus Δ119905)⟩

radic⟨[119877 (119866119894 119905) minus ⟨119877 (119866

119894 119905)⟩]2⟩ ⟨[119877 (119866

119894 119905 minus Δ119905) minus ⟨119877 (119866

119894 119905 minus Δ119905)⟩

2]⟩

(7)

where ⟨⟩ is a temporal average always performed over theinvestigated time period 119905 represents a time Δ119905 represents atime interval119877(119866

119894 119905) is the stock returns in a time intervalΔ119905

and119877(119866119894 119905) = ln[119901(119866

119894 119905)]minusln[119901(119866

119894 119905minusΔ119905)] is the logarithmic

difference of current stock price and the stock price before thetime interval Δ119905

According to the data point and time interval Δ119905 originaldata can be decomposed into several chunks The boundaryof 119876(119866

119894 Δ119905) is minus1 le 119876(119866

119894 Δ119905) le 1 If the 119876(119866

119894 Δ119905) is

positive the influence of chunk 119866119894on the time interval Δ119905 is

positive and vice versa We can calculate all the correlationintensity119876(119866

119894 Δ119905) according to (7) and get a119898times119898matrix of

4 Mathematical Problems in Engineering

Input data

SVR Mapping

Output model

CombiningDecomposingG1

G2

Gm

middot middot middot middot middot middot middot middot middot

SVG1

SVG2

SVGm

GSVM1

GSVM2

GSVM119898

GSVM

Figure 1 Model of directed-weighted chunking SVMs

IBM close value [1999-12-312013-12-31]Last 18757

Volume (millions)3619700

Moving average convergence divergence (12 26 9)

MACD 0957

Signal 0250

50

100

150

200

0

10

20

30

40

minus5

0

5

Dec 31 1999 Jun 01 2001 Dec 02 2002 Jun 01 2004 Dec 01 2005 Jun 01 2007 Dec 01 2008 Jun 01 2010 Dec 01 2011 Jun 03 2013

Figure 2 The IBM stock daily close prices (from December 31 1999 up to December 31 2013) Source Yahoo Finance httpfinanceyahoocom

correlation intensity Using these correlation intensity valueswe can get a relationship matrix with direction and weightNow dual form of the original optimization problem can bededuced as

max

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

+

119897

sum

119894=1

119886119894(119889119894minus 120576) minus 119886

lowast

119894(119889119894+ 120576)

Subject to 119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

0 le 119886lowast

119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

(8)

Now the solution of this equation is the overall optimalsolution of the original optimization problem

4 Experiments

In order to make a fair and thorough comparative betweendirected-weighted chunking SVMs and ordinary SVMs theIBM stock daily close prices are selected as shown in Figure 2The data points cover the time period from December 311999 up to December 31 2013 which has 21132 data points

Data points starting from December 31 1999 up toDecember 31 2007 (12075 data points) are used for trainingand data points starting from January 1 2008 up to theDecember 31 2013 (9057 data points) are used for testingNow we decomposed the training data into 1208 chunks bythe time intervals of 10 days and calculated the correlationintensity 119876(119866

119894 Δ119905) according to the function (7) Finally we

obtain a 1208 times 1208matrix of correlation intensity

41 Forecast Accuracy Assessment of Directed-Weighted ChunkSVMs The prediction performance is evaluated by using thefollowing statistical metrics namely the normalized meansquared error (NMSE) the mean absolute error (MAE) andthe directional symmetry (DS) These criteria are calculatedas (9) NMSE and MAE are the measures of the deviationbetween the actual and predicted valuesThe values of NMSE

Mathematical Problems in Engineering 5IB

M cl

ose v

alue

Real valuePredicted values in ordinary SVMPredicted values in D-W chunking SVM

100

150

200

Nov 2008 Nov 2009 Nov 2010 Nov 2011 Nov 2012 Oct 2013

Figure 3The predicted value of IBM stock daily close prices in testdata set (from January 01 2008 up to December 31 2013)

and MAE denote the accuracy of prediction A detaileddescription of performance metrics in financial forecastingcan be referred to in Abecasis [34]

NMSE = 1

1205902119873

119873

sum

119894=1

(119910119894minus 119910119894)2

where 1205902 = 1

119873 minus 1

119873

sum

119894=1

(119910119894minus 119910119894)2 119910119894=

1

119873

119873

sum

119894=1

119910119894

MAE = 1

119873

119873

sum

119894=1

1003816100381610038161003816119910119894minus 119910119894

1003816100381610038161003816

DS = 100

119873

119873

sum

119894=1

119889119894 where 119889

119894=

1 (119910119894minus 119910119894minus1) (119910119894minus 119910119894minus1)

0 otherwise(9)

The program of directed-weighted chunking SVMs algo-rithm is developed using R language In this paper theGaussian function is used as the kernel function of the SVMsThe experiments show that a width value of the Gaussianfunction of 002 is found to produce the best possible results119862 and 120576 are arbitrarily chosen to be 85 and 10minus3 respectively

We calculate the SVMs on the training data sets (fromDecember 31 1999 up to December 31 2007) and then obtainthe trained model Finally we obtain the predicted result byapplying the trained model on the test data set (from January1 2008 up to December 31 2013) In order to comparethe differences of various algorithms real value predictedvalue in ordinary SVMs and the predicted value in directed-weighted chunking SVMs are plotted in Figure 3

In Figure 3 we can clearly see that these two forecastingmethods are very precise but it is hard to tell which one ismore excellent So we calculated the performance criteriarespectively as shown in Table 1 By comparing these datawe find that the NMSE MAE and DS of directed-weightedchunking SVMs are 03760 01325 and 3829 on the trainingset and 10121 02846 and 4378 on the test set It is evidentthat these values are much smaller than those of ordinarySVMs which indicates that there is a smaller deviation

Table 1 Results of accuracy performance criteria

D-W chunking SVMs Ordinary SVMsNMSE MAE DS NMSE MAE DS

Training set 03760 01325 3829 05625 01778 4167Test set 10121 02846 4378 13574 02564 4897

Table 2 Performance comparison between the traditional SVMsand directed-weighted chunk SVMs

SVMmethod Sensitivity Specificity CPU time(ms)

Traditional SVMs 079 081 458231Directed-weightedchunk SVMs 083 085 65367

Dataset IBM stock daily close prices in training data set (12075 data pointsfrom December 31 1999 up to December 31 2007)SVMs parameters kernel is Gaussian function with 120590 = 002 119862 = 85 120576 =0001 and tolerance = 0001Chunking methods now we decomposed the training data into 1208 chunksby the time intervals of 10 daysThe CPU time covered the execution of entire algorithm excluding the fileIO time

between the actual values and predicted values with directed-weighted chunking SVMs

42 Calculation Performance of Directed-Weighted ChunkSVMs As is well known the performance of SVM dependson the parameters But it is difficult to choose suitableparameters for different problems Chunk algorithm reusedthe Hessian matrix elements from one to the next which canimprove the performance sharply

The calculation performance of all algorithms is mea-sured on an unloaded AMD E-350 16GHz processor run-ning Windows 7 and R 301 The same experiment will bedone on the data set of IBM stock daily close prices Theresults of the experiments are shown in Table 2

The primary purpose of these experiments is to examinedifferences of training times between two methods An over-all comparison of the SVMmethods can be found in Table 2Compared to the traditional SVMs directed-weighted chunkSVMs can improve the accuracy and decrease run timessharply Additionally the directed-weighted chunk SVMsmethod allows users to add machines to make the algorithmtraining even faster

43 Analysis of Optimal Number of Chunks in Directed-Weighted Chunk SVMs In the experiment described abovewe decomposed the training data into 1208 chunks by thetime intervals of 10 days arbitrarily and got a satisfactorypredictionOriginal training data set can be decomposed into500 chunks or 5000 chunks also Doing the same experimentson the same training sets by different chunks we will geta series of performance data Plotting the curve based onchunks number and NMSE values (in Figure 4) we canintuitively discover relationships between chunks numberand errors and get the optimal number of chunks

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 4: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

4 Mathematical Problems in Engineering

Input data

SVR Mapping

Output model

CombiningDecomposingG1

G2

Gm

middot middot middot middot middot middot middot middot middot

SVG1

SVG2

SVGm

GSVM1

GSVM2

GSVM119898

GSVM

Figure 1 Model of directed-weighted chunking SVMs

IBM close value [1999-12-312013-12-31]Last 18757

Volume (millions)3619700

Moving average convergence divergence (12 26 9)

MACD 0957

Signal 0250

50

100

150

200

0

10

20

30

40

minus5

0

5

Dec 31 1999 Jun 01 2001 Dec 02 2002 Jun 01 2004 Dec 01 2005 Jun 01 2007 Dec 01 2008 Jun 01 2010 Dec 01 2011 Jun 03 2013

Figure 2 The IBM stock daily close prices (from December 31 1999 up to December 31 2013) Source Yahoo Finance httpfinanceyahoocom

correlation intensity Using these correlation intensity valueswe can get a relationship matrix with direction and weightNow dual form of the original optimization problem can bededuced as

max

minus

1

2

119897

sum

119894=1

119897

sum

119895=1

(119886119894minus 119886lowast

119894) (119886119894minus 119886lowast

119894)119870 (119909

119894 119909119895)

+

119897

sum

119894=1

119886119894(119889119894minus 120576) minus 119886

lowast

119894(119889119894+ 120576)

Subject to 119897

sum

119894=1

(119886119894minus 119886lowast

119894) = 0

0 le 119886119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

0 le 119886lowast

119894le 119862119876 (119866

119894 Δ119905) 119894 = 1 2 119897

(8)

Now the solution of this equation is the overall optimalsolution of the original optimization problem

4 Experiments

In order to make a fair and thorough comparative betweendirected-weighted chunking SVMs and ordinary SVMs theIBM stock daily close prices are selected as shown in Figure 2The data points cover the time period from December 311999 up to December 31 2013 which has 21132 data points

Data points starting from December 31 1999 up toDecember 31 2007 (12075 data points) are used for trainingand data points starting from January 1 2008 up to theDecember 31 2013 (9057 data points) are used for testingNow we decomposed the training data into 1208 chunks bythe time intervals of 10 days and calculated the correlationintensity 119876(119866

119894 Δ119905) according to the function (7) Finally we

obtain a 1208 times 1208matrix of correlation intensity

41 Forecast Accuracy Assessment of Directed-Weighted ChunkSVMs The prediction performance is evaluated by using thefollowing statistical metrics namely the normalized meansquared error (NMSE) the mean absolute error (MAE) andthe directional symmetry (DS) These criteria are calculatedas (9) NMSE and MAE are the measures of the deviationbetween the actual and predicted valuesThe values of NMSE

Mathematical Problems in Engineering 5IB

M cl

ose v

alue

Real valuePredicted values in ordinary SVMPredicted values in D-W chunking SVM

100

150

200

Nov 2008 Nov 2009 Nov 2010 Nov 2011 Nov 2012 Oct 2013

Figure 3The predicted value of IBM stock daily close prices in testdata set (from January 01 2008 up to December 31 2013)

and MAE denote the accuracy of prediction A detaileddescription of performance metrics in financial forecastingcan be referred to in Abecasis [34]

NMSE = 1

1205902119873

119873

sum

119894=1

(119910119894minus 119910119894)2

where 1205902 = 1

119873 minus 1

119873

sum

119894=1

(119910119894minus 119910119894)2 119910119894=

1

119873

119873

sum

119894=1

119910119894

MAE = 1

119873

119873

sum

119894=1

1003816100381610038161003816119910119894minus 119910119894

1003816100381610038161003816

DS = 100

119873

119873

sum

119894=1

119889119894 where 119889

119894=

1 (119910119894minus 119910119894minus1) (119910119894minus 119910119894minus1)

0 otherwise(9)

The program of directed-weighted chunking SVMs algo-rithm is developed using R language In this paper theGaussian function is used as the kernel function of the SVMsThe experiments show that a width value of the Gaussianfunction of 002 is found to produce the best possible results119862 and 120576 are arbitrarily chosen to be 85 and 10minus3 respectively

We calculate the SVMs on the training data sets (fromDecember 31 1999 up to December 31 2007) and then obtainthe trained model Finally we obtain the predicted result byapplying the trained model on the test data set (from January1 2008 up to December 31 2013) In order to comparethe differences of various algorithms real value predictedvalue in ordinary SVMs and the predicted value in directed-weighted chunking SVMs are plotted in Figure 3

In Figure 3 we can clearly see that these two forecastingmethods are very precise but it is hard to tell which one ismore excellent So we calculated the performance criteriarespectively as shown in Table 1 By comparing these datawe find that the NMSE MAE and DS of directed-weightedchunking SVMs are 03760 01325 and 3829 on the trainingset and 10121 02846 and 4378 on the test set It is evidentthat these values are much smaller than those of ordinarySVMs which indicates that there is a smaller deviation

Table 1 Results of accuracy performance criteria

D-W chunking SVMs Ordinary SVMsNMSE MAE DS NMSE MAE DS

Training set 03760 01325 3829 05625 01778 4167Test set 10121 02846 4378 13574 02564 4897

Table 2 Performance comparison between the traditional SVMsand directed-weighted chunk SVMs

SVMmethod Sensitivity Specificity CPU time(ms)

Traditional SVMs 079 081 458231Directed-weightedchunk SVMs 083 085 65367

Dataset IBM stock daily close prices in training data set (12075 data pointsfrom December 31 1999 up to December 31 2007)SVMs parameters kernel is Gaussian function with 120590 = 002 119862 = 85 120576 =0001 and tolerance = 0001Chunking methods now we decomposed the training data into 1208 chunksby the time intervals of 10 daysThe CPU time covered the execution of entire algorithm excluding the fileIO time

between the actual values and predicted values with directed-weighted chunking SVMs

42 Calculation Performance of Directed-Weighted ChunkSVMs As is well known the performance of SVM dependson the parameters But it is difficult to choose suitableparameters for different problems Chunk algorithm reusedthe Hessian matrix elements from one to the next which canimprove the performance sharply

The calculation performance of all algorithms is mea-sured on an unloaded AMD E-350 16GHz processor run-ning Windows 7 and R 301 The same experiment will bedone on the data set of IBM stock daily close prices Theresults of the experiments are shown in Table 2

The primary purpose of these experiments is to examinedifferences of training times between two methods An over-all comparison of the SVMmethods can be found in Table 2Compared to the traditional SVMs directed-weighted chunkSVMs can improve the accuracy and decrease run timessharply Additionally the directed-weighted chunk SVMsmethod allows users to add machines to make the algorithmtraining even faster

43 Analysis of Optimal Number of Chunks in Directed-Weighted Chunk SVMs In the experiment described abovewe decomposed the training data into 1208 chunks by thetime intervals of 10 days arbitrarily and got a satisfactorypredictionOriginal training data set can be decomposed into500 chunks or 5000 chunks also Doing the same experimentson the same training sets by different chunks we will geta series of performance data Plotting the curve based onchunks number and NMSE values (in Figure 4) we canintuitively discover relationships between chunks numberand errors and get the optimal number of chunks

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 5: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

Mathematical Problems in Engineering 5IB

M cl

ose v

alue

Real valuePredicted values in ordinary SVMPredicted values in D-W chunking SVM

100

150

200

Nov 2008 Nov 2009 Nov 2010 Nov 2011 Nov 2012 Oct 2013

Figure 3The predicted value of IBM stock daily close prices in testdata set (from January 01 2008 up to December 31 2013)

and MAE denote the accuracy of prediction A detaileddescription of performance metrics in financial forecastingcan be referred to in Abecasis [34]

NMSE = 1

1205902119873

119873

sum

119894=1

(119910119894minus 119910119894)2

where 1205902 = 1

119873 minus 1

119873

sum

119894=1

(119910119894minus 119910119894)2 119910119894=

1

119873

119873

sum

119894=1

119910119894

MAE = 1

119873

119873

sum

119894=1

1003816100381610038161003816119910119894minus 119910119894

1003816100381610038161003816

DS = 100

119873

119873

sum

119894=1

119889119894 where 119889

119894=

1 (119910119894minus 119910119894minus1) (119910119894minus 119910119894minus1)

0 otherwise(9)

The program of directed-weighted chunking SVMs algo-rithm is developed using R language In this paper theGaussian function is used as the kernel function of the SVMsThe experiments show that a width value of the Gaussianfunction of 002 is found to produce the best possible results119862 and 120576 are arbitrarily chosen to be 85 and 10minus3 respectively

We calculate the SVMs on the training data sets (fromDecember 31 1999 up to December 31 2007) and then obtainthe trained model Finally we obtain the predicted result byapplying the trained model on the test data set (from January1 2008 up to December 31 2013) In order to comparethe differences of various algorithms real value predictedvalue in ordinary SVMs and the predicted value in directed-weighted chunking SVMs are plotted in Figure 3

In Figure 3 we can clearly see that these two forecastingmethods are very precise but it is hard to tell which one ismore excellent So we calculated the performance criteriarespectively as shown in Table 1 By comparing these datawe find that the NMSE MAE and DS of directed-weightedchunking SVMs are 03760 01325 and 3829 on the trainingset and 10121 02846 and 4378 on the test set It is evidentthat these values are much smaller than those of ordinarySVMs which indicates that there is a smaller deviation

Table 1 Results of accuracy performance criteria

D-W chunking SVMs Ordinary SVMsNMSE MAE DS NMSE MAE DS

Training set 03760 01325 3829 05625 01778 4167Test set 10121 02846 4378 13574 02564 4897

Table 2 Performance comparison between the traditional SVMsand directed-weighted chunk SVMs

SVMmethod Sensitivity Specificity CPU time(ms)

Traditional SVMs 079 081 458231Directed-weightedchunk SVMs 083 085 65367

Dataset IBM stock daily close prices in training data set (12075 data pointsfrom December 31 1999 up to December 31 2007)SVMs parameters kernel is Gaussian function with 120590 = 002 119862 = 85 120576 =0001 and tolerance = 0001Chunking methods now we decomposed the training data into 1208 chunksby the time intervals of 10 daysThe CPU time covered the execution of entire algorithm excluding the fileIO time

between the actual values and predicted values with directed-weighted chunking SVMs

42 Calculation Performance of Directed-Weighted ChunkSVMs As is well known the performance of SVM dependson the parameters But it is difficult to choose suitableparameters for different problems Chunk algorithm reusedthe Hessian matrix elements from one to the next which canimprove the performance sharply

The calculation performance of all algorithms is mea-sured on an unloaded AMD E-350 16GHz processor run-ning Windows 7 and R 301 The same experiment will bedone on the data set of IBM stock daily close prices Theresults of the experiments are shown in Table 2

The primary purpose of these experiments is to examinedifferences of training times between two methods An over-all comparison of the SVMmethods can be found in Table 2Compared to the traditional SVMs directed-weighted chunkSVMs can improve the accuracy and decrease run timessharply Additionally the directed-weighted chunk SVMsmethod allows users to add machines to make the algorithmtraining even faster

43 Analysis of Optimal Number of Chunks in Directed-Weighted Chunk SVMs In the experiment described abovewe decomposed the training data into 1208 chunks by thetime intervals of 10 days arbitrarily and got a satisfactorypredictionOriginal training data set can be decomposed into500 chunks or 5000 chunks also Doing the same experimentson the same training sets by different chunks we will geta series of performance data Plotting the curve based onchunks number and NMSE values (in Figure 4) we canintuitively discover relationships between chunks numberand errors and get the optimal number of chunks

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 6: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

6 Mathematical Problems in Engineering

1000 400030002000

02

10

08

06

04

The number of chunks on training set

NM

SE

Figure 4 NMSE value in different chunks (IBM stock daily closeprices data)

According to NMSE criteria we get the minimumNMSEvalue 03173 on the point of chunks 2350That means that thebest number of chunks is 2350 Under this chunking numberwe can get the best performance of prediction

From Figure 4 when the chunks number is increasedNMSE value is declined rapidly But when the decreasereaches a certain value NMSE value will increase converselyHowever this upward trend is not very large which indicatesthat the directed-weighted chunking SVM is not a funda-mental transform of SVM but a limited improvement Butfrom the perspective of processing massive-scale data thisimprovement is very important

5 Conclusions

In this paper we proposed a new chunks algorithm in SVMsregression which combined the support vectors according totheir importance The proposed algorithm can improve thecomputational speed without reducing prediction accuracy

In our directed-weighted chunking SVMsΔ119905 the criteriafor the chunking is a constant but in practice Δ119905 canbe variable or some form of function In addition furtherstudies on the different kernel functions and more suitableparameters 120576 and 119862 can be done in order to improve theperformance of directed-weighted chunking SVMs

Conflict of Interests

The authors declare that they have no conflict of interestsregarding the publication of this paper

Acknowledgments

This work was partly supported by the National NaturalScience Foundation of China (NSFC) under Project Grant(no 71162015) and the Inner Mongolia Autonomous Region

Higher Education Development Plan of Innovation TeamsrsquoProject Grant (no NMGIRT1404)

References

[1] Y S Abu-Mostafa and A F Atiya ldquoIntroduction to financialforecastingrdquoApplied Intelligence vol 6 no 3 pp 205ndash213 1996

[2] W Cheng W Wagner and C H Lin ldquoForecasting the 30-yearUS treasury bond with a system of neural networksrdquo Journalof Computational Intelligence in Finance vol 4 no 1 pp 10ndash161996

[3] K-J Kim and I Han ldquoGenetic algorithms approach to featurediscretization in artificial neural networks for the prediction ofstock price indexrdquo Expert Systems with Applications vol 19 no2 pp 125ndash132 2000

[4] HAhmadi ldquoTestability of the arbitrage pricing theory by neuralnetworkrdquo in Proceedings of the International Joint Conference onNeural Networks (IJCNN rsquo90) pp 385ndash393 June 1990

[5] R Tsaih YHsu andCC Lai ldquoForecasting SampP 500 stock indexfutures with a hybrid AI systemrdquo Decision Support Systems vol23 no 2 pp 161ndash174 1998

[6] G G Szpiro ldquoForecasting chaotic time series with geneticalgorithmsrdquo Physical Review E vol 55 no 3 pp 2557ndash25681997

[7] A Brabazon andM OrsquoNeill Biologically Inspired Algorithms forFinancial Modelling Springer Berlin Germany 2006

[8] X Cai N Zhang G K Venayagamoorthy andD CWunsch IIldquoTime series prediction with recurrent neural networks trainedby a hybrid PSO-EA algorithmrdquo Neurocomputing vol 70 no13ndash15 pp 2342ndash2353 2007

[9] F E H Tay and L Cao ldquoApplication of support vectormachinesin financial time series forecastingrdquo Omega vol 29 no 4 pp309ndash317 2001

[10] G Rubio H Pomares I Rojas and L J Herrera ldquoA heuristicmethod for parameter selection in LS-SVM application to timeseries predictionrdquo International Journal of Forecasting vol 27no 3 pp 725ndash739 2011

[11] V VapnikTheNature of Statical LearningTheory Springer NewYork NY USA 1995

[12] L Cao and F E H Tay ldquoFinancial forecasting using SupportVector Machinesrdquo Neural Computing and Applications vol 10no 2 pp 184ndash192 2001

[13] K-J Kim ldquoFinancial time series forecasting using supportvector machinesrdquo Neurocomputing vol 55 no 1-2 pp 307ndash3192003

[14] S Sun and D R Hardoon ldquoActive learning with extremelysparse labeled examplesrdquoNeurocomputing vol 73 no 16ndash18 pp2980ndash2988 2010

[15] V Ceperic G Gielen and A Baric ldquoRecurrent sparse supportvector regression machines trained by active learning in thetime-domainrdquo Expert Systems with Applications vol 39 no 12pp 10933ndash10942 2012

[16] S Parameswaran and K Q Weinberger ldquoLarge margin multi-task metric learningrdquo in Proceedings of the 24th Annual Con-ference on Neural Information Processing Systems (NIPS rsquo10)Vancouver Canada December 2010

[17] T Jebara ldquoMulti-task feature and kernel selection for SVMsrdquoin Proceedings of the 21st International Conference on MachineLearning pp 433ndash440 ACM July 2004

[18] Y Li S Gong and H Liddell ldquoSupport vector regression andclassification based multi-view face detection and recognitionrdquo

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 7: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

Mathematical Problems in Engineering 7

in Proceedings of the 4th IEEE International Conference onAutomatic Face and Gesture Recognition pp 300ndash305 2000

[19] K Bennett and A Demiriz ldquoSemi-supervised support vectormachinesrdquo Advances in Neural Information Processing Systemsvol 11 pp 368ndash374 1999

[20] C Brouard F DrsquoAlche-Buc and M Szafranski ldquoSemi-supervised penalized output kernel regression for linkpredictionrdquo in Proceedings of the 28th International Conferenceon Machine Learning pp 593ndash600 July 2011

[21] K Veropoulos C Campbell and N Cristianini ldquoControllingthe sensitivity of support vector machinesrdquo in Proceedings of theInternational Joint Conference on Artificial Intelligence pp 55ndash60 Citeseer 1999

[22] O L Mangasarian and D R Musicant ldquoSuccessive overrelax-ation for support vectormachinesrdquo IEEETransactions onNeuralNetworks vol 10 no 5 pp 1032ndash1037 1999

[23] V Vapnik Estimations of Dependences Based on Statistical DataSpringer Berlin Germany 1982

[24] L Kaufman ldquoSolving the quadratic programming problemarising in support vector classificationrdquo in Advances in KernelMethods pp 147ndash167 MIT Press 1999

[25] Y J Lee and O L Mangasarian ldquoRSVM reduced supportvector machinesrdquo in Proceedings of the First SIAM InternationalConference on Data Mining pp 5ndash7 Philadelphia Pa USA2001

[26] J C Platt ldquoUsing analytic QP and sparseness to speed trainingof support vector machinesrdquo in Proceedings of the Conference onAdvances in Neural Information Processing Systems pp 557ndash563Cambridge UK 1999

[27] E Osuna R Freund and F Girosi ldquoImproved training algo-rithm for support vector machinesrdquo in Proceedings of the 7thIEEEWorkshop on Neural Networks for Signal Processing (NNSPrsquo97) pp 276ndash285 September 1997

[28] F E H Tay and L J Cao ldquoImproved financial time seriesforecasting by combining support vector machines with self-organizing feature maprdquo Intelligent Data Analysis vol 5 no 4pp 339ndash354 2001

[29] F E H Tay and L J Cao ldquoModified support vector machines infinancial time series forecastingrdquo Neurocomputing vol 48 pp847ndash861 2002

[30] L Cao ldquoSupport vector machines experts for time seriesforecastingrdquo Neurocomputing vol 51 pp 321ndash339 2003

[31] K E Lee J W Lee and B H Hong ldquoComplex networks in astock marketrdquo Computer Physics Communications vol 177 no1-2 p 186 2007

[32] P Caraiani ldquoCharacterizing emerging European stock marketsthrough complex networks from local properties to self-similarcharacteristicsrdquo Physica A vol 391 no 13 pp 3629ndash3637 2012

[33] G Bonanno G Caldarelli F Lillo S Micciche N VandewalleandRNMantegna ldquoNetworks of equities in financialmarketsrdquoEuropean Physical Journal B vol 38 no 2 pp 363ndash371 2004

[34] S Abecasis E Lapenta and C Pedreira ldquoPerformance metricsfor financial time series forecastingrdquo Journal of ComputationalIntelligence in Finance vol 7 no 4 pp 5ndash22 1999

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of

Page 8: Research Article Financial Time Series Forecasting Using Directed … · 2019. 7. 31. · Financial Time Series Forecasting Using Directed-Weighted Chunking SVMs YongmingCai, 1 LeiSong,

Submit your manuscripts athttpwwwhindawicom

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical Problems in Engineering

Hindawi Publishing Corporationhttpwwwhindawicom

Differential EquationsInternational Journal of

Volume 2014

Applied MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Probability and StatisticsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Mathematical PhysicsAdvances in

Complex AnalysisJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

OptimizationJournal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

CombinatoricsHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Operations ResearchAdvances in

Journal of

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Function Spaces

Abstract and Applied AnalysisHindawi Publishing Corporationhttpwwwhindawicom Volume 2014

International Journal of Mathematics and Mathematical Sciences

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

The Scientific World JournalHindawi Publishing Corporation httpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Algebra

Discrete Dynamics in Nature and Society

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Decision SciencesAdvances in

Discrete MathematicsJournal of

Hindawi Publishing Corporationhttpwwwhindawicom

Volume 2014 Hindawi Publishing Corporationhttpwwwhindawicom Volume 2014

Stochastic AnalysisInternational Journal of


Recommended