Rainfall Prediction

Embed Size (px)

Citation preview

  • 8/2/2019 Rainfall Prediction

    1/4

    A Novel Nonlinear Combination Model Based on Support Vector Machine for

    Rainfall Prediction

    Kesheng LuDepartment of Mathematics and Computer Sciences

    Guangxi Normal University for Nationality

    Chongzui, Guangxi, China

    Email: [email protected]

    Lingzhi WangDepartment of Mathematics and Computer Science

    Liuzhou Teachers College

    Liuzhou, Guangxi, China

    Email: [email protected]

    AbstractIn this study, a novel modulartype Support Vec-tor Machine (SVM) is presented to simulate rainfall prediction.First of all, a bagging sampling technique is used to generatedifferent training sets. Secondly, different kernel function ofSVM with different parameters, i.e., base models, are then

    trained to formulate different regression based on the dif-ferent training sets. Thirdly, the Partial Least Square (PLS)technology is used to select choose the appropriate numberof SVR combination members. Finally, a SVM can beproduced by learning from all base models. The technique willbe implemented to forecast monthly rainfall in the Guangxi,China. Empirical results show that the prediction by usingthe SVM combination model is generally better than thoseobtained using other models presented in this study in termsof the same evaluation measurements. Our findings reveal thatthe nonlinear ensemble model proposed here can be used asan alternative forecasting tool for a Meteorological applica-tion in achieving greater forecasting accuracy and improvingprediction quality further.

    Keywords-support vector machine; kernel function; partial

    least square; rainfall prediction;

    I. INTRODUCTION

    Rainfall forecasting has been a difficult subject in hy-

    drology due to the complexity of the physical processes

    involved and the variability of rainfall in space and time [1],

    [2]. With the development of science and technology, in

    particular, the intelligent computing technology in the past

    few decades, many emerging techniques, such as artificial

    neural network (ANN), have been widely used in the rainfall

    forecasting and obtained good results [3], [4], [5]. ANN

    are computerized intelligence systems that simulate the

    inductive power and behavior of the human brain. Theyhave the ability to generalize and see through noise and

    distortion, to abstract essential characteristics in the presence

    of irrelevant data, and to provide a high degree of robustness

    and fault tolerance [6], [7].

    Many experimental results demonstrate that the rainfall

    forecasting of ANN model outperformed multiple regres-

    sion, moving average and exponent smoothing from the

    research literature. In addition, ANN approaches want of a

    strict theoretical support, effects of applications are strongly

    depended upon operators experience. In the practical appli-

    cation, ANN often exhibits inconsistent and unpredictable

    performance on noisy data [8].

    Recently, support vector regression (SVM), a novel neural

    network algorithm, was developed by Vapnik and his col-leagues [9], which is a learning machine based on statistical

    learning theory, and which adheres to the principle of

    structural risk minimization seeking to minimize an upper

    bound of the generalization error, rather than minimize the

    training error (the principle followed by ANN) [10], [11].

    When using SVM, the main problems is confronted: how

    to choose the kernel function and how to set the best kernel

    paraments. The proper parameters setting can improve the

    SVM regression accuracy. Different kernel function and dif-

    ferent parameter settings can cause significant differences in

    performance. Unfortunately, there are no analytical methods

    or strong heuristics that can guide the user in selecting an

    appropriate kernel function and good parameter values.In order to overcome these drawbacks, a novel technique

    is introduced. The generic idea consists of three phases.

    First, an initial data set is transformed into several different

    training sets. Based on the different training sets, different

    kernel function of SVM and different parameter settings are

    then trained to formulate different regression forecasting.

    Finally, a SVM can be produced by learning from all base

    models. The rainfall data of Guangxi is predicted as a

    case study for development of rainfall forecasting model.

    The rest of this study is organized as follows. Section 2

    elaborates a triple-phase SVM process is described in detail.

    For further illustration, this work employs the method to set

    up a prediction model for rainfall forecasting in Section 3.Finally, some concluding remarks are drawn in Section 4.

    I I . THE BUILDING PROCESS OF THE NONLINEAR

    ENSEMBLE MODEL

    Originally, SVM has been presented to solve pattern

    recognition problems. However, with the introduction of

    Vapniks -insensitive loss function, SVM has been de-

    veloped to solve nonlinear regression estimation problems,

    such as new techniques known as support vector regression

    2011 Fourth International Joint Conference on Computational Sciences and Optimization

    978-0-7695-4335-2/11 $26.00 2011 IEEE

    DOI 10.1109/CSO.2011.50

    1343

  • 8/2/2019 Rainfall Prediction

    2/4

    (SVR) [12], which have been shown to exhibit excellent

    performance. At present, SVR has been emerging as an

    alternative and powerful technique to solve the nonlinear

    regression problem. It has achieved great success in both

    academic and industrial platforms due to its many attractive

    features and promising generalization performance.

    A. Support Vector Regression

    The SVR model maps data nonlinearly into a higher-

    dimensional feature space, in which it undertakes linear

    regression. Rather than obtaining empirical errors, SVR

    aims to minimize the upper limit of the generalization

    error. Suppose we are given training data (x, )

    =1, where

    is the input vector; is the output value and is

    the total number of data dimension. The modelling aim is

    to identify a regression function, = (), that accuratelypredicts the outputs corresponding to a new set of input

    output examples, (, ). The linear regression function (inthe feature space) is described as follows:

    () =() + ,

    : , (1)

    where and are coefficients; () denotes the high dimen-sional feature space, which is nonlinearly mapped from the

    input space x. This primal optimization problem is a linearly

    constrained quadratic programming problem [13], which

    can be solved by introducing Lagrangian multipliers and

    applying Karush-Kuhn-Tucker (KKT) conditions to solve its

    dual problem:

    (, ) ==1

    ( ) =1

    ( + )

    12

    =1

    =1

    ( )(

    )(, )

    ..=1

    ( ) = 0

    0 , , = 1, 2, , (2)

    where and

    are the Lagrangian multipliers associated

    with the constraints, the term (, ) is defined as kernelfunction, where the value of kernel function equals the inner

    product of two vectors and in the feature space ()and (), meaning that (, ) = () (). Inmachine learning theories, the popular kernel functions are

    Linear kernel, Polynomial kernel and Guassian kernel.

    B. Generating individual SVR predictors

    With the work about biasvariance tradeoff of Breti-

    man [14], an ensemble model regression model consisting

    of diverse models with much disagreement is more likely

    to have a good generalization [15]. Therefore, how to

    generate diverse models is a crucial factor. For SVR model,

    several methods have been investigated for the generation of

    ensemble members making different errors. Such methods

    basically depended on different the kernel function, varying

    the parameters of SVR or utilizing different training sets.

    In this paper, there are three methods for generating diverse

    models.

    (1) Using different the type of SVR kernel function, such

    as the linear kernel function and the polynomial kernel

    function.(2) Utilizing different the parameters of SVR, such as

    different cluster center of the SVR, : through varying the

    cluster center of the SCR, different cluster radius of the

    SVR, different SVR can be produced.

    (3) Using different training data: by re-sampling and

    preprocessing data, different training sets can be obtained. .

    C. Selecting appropriate ensemble members

    After training, each individual neural predictor has gener-

    ated its own result. However, if there are a great number of

    individual members, we need to select a subset of represen-

    tatives in order to improve ensemble efficiency. In this paper,

    the Partial Least Square (PLS) regression technique [16] isadopted to select appropriate ensemble members. Interested

    readers can be referred to [16] for more details.

    D. -Support vector regression

    If the proper hyper parameters are picked up, SVR will

    gain good generalization performance and vice versa, so it

    is important to select right model. Instead of selecting an

    appropriate Schokopf et al. proposed a variant, called -

    support vector regression, which introduces a new parameter

    which can control the number of support vectors and

    training errors without defining a prior. To be more precise,

    they proved that is an upper bound on the fraction of

    margin errors and lower bound of the fraction of supportvectors [15].

    The -SVR regression in data sets can be described as

    follows:

    ( , ,) = 12

    + ( + 1

    =1

    ( + ))

    .. () + () + ,

    0, = 1, 2, , , 0.(3)

    where 0 1, is the regulator, and training data are mapped into a high (even infinite) dimensional feature

    space by the mapping function ().

    E. The Establishment of Combination Forecasting Model

    To summarize, the proposed nonlinear combination fore-

    casting model consists of four main stages. Generally speak-

    ing, in the first stage, the initial data set is divided into

    different training sets by used Bagging and Boosting tech-

    nology. In the second stage, these training sets are input

    to the different individual SVM regression models, and

    then various single SVM regression predictors are produced

    1344

  • 8/2/2019 Rainfall Prediction

    3/4

    based on diversity principle. In the third stage, PLS model

    is used to select choose the appropriate number of SVR

    ensemble members. In the four stage, -SVM regression

    is used to aggregate the selected combination members ( -

    SVR). In such a way final combination forecasting results

    can be obtained. The basic flow diagram can be shown in

    Fig.1.

    Original DataSet, DS

    v-SVR

    Ensemble

    Bagging

    Technology

    Training Se tTR 1

    SVR 1

    Output

    PLS Selectiom

    Training Se tTR 2

    Training Se tTR M-1

    Training Se tTR M

    SVR 3

    Output

    SVR 5

    Output

    SVR 6

    Output

    SVR 2

    Output

    SVR 4

    Output

    Figure 1. A flow diagram of the proposed semiparamentric ensembleforecasting model.

    III. EXPERIMENTAL RESULTS AND DISCUSSION

    A. Empirical Data

    This study has investigated Modeling -SVM regression

    to predict average monthly precipitation from January 1965

    to December 2009 in Guangxi. Thus the data set contained

    540 data points in time series, 500 data of whose were usedto train samples for -SVM regression learning, and the

    other 40 data were used to test sample for -SVM regression

    Generalization ability.

    Method of modeling is one-step ahead prediction, that is,

    the forecast is only one sample each time and the training

    samples is an additional one each time on the base of the

    previous training.

    B. Performance evaluation of model

    In order to measure the effectiveness of the proposed

    method, three types of errors are used in this paper, such

    as, Normalized Mean Squared Error (NMSE), the Mean

    Absolute Percentage Error (MAPE) and Pearson RelativeCoefficient (PRC), which be found in many paper [6].

    In order to investigate the effect of the proposed model,

    the simple averaging ensemble, the mean squared er-

    ror (MSE) based regression ensemble and variancebased

    weighted ensemble are established. Those are fitted the 500

    samples and forecasted the 40 samples by the those models,

    the comparison results are used to test the effect of predictive

    models.

    C. Analysis of the Results

    Table 1 illustrates the fitting accuracy and efficiency of

    the model in terms of various evaluation indices for 500

    training samples. From the table, we can generally see that

    learning ability of vSVM regression ensemble outperforms

    the other three models under the same network input. The

    more important factor to measure performance of a methodis to check its forecasting ability of testing samples in order

    for actual rainfall application.

    Table IA COMPARISON OF FITTING RESULT OF FOUR DIFFERENT MODELS

    ABOUT 50 0 TRAINING SAMPLES

    Ensemble Moel NMSE MAPE PRC

    simple averaging 0.0976 0.7360 0.7654MSE ensemble 0.1029 0.3456 0.7892variance weight ed 0.0452 0 .3211 0.9350v-SVM regression 0.0374 0.2486 0.9766

    Figure 3 shows the forecasting results of four differentmodels for 40 testing samples, we can see that the fore-

    casting results of vSVR ensemble model are best in all

    models. Table 2 shows that the forecasting performance of

    four different models from different perspectives in terms of

    various evaluation indices. From the graphs and table, we

    can generally see that the forecasting results are very promis-

    ing in the rainfall forecasting under the research where either

    the measurement of fitting performance is goodness or where

    the forecasting performance is effectiveness.

    5 10 15 20 25 30 35 400

    50

    100

    150

    200

    250

    300

    350

    400

    450

    500

    550

    Monthly

    Rainfall(mm

    )

    Actual monthly rainfallSimple averagingMSE regressionVariance based weightvSVR combination

    Figure 2. Testing results in June for 30 testing samples.

    As shown in Table 2 about the rainfall forecasting of fourdifferent model, the differences among the different models

    are very significant. For example, the NMSE of the simple

    averaging ensemble model is 0.1285. Similarly, the NMSE

    of the MSE ensemble model is 0.0955, the NMSE of the

    variance weighted ensemble model is 0.0653; however the

    NMSE of the vSVM regression model reaches 0.0221.

    The NMSE result of the vSVM regression model has

    obvious advantages over three other models. Subsequently,

    1345

  • 8/2/2019 Rainfall Prediction

    4/4

    Table IIA COMPARISON OF FORECASTING RESULT OF FOUR DIFFERENT

    MODELS ABOUT 40 TESTING SAMPLES

    Ensemble Moel NMSE MAPE PRC

    simple averaging 0.1285 0.8710 0.6726MSE ensemble 0.0955 0.4381 0.7965var iance weighted 0.0 653 0.410 9 0.8 820vSVM r egressio n 0.0 221 0.205 3 0.9 341

    for MAPE efficiency index, the proposed vSVM regression

    model is also the smallest.

    IV. CONCLUSION

    Accurate rainfall forecasting is crucial for a frequent unan-

    ticipated flash flood region to avoid life losing and economic

    loses. This paper proposes a novel nonlinear combination

    forecasting method in terms of vSVR principle. This model

    was applied to the forecasting fields of monthly rainfall

    in Guangxi. In terms of the different forecasting models,empirical results show that the developed model performs

    the best for monthly rainfall on the basis of different cri-

    teria. Our experimental results demonstrated the successful

    application of our proposed new model, vSVM regression,

    for the complex forecasting problem. It demonstrated that

    it increased the rainfall forecasting accuracy more than

    any other model employed in this study in terms of the

    same measurements. So the vSVM regression ensemble

    forecasting model can be used as an alternative tool for

    monthly rainfall forecasting to obtain greater forecasting

    accuracy.

    ACKNOWLEDGMENT

    The authors would like to express their sincere thanks

    to the editor and anonymous reviewers comments and sug-

    gestions for the improvement of this paper. This work was

    supported in part by Guangxi Natural Science Foundation

    under Grant No. 0832092, and in part by the Department of

    Guangxi Education under Grant No. 200707MS061.

    REFERENCES

    [1] Lihua Xiong, K. M. Connor, An empirical method to improvethe prediction limits of the GLUE methodology in rainfallrunoff modeling, Journal of Hydrology, Vol. 349, pp: 115124, 2008.

    [2] G.H. Schmitz, J. Cullmann, PAIOFF: A new proposal foronline flood forecasting in flash flood prone catchments,Journal of Hydrology, Vol. 360, pp: 114, 2008.

    [3] Jiansheng Wu, Liangyong Huang and Xiongming Pan, Anovel bayesian additive regression trees ensemble model basedon linear regression and nonlinear regression for torrentialrain forecasting, Proeedings of the Third Internatioal JointConference on Computational Sciences and Optimization, eds.K. K. Lai, Yingwen Song and Lean Yu, IEEE ComputerSociety Press, Vol. 2, pp:484487, 2010.

    [4] Jiansheng Wu, A novel nonparametric regression ensemblefor rainfall forecasting using particle swarm optimization tech-nique coupled with artificial neural network, Lecture NoteComputer Science, Vol. 5553, No. 3, pp: 4958, Springer-Verlag Berlin Heidelberg, 2009.

    [5] G. F. Lin and L. H. Chen, Application of an artificial neuralnetwork to typhoon rainfall forecasting, Hydrological Pro-

    cesses, Vol. 19, pp. 18251837, 2005.

    [6] Jiansheng Wu and Long Jin, Study on the meteorological pre-diction model using the learning algorithm of neural networkbased on pso algorithms, Journal of Tropical Meteorology,Vol. 16, No. 1, pp: 8388, 2009.

    [7] R. S. Govindaraju, Artificial neural network in hydrology,I: Preliminary concepts, Journal of Hydrologic Engineering,Vol.5, No.2, 115123.

    [8] W. C. Hong, Rainfall forecasting by technological machinelearning models, Applied Mathematics and Computation, Vol.200, pp: 4157, 2008.

    [9] V. Vapnik, The nature of statistical learning theory. New York:Springer Press, 1995.

    [10] F. E. H. Tay and L. Cao, Modified support vector machinesin financial time series forecasting, Neurocomputing, vol.48(14), pp: 847861, 2002.

    [11] V. Vapnik, S. Golowich and A. Smola, Support vectormethod for function approximation, regression estimation andsignal processing, In Edited by M. Mozer, M. Jordan andT. Petsche, Advance in neural information processing system,Vol. 9, pp: 281287. Cambridge, MA: MIT Press, 1997.

    [12] B. Scholkopf, A. Smola, R. C. Williamson and P. L. Bartlett.New support vector algorithms, Neural Computation, Vol. 5,pp: 12071245, 2000.

    [13] V. Vapnik, S. Golowich and A. Smola, Support vectormethod for function approximation, regression estimation andsignal processing, In Edited by M. Mozer, M. Jordan andT. Petsche, Advance in neural information processing system,Vol. 9, pp: 281287. Cambridge, MA: MIT Press, 1997.

    [14] L. Breiman, Combining Predictors,Proceedings of Combin-ing Artificial Neural Nets-Ensemble and Modular Multi-netSystems, the Springer Press, Berlin, Vol. 1, pp. 31-50, 1999.

    [15] J. A. Benediktsson, J. R. Sveinsson, O. K. Ersoy and P. H.Swain, Parallel Consensual Neural Neural Networks, IEEETransactions on Neural Networks, Vol. 8, pp. 54-64, 1997.

    [16] D. M. Pirouz. An overview of partial least square, Technicalreport, The Paul Merage School of Business, University ofCalifornia, Irvine, 2006.

    1346