A Study on Crude Oil Prices Modeled by Neurofuzzy Networks

22
Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions A Study on Crude Oil Prices Modeled by Neurofuzzy Networks (FUZZ-IEEE 2013) M. Panella 1 L. Liparulo 1 F. Barcellona 2 R. D’Ecclesia 3 1 Dept. of Information Engineering, Electronics and Telecommunications (DIET) 2 InfoSapienza Centre 3 Dept. of Methods and Models for Economics, Territory and Finance (MEMOTEF) University of Rome “La Sapienza” Email: [email protected] 2013 IEEE International Conference on Fuzzy Systems July 7-10, 2013, Hyderabad, India Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Transcript of A Study on Crude Oil Prices Modeled by Neurofuzzy Networks

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

A Study on Crude Oil Prices Modeledby Neurofuzzy Networks

(FUZZ-IEEE 2013)

M. Panella

1L. Liparulo

1F. Barcellona

2R. D’Ecclesia

3

1Dept. of Information Engineering, Electronics and Telecommunications (DIET)2InfoSapienza Centre

3Dept. of Methods and Models for Economics, Territory and Finance (MEMOTEF)

University of Rome “La Sapienza”Email: [email protected]

2013 IEEE International Conference on Fuzzy SystemsJuly 7-10, 2013, Hyderabad, India

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Outline

1 Introduction

2 Available methodologies

3 Datasets

4 Experimental setup

5 Numerical results

6 Conclusions

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Energy Commodities

+ )

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Motivation

To provide a powerful tool to replicate the dynamics of economic andfinancial series.

To apply a new methodology based on a neurofuzzy networks to forecastspecific standard price time series of energy commodities.

To backtest the results and set up risk management strategies.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Available methodologies

Quantitative techniques:

Explanatory methods

Time series based method

+

Forecasting process like a black box

Data-driven approaches to analize time series:

Autoregressive models (AR)

Moving average models (MA)

ARMA and ARIMA models

ARCH and GARCH models

Neural and Neurofuzzy networks

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

The proposed methodology for prediction (1/3)

We worked with log-prices yt = ln(St) and we considered a standard additivemodel for time series:

yt = µt + "t,

where µt is the deterministic component representing the forecast and "t is arandom variable which takes into account the uncertainty of prediction.

ARMA regression model8<

:µt = f

⇣X(µ)

t ;!(µ)t

X(µ)t =

⇥yt�1 yt�2 . . . yt�R "t�1 "t�2 . . . "t�M

where !

(µ)t is equal to the time varying parameter vectors of the regression

function f .

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

The proposed methodology for prediction (2/3)

ANFIS (Adaptive Neuro-Fuzzy Inference System) Network Architecture

Fuzzy decision: y(x) =PM

k=1 ↵(k)

(x)y

(k)(x)

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

The proposed methodology for prediction (3/3)

Our aim is to prove that ANFIS networks can solve the general regression problemby means of a set of P rules of Sugeno first-order type.

The kth rule, k = 1. . .P , has the following form:

(if yt�1 is B

(k)1 and . . . yt�R is B

(k)R and

"t�1 is B

(k)R+1 and . . . "t�M is B

(k)R+M then

µ

(k)t =

RX

j=1

a

(k)j yt�j +

MX

j=1

a

(k)R+j"t�j + a

(k)0

The structure of the fuzzy inference system is the following one:

µt = f

⇣X(µ)

t

⌘=

PPk=1 µ

(k)t VB(k)

⇣X(µ)

t

PPk=1 VB(k)

⇣X(µ)

t

⌘.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Datasets (1/2)

We used loarithmic transformed prices of both American and European energycommodities (Crude Oil):

US ! WTI (West Texas Intermediate)

EUR ! BRENT

Input-Output series

Input series (Training phase):8 two-years period from 2001� 2002 to 2008� 2009;

Output series (Testing phase):8 years from 2003 to 2010 (each follows the previous two years)

+

DATASETS

Unique time series for both phases (training and testing), described by data ofthree-year period (from 2001� 2002� 2003 to 2008� 2009� 2010)

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Datasets (2/2)

WTI daily log-prices from 2000 to 2012

Brent daily log-prices from 2000 to 2012

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Experimental setup (1/3)

Four data-driven modeling techniques are compared in this study:1

LSE (Least-Squares Estimation);2

RBF (Radial Basis Function neural networks), trained using the MatlabTM

software (version R2012a);3

MoG (Mixture of Gaussians neural networks), which are well suited tocomplex and non-convex data;

4ANFIS networks, where the Subtractive Clustering (SUBCL) method is usedfor the rule extraction and the rule parameters are obtained by means of astandard least-squares method coupled with the back-propagationoptimization.

Setup

M = 1 and R = 3 (Rule-of-thumb!)

Training window, NT = 500

Test window, NS = 250

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Experimental setup (2/3)

The prediction accuracy is measured by:

Mean Squared Error - MSE

MSE =

1

Ns

X

t

(yt � µt)2

Normalised Mean Square Error - NMSE

NMSE =

Pt (yt � µt)

2

Pt (yt � y)

2

Noise-to-signal ratio or NSR (dB)

) NSRdB = 10 log10

Pt(yt � µt)

2

Pt y

2

Mean absolute percentage error - MAPE

MAPE =

100

Ns

X

t

����yt � µt

yt

����

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Experimental setup (3/3)

In the testing phase of theoretical models, it is essential to define some criteria toassess the e�cacy of the chosen model.

In addition to the prediction error estimation, we also evaluated theSTATISTICAL FEATURES of the two series (Actual Vs Forecast)

+

Summarized in the four moments:Mean

Variance

Skewness

Kurtosis

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Results analysis 1/5 - 2003

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Results analysis 2/5 - 2009

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Results analysis 3/5 - Noise-to-Signal Ratio

Year Model WTI (NSR) Brent (NSR)

2003

ANFIS -41.32 -41.57

RBF -35.90 -41.00

MoG -20.57 -36.81

LSE -41.53 -42.00

2004

ANFIS -41.92 -24.16

RBF -35.27 -21.16

MoG -36.25 -19.37

LSE -32.89 -22.15

2005

ANFIS -44.28 -44.52

RBF -29.79 -24.11

MoG -28.22 -14.19

LSE -34.86 -44.67

2006

ANFIS -43.98 -45.85

RBF -35.42 -43.31

MoG -41.02 -20.24

LSE -36.51 -36.56

Year Model WTI (NSR) Brent (NSR)

2007

ANFIS -44.95 -46.02

RBF -35.09 -38.70

MoG -39.68 -42.32

LSE -35.71 -36.71

2008

ANFIS -35.56 -40.18

RBF -24.88 -14.80

MoG -23.94 -32.54

LSE -30.16 -30.55

2009

ANFIS -39.23 -40.51

RBF -33.63 -35.77

MoG -34.81 -36.70

LSE -30.03 -31.58

2010

ANFIS -46.65 -45.78

RBF -46.28 -47.12

MoG -42.93 -43.14

LSE -46.15 -36.97

The ANFIS model provides the best results in terms of NSR and this is ahomogeneous behavior across many years.The prediction performances of ANFIS are essentially better than the linearmodel LSE, which in turn behaves very similarly to the well-known naivepredictor for which yt = yt�1.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Results analysis 4/5 - From 2003 to 2006

Year StatisticWTI Brent

Estim. Actual Estim. Actual

2003

Mean 2.9669 2.9718 2.7643 2.7645

Variance 0.0067 0.0069 0.0088 0.0093

Kurtosis 3.1024 3.1255 2.3691 2.4811

Skewness 0.2533 0.2596 0.5063 0.5547

2004

Mean 3.2599 3.2599 2.8770 2.9508

Variance 0.0203 0.0183 0.0167 0.0219

Kurtosis 2.1745 2.1626 1.7541 2.2098

Skewness 0.3358 0.3369 -0.2293 0.0435

2005

Mean 3.5692 3.5727 3.3251 3.3297

Variance 0.0130 0.0125 0.0208 0.0203

Kurtosis 2.2378 2.2657 2.2683 2.2107

Skewness -0.2982 -0.3218 -0.5999 -0.5857

2006

Mean 3.7224 3.7314 3.4974 3.5018

Variance 0.0051 0.0070 0.0056 0.0068

Kurtosis 1.8276 1.7108 1.8171 1.8207

Skewness -0.0816 0.0545 -0.0973 -0.0988

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Results analysis 5/5 - From 2007 to 2010

Year StatisticWTI Brent

Estim. Actual Estim. Actual

2007

Mean 3.7974 3.8043 3.4994 3.5001

Variance 0.0278 0.0293 0.0138 0.0147

Kurtosis 2.1298 2.0256 2.4368 2.3630

Skewness 0.1930 0.2381 -0.1250 -0.1392

2008

Mean 4.0778 4.0999 3.6934 3.6952

Variance 0.0853 0.1137 0.0758 0.0815

Kurtosis 3.7970 3.7757 3.9390 3.8541

Skewness -1.3366 -1.2554 -1.2386 -1.2318

2009

Mean 3.6492 3.6477 3.3228 3.3277

Variance 0.0546 0.0556 0.0275 0.0251

Kurtosis 2.1135 2.3328 1.8815 1.8829

Skewness -0.6866 -0.7792 -0.5744 -0.5590

2010

Mean 3.8943 3.8958 3.5185 3.5285

Variance 0.0018 0.0019 0.0009 0.0010

Kurtosis 1.6653 1.6095 1.9534 1.8746

Skewness 0.0169 0.0614 0.2094 0.1035

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Conclusions & Future works

The approach proposed in this paper provides a highly predictive tool forCrude Oil prices, over a one year time horizon.

The quality of the results shows that neurofuzzy networks are extremelyuseful to describe time series with a complex dynamic.

Currently, we are experimenting this approach for the prediction of the returnseries.

We are further investigating more advanced techniques for a more reliableselection of the length of the training period, of the prediction order as wellas of the resultant complexity of the neurofuzzy models.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Thank you!

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Proposed algorithm (1/2)

Initialization. Let t = Ts be the first sample to be predicted and find theinitial condition for "k, k = (Ts �NT ) . . . (Ts � 1). We used in this regard anARMA model applied to the samples yk, k = (Ts �NT ) . . . (Ts � 1).

Step 1. At the current value of t, determine the training set to be used forthe model learning. It consists of a matrix where each row is an input-outputpattern that can be used for learning. In fact, the first R+M columns

represent the inputs X(µ)t and the last column is the expected value to be

estimated in correspondence with every pattern. The last row of the trainingset holds the most recent observation.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India

Introduction Available methodologies Datasets Experimental setup Numerical results Conclusions

Proposed algorithm (2/2)

Step 2. Determine, at the current time t, the parameters !(µ)t of the

regression function f , that is the ANFIS model or any other one, by using thetraining set and an appropriate learning algorithm according to the chosenregression model.

Step 3. By means of the parameters !(µ)t determined in the previous Step 2,

apply the ARMA model to forecast the conditional mean µt. Then, let"t = yt � µt, t � t+ 1, and go back to Step 1 if ts +NS , being NS thenumber of samples to be forecast.

Once the iteration is stopped, we have NS samples of conditional means(forecasts) and innovations pertaining to the time interval where prediction iscarried out.

Luca Liparulo FUZZ IEEE 2013, July 10, 2013 - Hyderabad, India