MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Sequential Optimizing Investing Strategy with Neural Networks



Similar documents
Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network

Forecasting the Direction and Strength of Stock Market Movement

What is Candidate Sampling

Recurrence. 1 Definitions and main statements

An Alternative Way to Measure Private Equity Performance

On the Optimal Control of a Cascade of Hydro-Electric Power Stations

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ).

Lecture 3: Force of Interest, Real Interest Rate, Annuity

A Genetic Programming Based Stock Price Predictor together with Mean-Variance Based Sell/Buy Actions

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic

How Sets of Coherent Probabilities May Serve as Models for Degrees of Incoherence

Using Series to Analyze Financial Situations: Present Value

Trade Adjustment and Productivity in Large Crises. Online Appendix May Appendix A: Derivation of Equations for Productivity

The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis

THE DISTRIBUTION OF LOAN PORTFOLIO VALUE * Oldrich Alfons Vasicek

Lecture 2: Single Layer Perceptrons Kevin Swingler

Financial Mathemetics

Time Delayed Independent Component Analysis for Data Quality Monitoring

7.5. Present Value of an Annuity. Investigate

A Novel Methodology of Working Capital Management for Large. Public Constructions by Using Fuzzy S-curve Regression

The Application of Fractional Brownian Motion in Option Pricing

An Evaluation of the Extended Logistic, Simple Logistic, and Gompertz Models for Forecasting Short Lifecycle Products and Services

1 Example 1: Axis-aligned rectangles

Prediction of Disability Frequencies in Life Insurance

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

APPLICATION OF PROBE DATA COLLECTED VIA INFRARED BEACONS TO TRAFFIC MANEGEMENT

Number of Levels Cumulative Annual operating Income per year construction costs costs ($) ($) ($) 1 600,000 35, , ,200,000 60, ,000

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting

Intra-day Trading of the FTSE-100 Futures Contract Using Neural Networks With Wavelet Encodings

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College

Course outline. Financial Time Series Analysis. Overview. Data analysis. Predictive signal. Trading strategy

Robust Design of Public Storage Warehouses. Yeming (Yale) Gong EMLYON Business School

Laddered Multilevel DC/AC Inverters used in Solar Panel Energy Systems

8.5 UNITARY AND HERMITIAN MATRICES. The conjugate transpose of a complex matrix A, denoted by A*, is given by

THE APPLICATION OF DATA MINING TECHNIQUES AND MULTIPLE CLASSIFIERS TO MARKETING DECISION

BERNSTEIN POLYNOMIALS

How To Understand The Results Of The German Meris Cloud And Water Vapour Product

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm

A Trading System for FTSE-100 Futures Using Neural Networks and Wavelets

Using Association Rule Mining: Stock Market Events Prediction from Financial News

NON-CONSTANT SUM RED-AND-BLACK GAMES WITH BET-DEPENDENT WIN PROBABILITY FUNCTION LAURA PONTIGGIA, University of the Sciences in Philadelphia

Implied (risk neutral) probabilities, betting odds and prediction markets

Portfolio Loss Distribution

Addendum to: Importing Skill-Biased Technology

Linear Circuits Analysis. Superposition, Thevenin /Norton Equivalent circuits

Support Vector Machines

Can Auto Liability Insurance Purchases Signal Risk Attitude?

An Interest-Oriented Network Evolution Mechanism for Online Communities

Activity Scheduling for Cost-Time Investment Optimization in Project Management

Construction Rules for Morningstar Canada Target Dividend Index SM

Modelling of Web Domain Visits by Radial Basis Function Neural Networks and Support Vector Machine Regression

) of the Cell class is created containing information about events associated with the cell. Events are added to the Cell instance

Chapter 4 ECONOMIC DISPATCH AND UNIT COMMITMENT

CHAPTER 14 MORE ABOUT REGRESSION

A COLLABORATIVE TRADING MODEL BY SUPPORT VECTOR REGRESSION AND TS FUZZY RULE FOR DAILY STOCK TURNING POINTS DETECTION

How To Calculate The Accountng Perod Of Nequalty

Binomial Link Functions. Lori Murray, Phil Munz

Application of an Improved BP Neural Network Model in Enterprise Network Security Forecasting

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification

The OC Curve of Attribute Acceptance Plans

1. Measuring association using correlation and regression

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION

Risk-based Fatigue Estimate of Deep Water Risers -- Course Project for EM388F: Fracture Mechanics, Spring 2008

Answer: A). There is a flatter IS curve in the high MPC economy. Original LM LM after increase in M. IS curve for low MPC economy

Luby s Alg. for Maximal Independent Sets using Pairwise Independence

PSYCHOLOGICAL RESEARCH (PYC 304-C) Lecture 12

Schedulability Bound of Weighted Round Robin Schedulers for Hard Real-Time Systems

The impact of hard discount control mechanism on the discount volatility of UK closed-end funds

8 Algorithm for Binary Searching in Trees

PERRON FROBENIUS THEOREM

IMPACT ANALYSIS OF A CELLULAR PHONE

Invoicing and Financial Forecasting of Time and Amount of Corresponding Cash Inflow

Calculation of Sampling Weights

Damage detection in composite laminates using coin-tap method

Brigid Mullany, Ph.D University of North Carolina, Charlotte

A Model of Private Equity Fund Compensation

Financial market forecasting using a two-step kernel learning method for the support vector regression

L10: Linear discriminants analysis

Abstract. 260 Business Intelligence Journal July IDENTIFICATION OF DEMAND THROUGH STATISTICAL DISTRIBUTION MODELING FOR IMPROVED DEMAND FORECASTING

DEFINING %COMPLETE IN MICROSOFT PROJECT

Statistical Approach for Offline Handwritten Signature Verification

Project Networks With Mixed-Time Constraints

When Network Effect Meets Congestion Effect: Leveraging Social Services for Wireless Services

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending

Transition Matrix Models of Consumer Credit Ratings

"Research Note" APPLICATION OF CHARGE SIMULATION METHOD TO ELECTRIC FIELD CALCULATION IN THE POWER CABLES *

Data Mining from the Information Systems: Performance Indicators at Masaryk University in Brno

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns

8.4. Annuities: Future Value. INVESTIGATE the Math Annuities: Future Value

PAS: A Packet Accounting System to Limit the Effects of DoS & DDoS. Debish Fesehaye & Klara Naherstedt University of Illinois-Urbana Champaign

SOLVING CARDINALITY CONSTRAINED PORTFOLIO OPTIMIZATION PROBLEM BY BINARY PARTICLE SWARM OPTIMIZATION ALGORITHM

NEURO-FUZZY INFERENCE SYSTEM FOR E-COMMERCE WEBSITE EVALUATION

Improved SVM in Cloud Computing Information Mining

An artificial Neural Network approach to monitor and diagnose multi-attribute quality control processes. S. T. A. Niaki*

Transcription:

MATHEMATICAL ENGINEERING TECHNICAL REPORTS Sequental Optmzng Investng Strategy wth Neural Networks Ryo ADACHI and Akmch TAKEMURA METR 2010 03 February 2010 DEPARTMENT OF MATHEMATICAL INFORMATICS GRADUATE SCHOOL OF INFORMATION SCIENCE AND TECHNOLOGY THE UNIVERSITY OF TOKYO BUNKYO-KU, TOKYO 113-8656, JAPAN WWW page: http://www.kesu.t.u-tokyo.ac.jp/research/techrep/ndex.html

The METR techncal reports are publshed as a means to ensure tmely dssemnaton of scholarly and techncal work on a non-commercal bass. Copyrght and all rghts theren are mantaned by the authors or by other copyrght holders, notwthstandng that they have offered ther works here electroncally. It s understood that all persons copyng ths nformaton wll adhere to the terms and constrants nvoked by each author s copyrght. These works may not be reposted wthout the explct permsson of the copyrght holder.

Sequental optmzng nvestng strategy wth neural networks Ryo Adach Akmch Takemura February 2010 Abstract In ths paper we propose an nvestng strategy based on neural network models combned wth deas from game-theoretc probablty of Shafer and Vovk. Our proposed strategy uses parameter values of a neural network wth the best performance untl the prevous round (tradng day) for decdng the nvestment n the current round. We compare performance of our proposed strategy wth varous strateges ncludng a strategy based on supervsed neural network models and show that our procedure s compettve wth other strateges. 1 Introducton A number of researches have been conducted on predcton of fnancal tme seres wth neural networks snce Rumelhart [10] developed back propagaton algorthm n 1986, whch s the most commonly used algorthm for supervsed neural network. Wth ths algorthm the network learns ts nternal structure by updatng the parameter values when we gve t tranng data contanng nputs and outputs. We can then use the network wth updated parameters to predct future events contanng nputs the network has never encountered. The algorthm s appled n many felds such as robotcs and mage processng and t shows a good performance n predcton of fnancal tme seres. Graduate School of Informaton Scence and Technology, Unversty of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, JAPAN Insttute of Industral Scence, Unversty of Tokyo, 4-6-1 Komaba, Meguro-ku, Tokyo 153-8505, JAPAN 1

Relevant papers on the use of neural network to fnancal tme seres nclude [5], [6], [8] and [14]. In these papers authors are concerned wth the predcton of tme seres and they to not pay much attenton to actual nvestng strateges, although the predcton s obvously mportant n desgnng practcal nvestng strateges. A forecast of tomorrow s prce does not mmedately tell us how much to nvest today. In contrast to these works, n ths paper we drectly consder nvestng strateges for fnancal tme seres based on neural network models and deas from game-theoretc probablty of Shafer and Vovk (2001) [11]. In the game-theoretc probablty establshed by Shafer and Vovk, varous theorems of probablty theory, such as the strong law of large numbers and the central lmt theorem, are proved by consderaton of captal processes of bettng strateges n varous games such as the con-tossng game and the bounded forecastng game. In game-theoretc probablty a player Investor s regarded as playng aganst another player Market. In ths framework nvestng strateges of Investor play a promnent role. Predcton s then derved based on strong nvestng strateges (cf. defensve forecastng n [13]). Recently n [9] we proposed sequental optmzaton of parameter values of a smple nvestng strategy n mult-dmensonal bounded forecastng games and showed that the resultng strategy s easy to mplement and shows a good performance n comparson to well-known strateges such as the unversal portfolo [3] developed by Thomas Cover and hs collaborators. In ths paper we propose sequental optmzaton of parameter values of nvestng strateges based on neural networks. Neural network models gve a very flexble framework for desgnng nvestng strateges. Wth smulaton and wth some data from Tokyo Stock Exchange we show that the proposed strategy shows a good performance. The organzaton of ths paper s as follows. In Secton 2 we propose sequental optmzng strategy wth neural networks. In Secton 3 we present some alternatve strateges for the purpose of comparson. In Secton 3.1 we consder an nvestng strategy usng supervsed neural network wth back propagaton algorthm. The strategy s closely related to and reflects exstng researches on stock prce predcton wth neural networks. In Secton 3.2 we consder Markovan proportonal bettng strateges, whch are much smpler than the strateges based on neural networks. In Secton 4 we evaluate performances of these strateges by Monte Carlo smulaton. In Secton 5 we apply these strateges to stock prce data from Tokyo Stock Exchange. Fnally we gve some concludng remarks n Secton 6. 2

2 Sequental optmzng strategy wth neural networks Here we ntroduce the bounded forecastng game of Shafer and Vovk [11] n Secton 2.1 and network models we use n Secton 2.2. In Secton 2.3 we specfy the nvestng rato by an unsupervsed neural network and we propose sequental optmzaton of parameter values of the network. 2.1 Bounded forecastng game We present the bounded forecastng game formulated by Shafer and Vovk n 2001 [11]. In the bounded forecastng game, Investor s captal at the end of round n s wrtten as K n (n = 1, 2,...) and ntal captal K 0 s set to be 1. In each round Investor frst announces the amount of money M n he bets ( M n < K n 1 ) and then Market announces her move x n [ 1, 1]. x n represents the change of the prce of a unt fnancal asset n round n. The bounded forecastng game can be consdered as an extenson of the classcal con-tossng game snce the bounded forecastng game results n the classcal con-tossng game f x n { 1, 1}. Wth K n, M n and x n, Investor s captal after round n s wrtten as K n = K n 1 + M n x n. The protocol of the bounded forecastng game s wrtten as follows. Protocol: K 0 =1. FOR n = 1, 2,... : Investor announces M n R. Market announces x n [ 1, 1]. K n = K n 1 + M n x n END FOR We can rewrte Investor s captal as K n = K n 1 (1 + α n x n ), where α n = M n /K n 1 s the rato of Investor s nvestment M n to hs captal K n 1 after round n 1. We call α n the nvestng rato at round n. We restrct α n as 1 < α n < 1 n order to prevent Investor becomng bankrupt. Furthermore we can wrte K n as K n = K n 1 (1 + α n x n ) = = Π n (1 + α k x k ). Takng the logarthm of K n we have log K n = n log(1 + α k x k ). (1) 3

The behavor of Investor s captal n (1) depends on the choce of α k. Specfyng a functonal form of α k s regarded as an nvestng strategy. For example, settng α k ɛ to be a constant ɛ for all k s called the ɛ-strategy whch s presented n [11]. In ths paper we consder varous ways to determne α k n terms of past values x k 1, x k 2,..., of x and and seek better α k n tryng to maxmze the future captal K n, n > k. Let u k 1 = (x k 1,..., x k L ) denote past L values of x and let α k depend on u k 1 and a parameter ω: α k = f(u k 1, ω). Then ω k 1 = argmax k 1 log(1 + f(u t 1, ω)x t ) t=1 s the best parameter value untl the prevous round. In our sequental optmzng nvestng strategy, we use ω n 1 to determne the nvestment M n at round n: M n = K n 1 f(u n 1, ω n 1). For the functon f we employ neural network models for ther flexblty, whch we descrbe n the next secton. 2.2 Desgn of the network We construct a three-layered neural network shown n Fgure 1. The nput layer has L neurons and they just dstrbute the nput u j (j = 1,..., L) to every neuron n the hdden layer. Also the hdden layer has M neurons and we wrte the nput to each neurons as I 2 whch s a weghted sum of u j s. As seen from Fgure 1, I 2 s obtaned as I 2 = L j=1 ω 1,2 j u j, where ω 1,2 j s called the weght representng the synaptc connectvty between the jth neuron n the nput layer and the th neuron n the hdden layer. Then the output of the th neuron n the hdden layer s descrbed as y 2 = tanh(i 2 ). As for actvaton functon we employ hyperbolc tangent functon. In a smlar way, the nput to the neuron n the output layer, whch we wrte I 3, s obtaned as I 3 = M =1 4 ω 2,3 y 2,

Fgure 1: Three-layered network where ω 1 s the weght between the th neuron n the hdden layer and the neuron n the output layer. Fnally, we have y 3 = tanh(i 3 ), whch s the output of the network. In the followng argument we use y 3 as an nvestment strategy. Thus we can wrte where α k = y 3 = f(u k 1, ω), ω = (ω 1,2, ω 2,3 ) = ( (ω 1,2 j ) =1,...,M, j=1,...,l, (ω 2,3 ) =1,...,M ). Investor s captal s wrtten as K n = K n 1 (1 + f(u n 1, ω)x n ). We need to specfy the number of nputs L and the number of neurons M n the hdden layer. It s dffcult to specfy them n advance. We compare varous choces of L and M n Secton 4 and Secton 5. Also n u n 1 we can nclude any nput whch s avalable before the start of round n, such as movng averages of past prces, seasonal ndcators or past values of other economc tme seres data. We gve further dscusson on the choce of u n 1 n Secton 6. 5

2.3 Sequental optmzng strategy wth neural networks In ths secton we propose a strategy whch we call Sequental Optmzng Strategy wth Neural Networks (SOSNN). We frst calculate ω = ω n 1 that maxmzes n 1 φ = log(1 + f(u k 1, ω)x k ). (2) Ths s the best parameter values untl the prevous round. If Investor uses α n = f(u n 1, ω n 1) as the nvestng rato, Investor s captal after round n s wrtten as K n = K n 1 (1 + f(u n 1, ω n 1)x n ). For maxmzaton of (2), we employ the gradent descent method. Wth ths method, the weght updatng algorthm of ω 2,3 wth the parameter β (called the learnng constant) s wrtten as where ω 2,3 = ω 2,3 + ω 2,3 = ω 2,3 + β φ ω 2,3, φ ω 2,3 = φ f = = n 1 n 1 f ω 2,3 = n 1 φ f f k I 3 k I 3 ω 2,3 x k 1 + f(u k 1, ω)x k (1 tanh 2 ( k I 3 )) k y 2 k δ k 1 y 2, and the left superscrpt k to I 3, y 2 ndexes the round. Thus we obtan ω 2,3 = β φ ω 2,3 = β n 1 k δ 1 k y 2. Smlarly, the weght updatng algorthm of ω 1,2 j s expressed as ω 1,2 j = ω 1,2 j + ω 1,2 j = ω 1,2 j 6 + β φ ω 1,2 j,

where φ ω 1,2 j Thus we obtan = = = n 1 n 1 n 1 φ k I 2 k I 2 ω 1,2 j = n 1 φ f f k I 3 k I 3 k y 2 k δ 1 ω 2,3 (1 tanh 2 ( k I 2 )) (u k 1 ) j k δ 2 (u k 1 ) j. n 1 k y 2 k I 2 ω 1,2 j = β φ ω 1,2 = β k δ 2 (u k 1 ) j. j Here we summarze the algorthm of SOSNN at round n. k I 2 ω 1,2 j 1. Gven the nput vector u k 1 = (x k 1,..., x k L ) (k = 1,..., n 1) and L the value of ω n 1, we frst evaluate k I 2 = ω 1,2 j (u k 1) j and then j=1 k y 2 = tanh( k I 2 ). Also we set the learnng constant β. 2. We calculate k I 3 = M =1 ω 2,3 k y 2 and then k y 3 = tanh( k I 3 ) wth k I 2 and k y 2 of the prevous step. Then we update weght ω wth the weght updatng formula ω 2,3 = ω 2,3 +β k δ k 1 y 2 and ω 1,2 j = ω 1,2 j +β k δ 2 (u k 1 ) j n 1 n 1. 3. Go back to step 1 replacng the weght ω n 1 wth updated values. After suffcent tmes of teraton, φ n (2) converges to a local maxmum wth respect to ω 1,2 j and ω 2,3 and we set ω 1,2 j = ω 1,2 j and ω 2,3 = ω 2,3, whch are elements of ωn 1. Then we evaluate Investor s captal after round n as K n = K n 1 (1 + f(u n 1, ωn 1)x n ). 3 Alternatve nvestment strateges Here we present some strateges that are desgned to be compared wth SOSNN. In Secton 3.1 we present a strategy wth back-propagatng neural network. The advantage of back-propagatng neural network s ts predctve ablty due to learnng as prevous researches show. In Secton 3.2 we show some sequental optmzng strateges that use rather smple functon for f than SOSNN does. 7

3.1 Optmzng strategy wth back propagaton In ths secton we consder a supervsed neural network and ts optmzaton by back propagaton. We call the strategy NNBP. It decdes the bettng rato by predctng actual up-and-downs of stock prces and can be regarded as ncorporatng exstng researches on stock prce predcton. Thus t s sutable as an alternatve to SOSNN. For supervsed network, we tran the network wth the data from a tranng perod, obtan the best value of the parameters for the tranng perod and then use t for the nvestng perod. These two perods are dstnct. For the tranng perod we need to specfy the desred output (target) T k of the network for each day k. We propose to specfy the target by the drecton of Market s current prce movement x k. Thus we set +1 x k > 0 T k = 0 x k = 0. 1 x k < 0 Note that ths T k s the best nvestng rato f Investor could use the current movement x k of Market for hs nvestment. Therefore t s natural to use T k as the target value for nvestng strateges. We keep on updatng ω k by cyclng through the nput-output pars of the days of the tranng perod and fnally obtan ω after suffcent tmes of teraton. Throughout the nvestng perod we use ω and Investor s captal after round n n the nvestng perod s expressed as K n = K n 1 (1 + f(u n 1, ω )x n ). Back propagaton s an algorthm whch updates weghts k ω 1,2 j and k ω 2,3 so that the error functon E k = 1 2 (T k k y 3 ) 2 decreases, where T k s the desred output of the network and k y 3 s the actual output of the network. The weght k ω 2,3 of day k s renewed to the weght of day k + 1 as k+1 ω 2,3 where k+1 ω 2,3 = k ω 2,3 + k ω 2,3 = k ω 2,3 = k ω 2,3 β k ɛ 1 k y 2, β E k k ω 2,3 = k ω 2,3 β E k k I 3 k I 3 k ω 2,3 k ɛ 1 = E k k I 3 = E k k y 3 k y 3 k I 3 = (T k k y 3 )(1 tanh 2 ( k I 3 )). 8

Also weght k ω 1,2 j s renewed as k+1 ω 1,2 j = k ω 1,2 j + k ω 1,2 j = k ω 1,2 j = k ω 1,2 j β k ɛ 2 (ũ k 1 ) j, β E k k ω 1,2 j = k ω 1,2 j β E k k I 3 k y 2 k I 3 k y 2 k I 2 where k ɛ 2 = E k k I 3 k y 2 = k ɛ 1 ω 2,3 k I 3 k y 2 k I 2 (1 tanh 2 ( k I 2 )). At the end of each step we calculate the tranng error defned as k I 2 k ω 1,2 j tranng error = 1 2m m (T k k y 3 ) 2 = 1 m m E k, (3) where m s the length of the tranng perod. We end the teraton when the the tranng error becomes smaller than the threshold µ, whch s set suffcently small. Here let us summarze the algorthm of NNBP n the tranng perod. 1. We set k = 1. 2. Gven the nput vector u k 1 = (x k 1,..., x k L ) and the value of ω k, L we frst evaluate k I 2 = k ω 1,2 j (u k 1) j and then k y 2 = tanh( k I 2 ). Also j=1 we set the learnng constant β. 3. We calculate k I 3 = M =1 k ω 2,3 k y 2 and then k y 3 = tanh( k I 3 ) wth k I 2 and k y 2 of the prevous step. Then we update weght ω k wth the weght updatng formula k+1 ω 2,3 = k ω 2,3 β k ɛ 1 k y 2 and k+1 ω 1,2 j = k ω 1,2 j β k ɛ 2 (u k 1 ) j. 4. Go back to step 2 settng k + 1 k and ω k+1 ω k whle 1 k m. When k = m we set k = 1 and ω 1 ω m+1 and contnue the algorthm untl the tranng error becomes less than µ. 3.2 Markovan proportonal bettng strateges In ths secton we present some sequental optmzng strateges that are rather smple compared to strateges wth neural network n Secton 2 and Secton 3.1. The strateges of ths secton are generalzatons of Markovan strategy n [12] for con-tossng games to bounded forecastng games. We 9

present these smple strateges for comparson wth SOSNN and observe how complexty n functon f ncreases or decreases Investor s captal processes n numercal examples n later sectons. Consder maxmzng the logarthm of Investor s captal n (1): log K n = n log(1 + α k x k ). We frst consder the followng smple strategy of [9] n whch we use α n = α n 1, where α n 1 = argmax Π n 1 (1 + αx k). In ths paper we denote ths strategy by MKV0. As a generalzaton of MKV0 consder usng dfferent nvestng ratos dependng on whether the prce went up or down on the prevous day. Let α k = α + k when x k 1 was postve and α k = α k when t was negatve. We denote ths strategy by MKV1. In the bettng on the nth day we use α n + = α n 1 + and αn = α n 1, where u k 1 = (x k 1 ) and (α n 1 +, αn 1 ) = argmax Π n 1 (1 + f(u k 1, α +, α )x k ), f(u k 1, α +, α ) = α + I {xk 1 0} + α I {xk 1 <0}. Here I { } denotes the ndcator functon of the event n { }. The captal process of MKV1 s wrtten n the form of (1) as log K n = n log(1 + f(u k 1, α + k 1, α k 1 )x k ). We can further generalze ths strategy consderng prce movements of past two days. Let u k 1 = (x k 1, x k 2 ) and let f(u k 1, α ++, α +, α +, α ) = α ++ I {xk 2 0, x k 1 0} + α + I {xk 2 0, x k 1 <0} + α + I {xk 2 <0, x k 1 0} + α I {xk 2 <,x k 1 <0}. We denote ths strategy by MKV2. We wll compare performances of the above Markovan proportonal bettng strateges wth strateges based on neural networks n the followng sectons. 10

4 Smulaton wth lnear models In ths secton we gve some smulaton results for strateges shown n Secton 2 and Secton 3. We use two lnear tme seres models to confrm the behavor of presented strateges. Lnear tme seres data are generated from the Box-Jenkns famly [2], autoregressve model of order 1 (AR(1)) and autoregressve movng average model of order 2 and 1 (ARMA(2,1)) havng the same parameter values as n [15]. AR(1) data are generated as and ARMA(2,1) data are generated as x n = 0.6x n 1 + ɛ n (4) x n = 0.6x n 1 + 0.3x n 2 + ɛ n 0.5ɛ n 1, (5) where we set ɛ n N(0, 1). After the seres s generated, we dvde each value by the maxmum absolute value to normalze the data to the admssble range [ 1, 1]. Here we dscuss some detals on calculaton of each strategy. Frst we set the ntal values of elements of ω as random numbers n [ 0.1, 0.1]. In SOSNN, we use the frst 20 values of x n as ntal values and the teraton process n gradent descent method s proceeded untl ω 1,2 j < 10 4 and ω 2,3 < 10 4 wth the upper bound of 10 4 steps. As for the learnng constant β, we use learnng-rate annealng schedules whch appear n Secton 3.13 of [7]. Wth annealng schedule called the search-then-converge schedule [4] we put β at the nth step of teraton as β(n) = β 0 1 + (n/τ), where β 0 and τ are constants and we set β 0 = 1.0 and τ = 5.0. In NNBP, we tran the network wth fve dfferent tranng sets of 300 observatons generated by (4) and (5). We contnue cyclng through the tranng set untl the tranng error becomes less than µ and we set µ = 10 2 wth the upper bound of 6 10 5 steps. Also we check the ft of the network to the data by means of the tranng error for some dfferent values of β, L and M. In Markovan strateges, we agan use the frst 20 values of x n as ntal values. We also adjust the data so that the bettng s conducted on the same data regardless of L n SOSNN and NNBP or dfferent number of nputs among Markovan strateges. In Table 1 we summarze the results of SOSNN, NNBP, MKV0, MKV1 and MKV2 under AR(1) and ARMA(2,1). The values presented are averages 11

of results for fve dfferent smulaton runs of (4) and (5). As for SOSNN, we smulate ffty cases (combnatons of L = 1,..., 5 and M = 1,..., 10), but only report the cases of L = 1, 2, 3 and some choces of M because the purpose of the smulaton s to test whether L = 1 works better than other choces of L under AR(1) and L = 2 works better under ARMA(2,1). For NNBP we only report the result for one case snce fttng the parameters to the data s qute a dffcult task due to the characterstc of desred output (target). Also once we obtan the value of β, L and M wth tranng error less than the threshold µ = 10 2, we fnd that the network has successfully learned the nput-output relatonshp and we do not test other choces of the above parameters. (See Appendx for more detal.) We set β = 0.07, L = 12 and M = 30 n smulaton wth AR(1) model and β = 0.08, L = 15 and M = 40 n smulaton wth ARMA(2,1) model. Investor s captal process for each choce of L and M n SOSNN, NNBP and each Markovan strategy s shown n three rows, correspondng to rounds 100, 200, 300 of the bettng (wthout the ntal 20 rounds n SOSNN and Markovan strateges). The fourth row of each result for NNBP shows the tranng error after learnng n the tranng perod. The best value among the choces of L and M n SOSNN or among each Markovan strategy s wrtten n bold and marked wth an astersk and the second best value s also wrtten n bold and marked wth two astersks. Also calculaton results wrtten wth are cases n that smulaton dd not gve proper values for some reasons. Notce that SOSNN whose bettng rato f s specfed by a complex functon gves better performance than rather smple Markovan strateges both under AR(1) and ARMA(2,1). Also the result that NNBP gves captal processes whch are compettve wth other strateges shows that the network has successfully learned the nput-output relatonshp n the tranng perod. 5 Comparson of performances wth some stock prce data In ths secton we present numercal examples calculated wth the stock prce data of three Japanese companes SONY, Nomura Holdngs and NTT lsted on the frst secton of the Tokyo Stock Exchange for comparng bettng strateges presented n Secton 2 and Secton 3. The nvestng perod (wthout days used for ntal values) s 300 days from March 1st n 2007 to June 19th n 2008 for all strateges and the tranng perod n NNBP s 300 days from December 1st n 2005 to February 20th n 2007. We use a shorter tranng perod than those n prevous researches, because longer perods resulted n 12

Table 1: Log captal processes of SOSNN, NNBP, MKV0, MKV1, MKV2 under AR(1) and ARMA(2,1) AR(1) model SOSNN L\M 1 2 3 4 5 6 7 8 9.890 9.168 10.021 10.007 9.982 9.559 10.009 9.821 1 18.285 17.436 17.241 17.709 18.480 17.542 18.260 16.991 30.151 23.574 29.465 29.949 32.483 28.941 7.476 7.132 7.864 8.004 7.267 8.738 6.009 7.771 2 14.983 12.710 15.518 17.350 13.914 13.614 12.016 26.211 25.785 26.090 32.144 21.981 23.137 19.492 8.128 9.084 5.427 2.922 7.291 5.209 5.989 3 13.934 12.242 11.166 8.607 11.539 8.311 10.994 23.232 19.889 20.013 16.330 18.807 16.055 20.388 NNBP MKV0 MKV1 MKV2 10.118 1.175 7.831 6.517 11.921 0.800 16.974 15.452 24.323 1.647 32.392 30.875 (5.28 10 3 ) ARMA(2,1) model SOSNN L\M 1 2 3 4 5 6 7 8 4.985 5.292 4.350 4.532 4.052 3.563 3.408 1.522 1 10.234 11.108 9.905 8.793 10.012 8.411 8.029 4.746 11.666 11.460 10.024 9.924 11.781 8.016 5.882 8.518 10.287 11.052 11.567 9.147 8.483 6.915 5.579 2 18.474 20.177 21.458 20.768 13.030 12.042 13.573 16.818 25.167 24.979 25.114 17.538 15.074 16.846 8.511 10.490 7.344 7.883 7.409 7.445 7.011 5.772 3 15.401 18.241 18.120 15.697 12.395 11.961 18.047 24.904 23.362 21.280 15.930 14.185 NNBP MKV0 MKV1 MKV2 7.813 0.729 3.160 7.566 15.420 0.132 10.483 17.619 25.819 2.422 13.375 22.911 (2.04 10 2 ) poor fttng. The data for 300 days from December 1st n 2005 to February 20th n 2007 s used for nput normalzaton of x n to [ 1, 1], whch s conducted accordng to the method shown n [1] and the procedure s as follows. For the data of daly closng prces n the above perod, we frst obtan the maxmum value of absolute daly movements and then dvde daly movements n the nvestng perod by that maxmum value. In case x n 1.0 or x n 1.0 we put x n = 1.0 or x n = 1.0. Thus we obtan x n n [ 1, 1] and we use them for nputs of the neural network. We tred perods of varous lengths for normalzaton and decded to choose a relatvely short perod to avod Investor s captal processes stayng almost constant. Also n NNBP we used β = 0.07, L = 12 and M = 90 for SONY, β = 0.07, L = 15 and M = 100 for 13

Nomura and β = 0.03, L = 15 and M = 120 for NTT wth upper bound of 10 5 teraton steps. Other detals of calculaton are the same as n Secton 4. We report the results n Table 2. In Fgure 2 we show the movements of closng prces of each company durng the nvestng perod. In Fgures 3-5 we show the log captal processes of the results shown n Table 2 to compare the performance of each strategy. Fgure 3 s for SONY, Fgure 4 s for Nomura Holdngs and Fgure 5 s for NTT. For SOSNN we plotted the result of L and M that gave the best performance at n = 300 (the bottom row of the three rows) n Table 2. 7000 Sony Nomura NTT 2 SOSNN NNBP MKV1 6000 1 5000 0 Closng Prce 4000 3000 Log Captal -1-2 2000-3 1000-4 0-5 0 50 100 150 200 250 300 0 50 100 150 200 250 300 Days Rounds Fgure 2: Closng prces Fgure 3: SONY 2 SOSNN NNBP 1 SOSNN NNBP MKV0 MKV0 1 0 0-1 Log Captal -1-2 Log Captal -2-3 -3-4 -4-5 -5 0 50 100 150 200 250 300 Rounds 0 50 100 150 200 250 300 Rounds Fgure 4: Nomura Fgure 5: NTT As we see from above fgures, NNBP whch shows compettve performance for two lnear models n Secton 4 gves the worst result. Thus t s obvous that the network has faled to capture trend n the bettng perod even f t fts n the tranng perod. Also the results are favorable to SOSNN f we adopt approprate numbers for L and M. 14

Table 2: Log captal process of SOSNN, NNBP, MKV0, MKV1 and MKV2 for TSE stocks SONY SOSNN L\M 1 2 4 5 7 8 9 0.339 0.292 0.144 0.832 0.964 0.896 0.082 1 0.572 0.520 1.438 0.401 0.383 0.433 2.012 0.153 0.220 0.461 0.633 0.762 1.283 0.582 0.329 0.163 1.273 0.675 0.349 2.056 2.541 2 0.000 0.260 1.825 1.420 0.072 1.100 1.256 0.565 0.412 2.274 2.006 0.980 2.490 1.981 0.230 1.448 1.402 1.232 0.397 0.627 1.404 3 0.309 1.823 0.438 0.695 0.970 0.599 0.281 0.307 2.237 1.333 1.843 0.212 1.827 2.749 NNBP MKV0 MKV1 MKV2 1.039 0.578 0.459 1.285 2.557 0.979 1.414 0.482 3.837 1.260 0.212 2.297 (3.67 10 2 ) Nomura SOSNN L\M 1 2 4 5 7 8 9 0.200 0.212 1.309 0.479 0.839 0.679 0.726 1 1.193 0.754 2.417 0.370 2.650 0.005 1.212 3.326 2.888 2.333 0.581 6.938 4.979 0.504 0.819 0.338 4.148 0.229 4.787 1.662 0.754 2 0.969 1.030 4.920 0.007 11.478 3.385 10.046 0.202 1.127 4.941 7.003 18.795 10.897 22.861 1.066 2.076 1.458 0.389 2.926 1.783 2.451 3 1.111 1.002 5.307 0.198 2.897 5.254 9.595 3.672 1.599 10.570 3.885 0.621 8.420 14.021 NNBP MKV0 MKV1 MKV2 0.743 0.883 1.911 3.789 8.087 0.753 1.354 4.970 16.051 1.390 1.952 6.410 (2.03 10 2 ) NTT SOSNN L\M 1 2 4 5 7 8 9 0.674 0.365 1.824 0.880 5.673 0.888 4.757 1 0.825 0.411 0.246 4.248 4.472 7.931 4.386 1.115 1.269 6.049 5.476 2.769 7.899 6.606 4.134 2 1.175 8.333 9.581 5.377 11.071 9.206 5.250 0.498 10.660 7.137 13.900 0.377 3.131 2.192 4.449 1.670 10.896 8.633 3 0.431 5.861 1.366 9.349 16.990 10.811 12.315 0.092 5.463 14.584 15.456 NNBP MKV0 MKV1 MKV2 4.155 0.902 2.048 4.788 3.161 1.272 2.642 6.807 5.669 1.566 3.391 8.687 (3.73 10 2 ) 15

6 Concludng remarks We proposed nvestng strateges based on neural networks whch drectly consder Investor s captal process and are easy to mplement n practcal applcatons. We also presented numercal examples for smulated and actual stock prce data to show advantages of our method. In ths paper we only adopted normalzed values of past Market s movements for the nput u n 1 whle we can use any data avalable before the start of round n as a part of the nput as we mentoned n Secton 2.2. Let us summarze other possbltes consdered n exstng researches on fnancal predcton wth neural networks. The smplest choce s to use raw data wthout any normalzaton as n [6], n whch they analyze tme seres of Athens Stock ndex to predct future daly ndex. In [5] they adopt prce of FAZ-ndex (one of the German equvalents of the Amercan Dow-Jones- Index), movng averages for 5, 10 and 90 days, bond market ndex, order ndex, US-Dollar and 10 successve FAZ-ndex prces as nputs to predct the weekly closng prce of the FAZ-ndex. Also n [8] they use 12 techncal ndcators to predct the S&P 500 stock ndex one month n the future. From these researches we see that for longer predcton terms (such as monthly or yearly), longer movng averages or seasonal ndexes become more effectve. Thus those long term ndcators may not have much effect n daly prce predcton whch we presented n ths paper. On the other hand, adoptng data whch seems to have a strong correlaton wth closng prces of Tokyo Stock Exchange such as closng prces of New York Stock Exchange of the prevous day may ncrease Investor s captal processes presented n ths paper. Snce there are numercal dffcultes n optmzng neural networks, t s better to use small number of effectve nputs for a good performance. Another mportant generalzaton of the method of ths paper s to consder portfolo optmzaton. We can easly extend the method n ths paper to the bettng on multple assets. Let the output layer of the network have P neurons as shown n Fgure 7 and the output of each neuron s expressed as yh 3, h = 1,..., P. Then we obtan a vector y3 = (y1, 3..., yp 3 ) of outputs. The number of neurons P refers to the number of dfferent stocks Investor nvests. Investor s captal after round n s wrtten as K n = K n 1 (1 + P f h (u n 1, ω h )x n,h ), h=1 where ω h = ( (ω 1,2 j ) =1,...,M, j=1,...,l, (ω 2,3 h ) =1,...,M). 16

Fgure 6: Three-layered network for portfolo cases Thus also n portfolo cases we see that our method s easy to mplement and we can evaluate Investor s captal process n practcal applcatons. Appendx Here we dscuss tranng error n the tranng perod of NNBP. In ths paper we set the threshold µ for endng the teraton to 10 2, whle the value commonly adopted n many prevous researches s smaller, for nstance, µ = 10 4. We gve some detals on our choce of µ. Let us examne the case of Nomura Holdngs n Secton 5. In Fgure 7 we show the tranng error after each step of teraton n the tranng perod calculated wth (3). Whle the plotted curve has a typcal shape as those of prevous researches, t s unlkely that the tranng error becomes less than 10 2. Also n Fgure 8 we plot E k = 1 2 (T k k y 3 ) 2 for each k calculated wth parameter values ω after learnng. We observe that the network fals to ft for some ponts (actually 9 days out of 300 days) but perfectly fts for all other days. It can be nterpreted that the network gnores some outlers and adjust to capture the trend of the whole data. 17

0.14 2 0.12 0.1 1.5 Tranng Error 0.08 0.06 Tranng Error 1 0.04 0.5 0.02 0 0 0 20000 40000 60000 80000 100000 Epochs 0 50 100 150 200 250 300 Perods Fgure 7: tranng error (epochs) Fgure 8: tranng error (perods) References [1] E. M. Azoff. Neural Network Tme Seres Forecastng of Fnancal Markets. Wley, Chchester, 1994. [2] G. P. E. Box and G. M. Jenkns. Tme Seres: Analyss Forecastng and Control. Holden-Day, San Francsco, 1970. [3] T. M. Cover. Unversal portfolos. Mathematcal Fnance, 1, No.1, 1 29, 1991. [4] C. Darken, J. Chang and J. Moody. Learnng rate schedules for faster stochastc gradent search. IEEE Second Workshop on Neural Networks for Sgnal Processng, 3 12, 1992. [5] B. Fresleben. Stock market predcton wth backpropagaton networks. Industral and Engneerng Applcatons of Artfcal Intellgence and Expert System 5th Internatonal Conference, 451 460, 1992. [6] M. Hanas, P. Curts and J. Thalassnos. Predcton wth neural networks: The Athens stock exchange prce ndcator. European Journal of Economcs, Fnance and Admnstratve Scences, 9, 21 27, 2007. [7] S. S. Haykn. Neural Networks and Learnng Machnes. 3rd ed., Prentce Hall, New York, 2008. [8] N. L. D. Khoa, K. Sakakbara and I. Nshkawa. Stock prce forecastng usng back propagaton neural networks wth tme and proft based adjusted weght factors. SICE-ICASE Internatonal Jont Conference, 5484 5488, 2006. 18

[9] M. Kumon, A. Takemura and K. Takeuch. Sequental optmzng strategy n mult-dmensonal bounded forecastng games. arxv:0911.3933v1, 2009. [10] D. E. Rumelhart, G. E. Hnton and R. J. Wllams. Learnng nternal representaton by backpropagatng errors. Nature, 323, 533 536, 1986. [11] G. Shafer and V. Vovk. Probablty and Fnance: It s Only a Game!. Wley, New York, 2001. [12] K. Takeuch, M. Kumon and A. Takemura. Multstep Bayesan strategy n con-tossng games and ts applcaton to asset tradng games n contnuous tme. arxv:0802.4311v2, 2008. Condtonally accepted to Stochastc Analyss and Applcatons. [13] V. Vovk, A. Takemura and G. Shafer. Defensve Forecastng. Proceedngs of the 10th Internatonal Workshop on Artfcal Intellgence and Statstcs (R. G. Cowell and Z. Ghahraman edtors), 365 372, 2005. [14] Y. Yoon and G. Swales. Predctng stock prce performance: A neural network approach. Proceedngs of the 24th Annual Hawa Internatonal Conference on System, 4, 156 162, 1991. [15] G. P. Zhang. An nvestgaton of neural networks for lnear tme-seres forecastng. Computers & Operatons Research, 28, No.12, 1183 1202, 2001. 19