Modelling of Web Domain Visits by Radial Basis Function Neural Networks and Support Vector Machine Regression

Similar documents
The Development of Web Log Mining Based on Improve-K-Means Clustering Analysis

Forecasting the Direction and Strength of Stock Market Movement

Institute of Informatics, Faculty of Business and Management, Brno University of Technology,Czech Republic

Causal, Explanatory Forecasting. Analysis. Regression Analysis. Simple Linear Regression. Which is Independent? Forecasting

Support Vector Machines

Can Auto Liability Insurance Purchases Signal Risk Attitude?

A study on the ability of Support Vector Regression and Neural Networks to Forecast Basic Time Series Patterns

Feature selection for intrusion detection. Slobodan Petrović NISlab, Gjøvik University College

Forecasting the Demand of Emergency Supplies: Based on the CBR Theory and BP Neural Network

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Project Networks With Mixed-Time Constraints

ANALYZING THE RELATIONSHIPS BETWEEN QUALITY, TIME, AND COST IN PROJECT MANAGEMENT DECISION MAKING

Lecture 2: Single Layer Perceptrons Kevin Swingler

IMPACT ANALYSIS OF A CELLULAR PHONE

benefit is 2, paid if the policyholder dies within the year, and probability of death within the year is ).

Performance Analysis and Coding Strategy of ECOC SVMs

BERNSTEIN POLYNOMIALS

Conversion between the vector and raster data structures using Fuzzy Geographical Entities

Bayesian Network Based Causal Relationship Identification and Funding Success Prediction in P2P Lending

POLYSA: A Polynomial Algorithm for Non-binary Constraint Satisfaction Problems with and

A DATA MINING APPLICATION IN A STUDENT DATABASE

CS 2750 Machine Learning. Lecture 3. Density estimation. CS 2750 Machine Learning. Announcements

8.5 UNITARY AND HERMITIAN MATRICES. The conjugate transpose of a complex matrix A, denoted by A*, is given by

SVM Tutorial: Classification, Regression, and Ranking

Development of an intelligent system for tool wear monitoring applying neural networks

Gender Classification for Real-Time Audience Analysis System

L10: Linear discriminants analysis

INVESTIGATION OF VEHICULAR USERS FAIRNESS IN CDMA-HDR NETWORKS

A Novel Methodology of Working Capital Management for Large. Public Constructions by Using Fuzzy S-curve Regression

How To Know The Components Of Mean Squared Error Of Herarchcal Estmator S

Logistic Regression. Lecture 4: More classifiers and classes. Logistic regression. Adaboost. Optimization. Multiple class classification

Time Delayed Independent Component Analysis for Data Quality Monitoring

Damage detection in composite laminates using coin-tap method

Searching for Interacting Features for Spam Filtering

Frequency Selective IQ Phase and IQ Amplitude Imbalance Adjustments for OFDM Direct Conversion Transmitters

The Journal of Systems and Software

Calculation of Sampling Weights

An Alternative Way to Measure Private Equity Performance

New Approaches to Support Vector Ordinal Regression

Credit Limit Optimization (CLO) for Credit Cards

NEURO-FUZZY INFERENCE SYSTEM FOR E-COMMERCE WEBSITE EVALUATION

On the Interaction between Load Balancing and Speed Scaling

Support vector domain description

RELIABILITY, RISK AND AVAILABILITY ANLYSIS OF A CONTAINER GANTRY CRANE ABSTRACT

Single and multiple stage classifiers implementing logistic discrimination

Risk Model of Long-Term Production Scheduling in Open Pit Gold Mining

A Genetic Programming Based Stock Price Predictor together with Mean-Variance Based Sell/Buy Actions

Traffic State Estimation in the Traffic Management Center of Berlin

Financial market forecasting using a two-step kernel learning method for the support vector regression

What is Candidate Sampling

Vision Mouse. Saurabh Sarkar a* University of Cincinnati, Cincinnati, USA ABSTRACT 1. INTRODUCTION

Improved SVM in Cloud Computing Information Mining

Brigid Mullany, Ph.D University of North Carolina, Charlotte

Data Broadcast on a Multi-System Heterogeneous Overlayed Wireless Network *

Fuzzy Set Approach To Asymmetrical Load Balancing In Distribution Networks

Recurrence. 1 Definitions and main statements

Activity Scheduling for Cost-Time Investment Optimization in Project Management

Automated Network Performance Management and Monitoring via One-class Support Vector Machine

Research Article Enhanced Two-Step Method via Relaxed Order of α-satisfactory Degrees for Fuzzy Multiobjective Optimization

DEFINING %COMPLETE IN MICROSOFT PROJECT

On-Line Fault Detection in Wind Turbine Transmission System using Adaptive Filter and Robust Statistical Features

GRAVITY DATA VALIDATION AND OUTLIER DETECTION USING L 1 -NORM

8 Algorithm for Binary Searching in Trees

SIMPLE LINEAR CORRELATION

CHAPTER 14 MORE ABOUT REGRESSION

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. Sequential Optimizing Investing Strategy with Neural Networks

CHOLESTEROL REFERENCE METHOD LABORATORY NETWORK. Sample Stability Protocol

Risk-based Fatigue Estimate of Deep Water Risers -- Course Project for EM388F: Fracture Mechanics, Spring 2008

A COLLABORATIVE TRADING MODEL BY SUPPORT VECTOR REGRESSION AND TS FUZZY RULE FOR DAILY STOCK TURNING POINTS DETECTION

On the Interaction between Load Balancing and Speed Scaling

Study on Model of Risks Assessment of Standard Operation in Rural Power Network

APPLICATION OF COMPUTER PROGRAMMING IN OPTIMIZATION OF TECHNOLOGICAL OBJECTIVES OF COLD ROLLING

A DYNAMIC CRASHING METHOD FOR PROJECT MANAGEMENT USING SIMULATION-BASED OPTIMIZATION. Michael E. Kuhl Radhamés A. Tolentino-Peña

Using Multi-objective Metaheuristics to Solve the Software Project Scheduling Problem

An artificial Neural Network approach to monitor and diagnose multi-attribute quality control processes. S. T. A. Niaki*

Research Article Integrated Model of Multiple Kernel Learning and Differential Evolution for EUR/USD Trading

Face Verification Problem. Face Recognition Problem. Application: Access Control. Biometric Authentication. Face Verification (1:1 matching)

On the Optimal Control of a Cascade of Hydro-Electric Power Stations

A hybrid global optimization algorithm based on parallel chaos optimization and outlook algorithm

Luby s Alg. for Maximal Independent Sets using Pairwise Independence

THE APPLICATION OF DATA MINING TECHNIQUES AND MULTIPLE CLASSIFIERS TO MARKETING DECISION

The Greedy Method. Introduction. 0/1 Knapsack Problem

How To Solve An Onlne Control Polcy On A Vrtualzed Data Center

Forecasting and Modelling Electricity Demand Using Anfis Predictor

Biometric Signature Processing & Recognition Using Radial Basis Function Network

Statistical Methods to Develop Rating Models

The Application of Fractional Brownian Motion in Option Pricing

A Simple Approach to Clustering in Excel

The Network flow Motoring System based on Particle Swarm Optimized

Politecnico di Torino. Porto Institutional Repository

Transcription:

Modellng of Web Doman Vsts by Radal Bass Functon Neural Networks and Support Vector Machne Regresson Vladmír Olej, Jana Flpová Insttute of System Engneerng and Informatcs Faculty of Economcs and Admnstraton, Unversty of Pardubce, Studentská 84 532 10 Pardubce, Czech Republc, {vladmr.olej@upce.cz, jana.flpova@upce.cz} Abstract. The paper presents basc notons of web mnng, radal bass functon (RBF) neural networks and -nsenstve support vector machne regresson ( - SVR) for the predcton of a tme seres for the webste of the Unversty of Pardubce. The model ncludes pre-processng tme seres, desgn RBF neural networks and -SVR structures, comparson of the results and tme seres predcton. The predctons concernng short, ntermedate and long tme seres for varous ratos of tranng and testng data. Predcton of web data can be beneft for a web server traffc as a complcated complex system. Keywords: Web mnng, radal bass functon neural networks, -nsenstve support vector machne regresson, tme seres, predcton. 1 Introducton Data modellng (predcton, classfcaton, optmzaton) obtaned by web mnng [1] from the log fles and data on a vrtual server, as a complcated complex system that affects these data, there are a few problems worked out. The web server represents a complcated complex system; vrtual system usually works wth several vrtual machnes that operate over multple databases. The qualty of the actvtes of the system also affects the data obtaned usng web mnng. Gven vrtual system s characterzed by ts operatonal parameters, whch are changng over tme, so t s a dynamc system. The data show a nonlnear characterstcs, are heterogeneous, nconsstent, mssng and uncertan. Currently there are a number of methods for modellng of the data obtaned by web mnng. These methods can generally be dvded nto methods of modellng wth unsupervsed learnng [2,3] and supervsed learnng [4]. The present work bulds on the modellng of web domans vst wth uncertanty [5,6]. In the paper s presented a problem formulaton wth the am of descrbng the tme seres web upce.cz (web presentaton vsts), ncludng possbltes of pre-processng whch s realzed by means of smple mathematc-statstc methods. Next, there are ntroduced basc notons of RBFs [7] neural networks and -SVRs [4,8] for tme

seres predcton. Further, the paper ncludes a comparson of the predcton results of the model desgned. The model represents a tme seres predcton for web upce.cz wth the pre-processng done by smple mathematcal statstcal methods. The predcton s carred out through RBF neural networks and -SVRs for dfferent duratons of tme seres and varous ratos O tran :O test tranng O tran and testng O test dates (O= O tran O test ). There s expected, that for shorter tme seres predcton t would be better to use RBF neural networks and for longer -SVRs. 2 Problem Formulaton The data for predcton of the tme seres web upce.cz over a gven tme perod was obtaned from Google Analytcs. Ths web mnng tool, whch makes use of Java- Scrpt code mplemented for web presentaton, offers a wde spectrum of operaton characterstcs (web metrcs). Metrcs provded by Google Analytcs can be dvded nto followng sectons: vsts, sources of access, contents and converson. In the secton vsts t can be montored for example the number of vstors, the number of vsts and number of pages vewed as well as the rato of new and returnng vstors. Indcator geolocaton,.e. from whch country are vstors most often, s needed to be known because of language mutatons, for example. In order to predct the vst rate to the Unversty of Pardubce, Czech Republc webste (web upce.cz) t s mportant montorng the ndcator of the number of vsts wthn a gven tme perod. One vst here s defned as an unrepeated combnaton of IP address and cookes. A submetrcs s absolutely unque vst defned by an unrepeatable IP address and cookes wthn gven tme perod. The basc nformaton obtaned from Google Analytcs about web upce.cz durng May 2009 consdered of the followng: The total vst rate durng gven monthly cycles. A clear trend s obvous there, wth Monday havng the hghest vst rate, whch n turn decreases as the week progresses; Saturday has the lowest vst rate; The average number of pages vsted s more than three; A vstor stays on certan page fve and half a mnutes on average; The bounce rate s approxmately 60%; Vstors generally come drectly to the webste, whch s postve; The favourte pages s the man page, followed by the pages of the Faculty of Economc and Admnstraton and the Faculty of Phlosophy. The measurement of the vst rate of the Unversty of Pardubce web page, (web upce.cz) took place tme perods of regular, controlled ntervals. The result represents a tme seres. The pre-processng of data was realzed by means of smple mathematc-statstc methods: a smple movng average (SMA), a central movng average (CMA), a movng medan (MM) along wth smple exponental smoothng (SES) and double exponental smoothng (DES) at tme t. Pre-processng was used wth am of smoothng the outlers, whle mantanng the physcal nterpretaton of data. The general formulaton of the model of predcton of the vst rate for upce.cz can be stated n thusly: y =f(x t 1,x t 2,,x t m), m=5, where y s the number of daly web vsts n tme t+1, y s the number of daly web vsts n tme t, x t 1 s SMA, x t 2 s CMA, x t 3 s MM, x t 4 s SES, and x t 5 s DES at tme t. An example of the preprocessng of the tme seres of web upce.cz s represented n Fg. 1.

Fg. 1. The pre-processng of web upce.cz vsts by SMA On the bass of the dfferent duratonal tme seres of web upce.cz, concrete predcton models for the vst rate to upce.cz web can be defned thusly y =f(ts S SMA,TS S CMA,TS S MM,TS S SES,TS S DES), y =f(ts I SMA,TS I CMA,TS I MM,TS I SES,TS I DES), y =f(ts L SMA,TS L CMA,TS L MM,TS L SES,TS L DES), where TS - tme seres, S - short TS (264 days), I - ntermedate TS (480 days), L - long TS (752 days) at tme t. (1) 3 Basc Notons of RBF Neural Networks and -SVR The term RBF neural network [7] refers to any knd of feed-forward neural networks that uses RBF as an actvaton functon. RBF neural networks are based on supervsed learnng. The output f(x,h,w) RBF of a neural network can be defned ths way q f ( x, H, w) w h ( x), (2) 1 where H={h 1 (x),h 2 (x),,h (x),,h q (x)} s a set of actvaton functons RBF of neurons (of RBF functons) n the hdden layer and w are synapse weghts. Each of the m components of vector x=(x 1,x 2,,x k,,x m ) s an nput value for the q actvaton functons h (x) of RBF neurons. The output f(x,h,w) of RBF neural network represents a lnear combnaton of outputs from q RBF neurons and correspondng synapse weghts w. The actvaton functon h (x) of an RBF neural network n the hdden layer belongs to a specal class of mathematcal functons whose man characterstc s a monotonous rsng or fallng at an ncreasng dstance from center c of the actvaton functon h (x) of an RBF. Neurons n the hdden layer can use one of several actvaton functons h (x) of an RBF neural network, for example a Gaussan actvaton functon (a one-dmensonal actvaton functon of RBF), a rotary Gaussan actvaton functon (a two-dmensonal RBF actvaton functon), multsquare and nverse multsquare actvaton functons or Cauchy s functons. Results may be presented n ths manner

q x c (3) h(x,c,r)= exp ( ), 1 r where x=(x 1,x 2,,x k,,x m ) represents the nput vector, C={c 1,c 2,,c,,c q } are centres of actvaton functons h (x) an RBF neural network and R={r 1,r 2,,r,,r q } are the raduses of actvaton functons h (x). The neurons n the output layer represent only a weghted sum of all nputs comng from the hdden layer. An actvaton functon of neurons n the output layer can be lnear, wth the unt of the output eventually beng convert by jump nstructon to bnary form. The RBF neural network learnng process requres a number of centres c of actvaton functon h (x) of the RBF neural networks to be set as well as for the most sutable postons for RBF centres c to be found. Other parameters are raduses of centres c, rate of actvaton functons h (x) of RBFs and synapse weghts W(q,n). These are setup between the hdden and output layers. The desgn of an approprate number of RBF neurons n the hdden layer s presented n [4]. Possbltes of centres recognton c are mentoned n [7] as a random choce. The poston of the neurons s chosen randomly from a set of tranng data. Ths approach presumes that randomly pcked centres c wll suffcently represent data enterng the RBF neural network. Ths method s sutable only for small sets of nput data. Use on larger sets, often results n a quck and needless ncrease n the number of RBF neurons n the hdden layer, and therefore an unjustfed complexty of the neural network. The second approach to locatng centres c of actvaton functons h (x) of RBF neurons can be realzed by a K-means algorthm. In nonlnear regresson -SVR [4,8,9,10,11] mnmzes the loss functon L(d,y) wth nsenstve [4,11]. Loss functon L(d,y)= d-y where d s the desred response and y s the output estmate. The constructon of the -SVR for approxmatng the desred response d can be used for the extenson of loss functon L(d,y) as follows d y L ( d, y) for d - y, (4) 0 else where s a parameter. Loss functon L (d,y) s called a loss functon wth nsenstve. Let the nonlnear regresson model n whch the dependence of the scalar d vector x expressed by d=f(x) + n. Addtve nose n s statstcally ndependent of the nput vector x. The functon f(.), and nose statstcs are unknown. Next, let the sample tranng data (x,d ), =1,2,...,N, where x and d s the correspondng value of the output model d. The problem s to obtan an estmate of d, dependng on x. For further progress s expected to estmate d, called y, whch s wdespread n the set of nonlnear bass functons j x), j=0,1,...,m 1 ths way y m j 1 0 w j j 2 ( x) wt ( x), (5) where (x)=[ (x), (x),, m 1 (x)] T and w=[w 0,w 1,...,w m1 ]. It s assumed that (x)=1 n order to the weght w 0 represents bas b. The soluton to the problem s to mnmze the emprcal rsk

1 N R emp L ( d, y ), (6) N 1 under condtons of nequalty w c 0, where c 0 s a constant. The restrcted optmzaton problem can be rephrased usng two complementary sets of non-negatve varables. Addtonal varables and descrbe loss functon L (d,y) wth nsenstvty. The restrcted optmzaton problem can be wrtten as an equvalent to mnmzng the cost functonal N 1 ( w,, ) C( ( wt )) w, (7) 1 2 under the constrants of two complementary sets of non-negatve varables and. The constant C s a user-specfed parameter. Optmzaton problem (7) can be easly solved n the dual form. The basc dea behnd the formulaton of the dual-shaped structure s the Lagrangan functon [9], the objectve functon and restrctons. Can then be defned Lagrange multplers wth ther functons and parameters whch ensure optmalty of these multplers. Optmzaton the Lagrangan functon only descrbes the orgnal regresson problem. To formulate the correspondng dual problem a convex functon can be obtaned (for shorthand) N Q (, ) d ( ) ( ) 1 1 N N ( )( j j ) K ( x, x j ), 2 1 j 1 where K(x,x j ) s kernel functon defned n accordance wth Mercers theorem [4,8]. Solvng optmzaton problem s obtaned by maxmzng Q( ) wth respect to Lagrange multplers and and provded a new set of constrants, whch hereby ncorporated constant C contaned n the functon defnton (w, ). Data ponts covered by the, defne support vectors [4,8,9]. N 1 (8) 4 Modellng and Analyss of the Results The desgned model n Fg. 2 demonstrates predcton modellng of the tme seres web upce.cz. Data pre-processng s carred out by means of data standardzaton. Thereby, the dependency on unts s elmnated. Then the data are pre-processed through smple mathematcal statstcal methods (SMA,CMA,MM,SES, and DES). Data pre-processng makes the sutable physcal nterpretaton of results possble. The pre-processng s run for tme seres of dfferent duraton namely (TS S SMA, TS S CMA,TS S MM,TS S SES,TS S DES),(TS I SMA,TS I CMA,TS I MM,TS I SES,TS I DES),(TS L SMA,TS L CMA, TS L MM,TS L SES,TS L DES). The predcton s made for the aforementoned pre-processed tme seres S, I, and L wth the help of RBF neural networks, -SVR wth a polynomal kernel functon and -SVR wth a RBF kernel functon for varous sets of tranng O tran and testng O test data.

y Data standardzaton Data pre-processng by SMA,CMA,MM,SES,DES Modellng by RBF neural networks Modellng by -SVR wth RBF kernel functon Modellng by -SVR wth polynomal kernel functon Comparson of the results and predcton y =f(x t 1,x t 2,,x t m), m=5 Fg. 2. Predcton modellng of tme seres for web.upce.cz In Fg. 3 are shown the dependences of Root Mean Squared Error (RMSE) on the number of neurons q n the hdden layer. The dependences RMSE on the parameter are represented n Fg. 4. In Fg. 5 the dependences RMSE are on the parameter for tranng O tran and as well as the testng O test set, where: - O tran_s, - O tran_i, - O tran_l, - O test_s, - O test_i, - O test_l. The parameter allows for an overrun of the local extreme n the learnng process and the followng progress of learnng. The parameter represent the selecton of centers RBFs as well as guarantees the correct allocaton of neurons n the hdden layer for the gven data enterng the RBF neural network. Fg. 3. RMSE dependences on the parameter q Fg. 4. RMSE dependences on the parameter Fg. 5. RMSE dependences on the parameter Conclusons presented n [4,7,10] are verfed by the analyss of results (wth 10- fold cross valdaton). RMSE tran s lowered untl t reaches to the value 0.3, when q neurons n the hdden layer ncreases n sze 125 and wth the greater rato O tran :O test.

RMSE test decreases only when the number q RBF neurons ncreases at a sgnfcantly slower rato than the rato O tran :O test. Mnmum RMSE test moves rght wth the ncreasng rato O tran :O test. Next, determnaton of the optmal number of q neurons n the hdden layer s necessary. Respectvely, a lower number q neurons represents the ncreasng RMSE test. In Table 1, Table 2, and Table 3 are shown optmzed results of the analyss of the experments for dfferent parameters of the RBF neural networks (wth RBF actvaton functon) wth dfferent duratons of tme seres, varous ratos of O tran :O test tranng O tran and testng O test of data sets and same amount of learnng at p=600 cycles. Table 2 bulds on the best set of q n Table1 and Table 3 bulds on the best set of q (Table 1) and (Table 2). Table 1. Optmzed results of RMSE analyss usng the amount of neurons q n the hdden layer ( and are constant for gven TS) TS O tran :O test Number q of neurons RMSE tran RMSE test S 50:50 80 0.9 1 0.383 0.408 S 66:34 80 0.9 1 0.331 0.463 S 80:20 80 0.9 1 0.343 0.365 I 50:50 125 0.9 1 0.372 0.316 I 66:34 125 0.9 1 0.295 0.440 I 80:20 125 0.9 1 0.352 0.326 L 50:50 25 0.9 1 0.409 0.402 L 66:34 25 0.9 1 0.404 0.415 L 80:20 25 0.9 1 0.409 0.390 Table 2. Optmzed results of RMSE analyss usng parameter q and are constant for gven TS) TS O tran :O test Number q of neurons RMSE tran RMSE test S 50:50 80 0.5 1 0.473 0.385 S 66:34 80 0.5 1 0.341 0.445 S 80:20 80 0.5 1 0.311 0.408 I 50:50 125 0.3 1 0.313 0.344 I 66:34 125 0.3 1 0.302 0.428 I 80:20 125 0.3 1 0.343 0.365 L 50:50 25 0.3 1 0.400 0.387 L 66:34 25 0.3 1 0.398 0.409 L 80:20 25 0.3 1 0.406 0.391 Table 3. Optmzed results of RMSE analyss usng parameter q and are constant for gven TS) TS O tran :O test Number q of neurons RMSE tran RMSE test S 50:50 80 0.5 1 0.385 0.473 S 66:34 80 0.5 1 0.341 0.445 S 80:20 80 0.5 1 0.311 0.408 I 50:50 125 0.3 3 0.357 0.375 I 66:34 125 0.3 3 0.327 0.405 I 80:20 125 0.3 3 0.376 0.441 L 50:50 25 0.3 1 0.400 0.387 L 66:34 25 0.3 1 0.398 0.409 L 80:20 25 0.3 1 0.406 0.391

The dependences RMSE on the parameter C are represented n Fg. 6, the dependences RMSE on the parameter n Fg. 7, the dependences RMSE on the parameter n Fg. 8 for tranng O tran as well as testng O test set, where: - O tran_s, - O tran_i, - O tran_l, - O test_s, - O test_i, - O test_l. The parameters C, are functons of kernel functons k(x,x ) [4,11] varatons. In the learnng process -SVR are set usng 10-fold cross valdaton. Parameter C controls the trade off between errors of the - SVR of tranng data and margn maxmzaton; [4] selects support vectors n the regresson structures, and represents the rate of polynomal kernel functon k(x,x ). The coeffcent characterzes polynomal and RBF kernel functon. Fg. 6. RMSE dependences on the parameter C Fg. 7. RMSE dependences on the parameter Fg. 8. RMSE dependences on the parameter The confrmaton of conclusons presented n [4,11] s verfed by the analyss of results. RMSE test for the -SVR wth RBF kernel functon lowers toward zero value wth decreasng C (n the case the user expermentaton) and hgher rato of O tran :O test. In the use of 10-fold cross valdaton, RMSE test moves towards zero wth an ncrease of parameter for O tran_s. In the -SVR wth polynomal kernel functon RMSE test sgnfcantly decreases when the parameter decreases (Fg. 7). Mnmum RMSE test moves rghts wth an ncrease of the parameter (the mnmum s between 0.2 to 0.3), whereas an ndrect correlaton between the rato O tran :O test and RMSE test obtans. In Table 4 and Table 5 are shown the optmzed results of the analyss of the experments for dfferent parameters of the -SVR (wth RBF and a polynomal kernel functon) wth dfferent duratons of tme seres, varous ratos O tran :O tes tranng O tran and testng O test of data sets and same amount of learnng at p=600 cycles. The tables are not represented by the partal results for the varous parameters, but only the resultng set of parameters for the TS and the rato O tran :O test.

Table 4. Optmzed results of RMSE analyss usng parameter C TS O tran :O test C RMSE tran RMSE test S 50:50 8 0.2 0.4 0.343 0.414 S 66:34 10 0.1 0.4 0.365 0.314 S 80:20 10 0.1 0.4 0.355 0.306 I 50:50 10 0.1 0.4 0.331 0.369 I 66:34 10 0.1 0.4 0.342 0.361 I 80:20 9 0.1 0.4 0.350 0.336 L 50:50 6 0.1 0.2 0.382 0.409 L 66:34 4 0.1 0.2 0.383 0.416 L 80:20 4 0.1 0.2 0.384 0.445 Table 5. Optmzed results of RMSE analyss usng parameters and TS O tran :O test C RMSE tran RMSE test S 50:50 10 0.1 0.2 1 0.411 0.319 S 66:34 10 0.2 0.2 2 0.474 0.379 S 80:20 10 0.1 0.2 3 0.587 0.509 I 50:50 6 0.1 0.3 1 0.385 0.416 I 66:34 10 0.1 0.3 2 0.436 0.486 I 80:20 6 0.1 0.3 3 0.526 0.567 L 50:50 10 0.1 0.2 2 0.455 0.468 L 66:34 6 0.1 0.2 1 0.385 0.416 L 80:20 8 0.1 0.2 1 0.385 0.443 The orgnal value of TS S (O test ) y compared wth predcted values of y (n whch tme s t+ t, t=1 day) usng RBF neural network s dsplayed n Fg. 9. In Fg. 10 TS S (O test ) y s then compared to predcted values y usng the -SVR (wth RBF kernel functon). Fg. 9. TS S wth values predcated by RBF Fg. 10. TS S wth values predcated by SVR In Table 6, there s a comparson of the RMSE tran and RMSE test on the tranng and testng set to other desgned and analyzed structures. For example, t was used a fuzzy nference system (FIS) Takag-Sugeno [5,6], ntutonstc fuzzy nference system (IFIS) Takag-Sugeno [5,6], feed-forward neural networks (FFNNs), RBF neural networks and -SVR 1 ( -SVR 2 ) wth RBF (polynomal kernel functon) wth preprocessng nput data by smple mathematcal statstcal methods. Table 6. Comparson of the RMSE tran and RMSE test on the tranng and testng data to other desgned and analyzed structures of fuzzy nference systems and neural networks FIS IFIS FFNN RBF -SVR 1 -SVR 2 RMSE tran 0.221 0.224 0.593 0.311 0.331 0.385 RMSE test 0.237 0.239 0.687 0.408 0.369 0.416

5 Concluson The proposed model conssts of data pre-processng and actual predcton usng RBF neural networks as well as -SVR wth polynomal and RBF kernel functons. Furthermore, the modellng was done for tme seres of dfferent lengths and dfferent parameters of neural networks. The analyss results for varous ratos of O tran :O test show trends for RMSE tran and RMSE test. From the analyss of all obtaned results of modellng tme seres by RBF neural network ( -SVR) shows that RMSE test takes mnmum values [4,10,11] for TS I (TS S ). Further drecton of research n the area of modellng web doman vsts (excludng the use of uncertanty [5,6] and modellng usng RBF neural networks and machne learnng usng -SVR) s focused on dfferent structures of neural networks. The crux of modellng are dfferent lengths of tme seres, varous ratos of O tran :O test and dfferent technques of ther parttonng. Predcton usng web mnng gves better characterzaton of webspace. On the bass of that system engneers can better characterze the load of complex vrtual system and ts dynamcs. The RBF neural networks ( -SVR) desgn was carred out n SPSS Clementne (STATISTICA). Acknowledgments. Ths work was supported by the scentfc research project of Mnstry of Envronment, the Czech Republc under Grant No: SP/42/60/07. References [1] Cooley, R., Mobasher, B., Srvstava, J.: Web Mnng: Informaton and Pattern Dscovery on the World Wde Web. In: 9 th IEEE Internatonal Conference on Tools wth Artfcal Intellgence, ICTAI 97, Newport Beach, CA (1997) [2] Krshnapuram, R., Josh, A., Y, L.: A Fuzzy Relatve of the K-medods Algorthm wth Applcaton to Document and Snppet Clusterng. In: IEEE Internatonal Conference on Fuzzy Systems, pp.1281-1286. Korea (1999) [3] Pedrycz, W.: Condtonal Fuzzy C-means. Pattern Recognton Letters, 17, 625-632 (1996) [4] Haykn, S.: Neural Networks: A Comprehensve Foundaton. Prentce-Hall Inc., New Jersey (1999) [5] Olej, V., Hájek, P., Flpová, J.: Modellng of Web Doman Vsts by IF-Inference System. WSEAS Transactons on Computers, Issue 10, 9, 1170-1180 (2010) [6] Olej, V., Flpová, J., Hájek, P.: Tme Seres Predcton of Web Doman Vsts by IF- Inference System. In: Proc. of the 14 th WSEAS Internatonal Conference on Systems, Latest Trends on Computers, N. Mastoraks at al. (eds.), Vol.1, Greece, pp.156-161 (2010) [7] Broomhead, D.S., Lowe, D.: Multvarate Functonal Interpolaton and Adaptve Networks. Complex Systems, 2, 321-355 (1988) [8] Crstann, N., Shawe-Taylor, J.: An Introducton to Support Vector Machnes and other Kernel-based Learnng Methods. Cambrdge Unversty Press, Cambrdge (2000) [9] Smola, A., Scholkopf, J.: A Tutoral on Support Vector Regresson. Statstcs and Computng, 14, 199 222 (2004) [10] Nyog, P., Gros, F.: On the Relatonshp between Generalzaton Error, Hypothess Complexty, and Sample Complexty for Radal Bass Functons. Massachusetts Insttute of Technology Artfcal Intellgence Laboratory, Massachusetts (1994) [11] Vapnk, V.N.: The Nature of Statstcal Learnng Theory. Sprnger-Verlag New York (1995)