Automated Methods for Fuzzy Systems
|
|
- Irene Cole
- 7 years ago
- Views:
Transcription
1 Automated Methods for Fuzzy Systems Gradient Method Adriano Joaquim de Oliveira Cruz PPGI-UFRJ September 2012 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
2 Summary 1 Introduction Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
3 Summary 1 Introduction 2 Training Standard Fuzzy System Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
4 Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
5 Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
6 Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
7 Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
8 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
9 A precise model is a contradiction. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
10 Bibliography Kevin M. Passino, Stephen Yurkovich Fuzzy Control in Chapter 5. Addison Wesley Longman, Inc, USA, Timothy J. Ross Fuzzy Logic with Engineering Applications. John Wiley and Sons, Inc, USA, J. R. Jang, C. Sun, E. Mizutani Neuro Fuzzy and Soft Computing: A Computational Approach to Learning and Machine Intelligence Prentice Hall, NJ, USA, 1997 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
11 Constructing fuzzy systems How to construct a fuzzy system from numeric data? Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
12 Constructing fuzzy systems How to construct a fuzzy system from numeric data? Using data obtained experimentally from a system, it is possible to identify the model. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
13 Constructing fuzzy systems How to construct a fuzzy system from numeric data? Using data obtained experimentally from a system, it is possible to identify the model. Find a model that fits the data by using fuzzy interpolation capabilities. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
14 Introduction We need to construct a fuzzy system f(x,θ) that approximate the function g represented in the training data G. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
15 Introduction We need to construct a fuzzy system f(x,θ) that approximate the function g represented in the training data G. There is no guarantee that it will succeed. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
16 Introduction We need to construct a fuzzy system f(x,θ) that approximate the function g represented in the training data G. There is no guarantee that it will succeed. It provides a method to tune all parameters of a fuzzy system. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
17 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
18 The System Gaussian input membership functions with centers c i j and spreads σ i j. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
19 The System Gaussian input membership functions with centers c i j and spreads σ i j. Output membership function centers b i. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
20 The System Gaussian input membership functions with centers c i j and spreads σ i j. Output membership function centers b i. Product for premise and implication. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
21 The System Gaussian input membership functions with centers cj i and spreads σj i. Output membership function centers b i. Product for premise and implication. Center-average defuzzification. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
22 The System Gaussian input membership functions with centers c i j and spreads σ i j. Output membership function centers b i. Product for premise and implication. Center-average defuzzification. It is described by f(x θ) = [ R i=1 b n i j=1 exp 1 2 ( ) ] 2 x j cj i σj i [ ( ) ] R 2 n i=1 j=1 exp 1 x j cj i 2 σj i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
23 Error Suppose that you have the m th training data pair (x,y) G. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
24 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
25 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. The equation for the error surface is: e m = 1 2 [f(x θ) y]2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
26 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. The equation for the error surface is: e m = 1 2 [f(x θ) y]2 We seek to minimize e m by choosing the parameters θ that are Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
27 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. The equation for the error surface is: e m = 1 2 [f(x θ) y]2 We seek to minimize e m by choosing the parameters θ that are b i,cj i and θj i, i = 1,2,...,R, j = 1,2,...,n. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
28 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. The equation for the error surface is: e m = 1 2 [f(x θ) y]2 We seek to minimize e m by choosing the parameters θ that are b i,cj i and θj i, i = 1,2,...,R, j = 1,2,...,n. R rules, n input variables. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
29 Error Suppose that you have the m th training data pair (x,y) G. The GM s goal is to minimize the error between the predicted output value, f(x m θ) and the actual output value y m. The equation for the error surface is: e m = 1 2 [f(x θ) y]2 We seek to minimize e m by choosing the parameters θ that are b i,cj i and θj i, i = 1,2,...,R, j = 1,2,...,n. R rules, n input variables. θ(k) will be used to denote these parameter s values at time k. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
30 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
31 b i Update Law How to adjunt the b i to minimize e m. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
32 b i Update Law How to adjunt the b i to minimize e m. We will use b i (k +1) = b i (k) λ 1 e m b i k Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
33 b i Update Law How to adjunt the b i to minimize e m. We will use where i = 1,2,...,R b i (k +1) = b i (k) λ 1 e m b i k Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
34 b i Update Law How to adjunt the b i to minimize e m. We will use where i = 1,2,...,R b i (k +1) = b i (k) λ 1 e m b i This is the gradiante descent approach. k Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
35 Gradient Descent The update method would move b i along the negative gradient of the error surface. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
36 Gradient Descent The update method would move b i along the negative gradient of the error surface. The parameter λ 1 > 0 characterizes the step size. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
37 Gradient Descent The update method would move b i along the negative gradient of the error surface. The parameter λ 1 > 0 characterizes the step size. If λ 1 is chosen too small, then b i is adjusted very slowly. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
38 Gradient Descent The update method would move b i along the negative gradient of the error surface. The parameter λ 1 > 0 characterizes the step size. If λ 1 is chosen too small, then b i is adjusted very slowly. If λ 1 is chosen too big, then it may step over the minimum value of e m. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
39 Gradient Descent The update method would move b i along the negative gradient of the error surface. The parameter λ 1 > 0 characterizes the step size. If λ 1 is chosen too small, then b i is adjusted very slowly. If λ 1 is chosen too big, then it may step over the minimum value of e m. Some algorithms try to adaptively choose the step size. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
40 Gradient Descent The update method would move b i along the negative gradient of the error surface. The parameter λ 1 > 0 characterizes the step size. If λ 1 is chosen too small, then b i is adjusted very slowly. If λ 1 is chosen too big, then it may step over the minimum value of e m. Some algorithms try to adaptively choose the step size. If the error is big increase λ 1, but if they are decreasing take small steps. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
41 b i Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
42 b i Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m b i = (f(x m θ) y m ) f(xm θ) b i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
43 b i Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m = (f(x m θ) y m ) f(xm θ) b i b i Since f(x θ) = R i=1 b n i j=1 exp 1 2 ( x j c i j ) 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
44 b i Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m = (f(x m θ) y m ) f(xm θ) b i b i Since f(x θ) = R i=1 b n i j=1 exp 1 2 then e m = (f(x m θ) y m ) b i ( x j c i j ) 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i n j=1 exp 1 2 ( x j c i j ) 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
45 b i Update Formula II Let µ i (x m,k) = ( ) n j=1 exp 1 x j c j i 2 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
46 b i Update Formula II Let µ i (x m,k) = ( ) n j=1 exp 1 x j c j i 2 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i Let ǫ m (k) = f(x m θ(k)) y m Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
47 b i Update Formula II Let µ i (x m,k) = ( ) n j=1 exp 1 x j c j i 2 2 σ j i ( ) R n i=1 j=1 exp 1 x j c j i 2 2 σ j i Let ǫ m (k) = f(x m θ(k)) y m Then µ i (x m,k) b i (k +1) = b i (k) λ 1 ǫ m (k) R i=1 µ i(x m,k) Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
48 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
49 c i j Update Law We will use c i j (k +1) = ci j (k) λ 2 e m c i j k Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
50 c i j Update Law We will use c i j (k +1) = ci j (k) λ 2 e m c i j k where λ 2 > 0, i = 1,2,...,R and j = 1,2,...,n Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
51 c i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
52 c i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m c i j = ǫ m (k) f(xm θ(k)) µ i (x m,k) µ i (x m,k) cj i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
53 c i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m c i j Now f(xm θ(k)) µ i (x m,k) = ǫ m (k) f(xm θ(k)) µ i (x m,k) µ i (x m,k) cj i = ( R i=1 µ i(x m,k))b i (k) ( R i=1 b i(k)µ i (x m,k))(1) ( R i=1 µ i(x m,k)) 2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
54 c i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m c i j Now f(xm θ(k)) µ i (x m,k) So that f(xm θ(k)) µ i (x m,k) = ǫ m (k) f(xm θ(k)) µ i (x m,k) µ i (x m,k) cj i = ( R i=1 µ i(x m,k))b i (k) ( R i=1 b i(k)µ i (x m,k))(1) ( R i=1 µ i(x m,k)) 2 = b i(k) f(x m θ(k)) R i=1 µ i(x m,k) Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
55 c i j Update Formula II Also we have µ i(x m,k) c i j ( ) = µ i (x m xj,k) m c j i (k) (σj i(k))2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
56 c i j Update Formula II Also we have µ i(x m,k) c i j ( ) = µ i (x m xj,k) m c j i (k) (σj i(k))2 The update formula for c i j is Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
57 c i j Update Formula II Also we have µ i(x m,k) c i j ( ) = µ i (x m xj,k) m c j i (k) (σj i(k))2 The update formula for c i j is Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
58 c i j Update Formula II Also we have µ i(x m ( ),k) cj i = µ i (x m xj,k) m c j i (k) (σj i(k))2 The update formula for cj i is ( ) ( cj(k+1) i = cj(k) λ i b i (k) f(x m θ(k)) x m 2 ǫ m (k) R i=1 µ µ i (x m j cj i,k) (k) ) i(x m,k) (σj i(k))2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
59 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
60 σ i j Update Law We will use σ i j (k +1) = σi j (k) λ 3 e m σ i j k Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
61 σ i j Update Law We will use σ i j (k +1) = σi j (k) λ 3 e m σ i j k where λ 3 > 0, i = 1,2,...,R and j = 1,2,...,n Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
62 σ i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
63 σ i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m σ i j = ǫ m (k) f(xm θ(k)) µ i (x m,k) µ i (x m,k) σj i Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
64 σ i j Update Formula I Erro: e m = 1 2 [f(x θ) y]2 Regra da Cadeia: e m σ i j We already calculated f(xm θ(k)) µ i (x m,k) = ǫ m (k) f(xm θ(k)) µ i (x m,k) µ i (x m,k) σj i = b i(k) f(x m θ(k)) R i=1 µ i(x m,k) Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
65 σ i j Update Formula II Also we have µ i(x m,k) σ i j ( ) = µ i (x m (x m,k) j cj i(k)2 ) (σj i(k))3 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
66 σ i j Update Formula II Also we have µ i(x m,k) σ i j ( ) = µ i (x m (x m,k) j cj i(k)2 ) (σj i(k))3 The update formula for σ i j is Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
67 σ i j Update Formula II Also we have µ i(x m,k) σ i j ( ) = µ i (x m (x m,k) j cj i(k)2 ) (σj i(k))3 The update formula for σ i j is Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
68 σ i j Update Formula II Also we have µ i(x m ( ),k) σj i = µ i (x m (x m,k) j cj i(k)2 ) (σj i(k))3 The update formula for σj i is ) 2 σj(k i +1) = σj(k) λ i 3 ǫ m (k) b i(k) f(x m θ(k)) (x R i=1 µ µ i (x m j m cj i(k),k) i(x m,k) (σj i(k))3 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
69 Section Summary 1 Introduction 2 Training Standard Fuzzy System 3 Output Membership Function Centers Update Law 4 Input Membership Function Centers Update Law 5 Input Membership Function Spreads Update Law 6 Example Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
70 Training Data Set We will use the training data set of the table to illustrate the algorithm. x 1 x 2 y x x x Table: Z = [([x 1,x 2 ],y)] Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
71 Choosing the step size The algorithm requires that a step size λ be specified for each of the three parameters. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
72 Choosing the step size The algorithm requires that a step size λ be specified for each of the three parameters. Selecting a large λ will converge faster but may risk overstepping the minimum. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
73 Choosing the step size The algorithm requires that a step size λ be specified for each of the three parameters. Selecting a large λ will converge faster but may risk overstepping the minimum. Selecting a small step means converging very slowly. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
74 Choosing the step size The algorithm requires that a step size λ be specified for each of the three parameters. Selecting a large λ will converge faster but may risk overstepping the minimum. Selecting a small step means converging very slowly. In this example the same value will be chosen, so λ 1 = λ 2 = λ 3 = 1. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
75 Choosing initial values Initial values for the rules must be designated. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
76 Choosing initial values Initial values for the rules must be designated. For the first rule, we choose x 1 1,x1 2,y1 as the input and output membership centers. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
77 Choosing initial values Initial values for the rules must be designated. For the first rule, we choose x 1 1,x1 2,y1 as the input and output membership centers. For the second rule, we choose x 2 1,x2 2,y2 as the input and output membership centers. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
78 Choosing initial values Initial values for the rules must be designated. For the first rule, we choose x 1 1,x1 2,y1 as the input and output membership centers. For the second rule, we choose x 2 1,x2 2,y2 as the input and output membership centers. Select spread equals to 1. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
79 Choosing initial values Initial values for the rules must be designated. For the first rule, we choose x 1 1,x1 2,y1 as the input and output membership centers. For the second rule, we choose x 2 1,x2 2,y2 as the input and output membership centers. Select spread equals to 1. These values correspond to the zero time step (k = 0). Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
80 Choosing initial values [ c 1 1 (0) c 1 2 (0) [ c 2 1 (0) c 2 2 (0) Rule1 ] = Rule2 ] = [ 0 2 [ 2 4 ] [ σ 1 1 (0) σ 1 2 (0) ] [ σ 2 1 (0) σ 2 2 (0) ] = ] = [ 1 1 [ 1 1 ] ] b 1 (0) = 1 b 2 (0) = 5 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
81 Plotting initial values 1 µ(x 1 ) 0.5 c c x 1 1 µ(x 2 ) 0.5 c 21 c x 2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
82 Calculating predicted outputs Calculate the membership values of the implication of each rule using: ( ) n µ i (x m,k = 0) = exp 1 x m j cj i 2 (k = 0) 2 σj i (k = 0) j=1 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
83 Calculating predicted outputs Calculate the membership values of the implication of each rule using: ( ) n µ i (x m,k = 0) = exp 1 x m j cj i 2 (k = 0) 2 σj i (k = 0) j=1 Calculate the outputs using (defuzzification): f(x m θ(k = 0)) = R i=1 b i(0)µ i (x m,k = 0) µ i (x m,k = 0) Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
84 Membership degrees rule 1 [ µ 1 (x 1,0) = exp 1 ( ) ] [ exp 1 ( ) ] = [ µ 1 (x 2,0) = exp 1 ( ) ] [ exp 1 ( ) ] = [ µ 1 (x 3,0) = exp 1 ( ) ] [ exp 1 ( ) ] = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
85 Membership degrees rule 2 [ µ 2 (x 1,0) = exp 1 ( ) ] [ exp 1 ( ) ] = [ µ 2 (x 2,0) = exp 1 ( ) ] [ exp 1 ( ) ] = [ µ 2 (x 3,0) = exp 1 ( ) ] [ exp 1 ( ) ] = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
86 Defuzzification f(x 1 θ(0)) = b 1(0) µ 1 (x 1,0)+b 2 (0) µ 2 (x 1,0) µ 1 (x 1,0)+µ 2 (x 1,0) f(x 1 θ(0)) = f(x 1 θ(0)) = f(x 2 θ(0)) = b 1(0) µ 1 (x 2,0)+b 2 (0) µ 2 (x 2,0) µ 1 (x 2,0)+µ 2 (x 2,0) f(x θ(0)) = f(x 2 θ(0)) = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
87 Defuzzification f(x 3 θ(0)) = b 1(0) µ 1 (x 3,0)+b 2 (0) µ 2 (x 3,0) µ 1 (x 3,0)+µ 2 (x 3,0) f(x 3 θ(0)) = f(x 3 θ(0)) = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
88 Calculating erros e m = 1 2 [f(xm θ(k = 0)) y m ] 2 e 1 = 1 2 [ ]2 = e 2 = 1 2 [ ]2 = e 3 = 1 2 [ ]2 = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
89 Calculating erros e m = 1 2 [f(xm θ(k = 0)) y m ] 2 e 1 = 1 2 [ ]2 = e 2 = 1 2 [ ]2 = e 3 = 1 2 [ ]2 = The first two data points are mapped better than the third. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
90 Calculating erros e m = 1 2 [f(xm θ(k = 0)) y m ] 2 e 1 = 1 2 [ ]2 = e 2 = 1 2 [ ]2 = e 3 = 1 2 [ ]2 = The first two data points are mapped better than the third. The result can be improved by cycling through the model. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
91 Calculating erros e m = 1 2 [f(xm θ(k = 0)) y m ] 2 e 1 = 1 2 [ ]2 = e 2 = 1 2 [ ]2 = e 3 = 1 2 [ ]2 = The first two data points are mapped better than the third. The result can be improved by cycling through the model. The GM will update the rule-base parameters b i,c i j and σ i j using the first time step. Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
92 Updating... ǫ m (k = 0) = f(x m θ(k = 0)) y m ǫ 1 (0) = = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
93 Updating b i b i (k) = b i (k 1) λ 1 (ǫ k (k 1)) µ i (x k,k 1) R i=1 µ i(x k,k 1) µ 1 (x 1,0) b 1 (1) = b 1 (0) λ 1 (ǫ 1 (0)) µ 1 (x 1,0)+µ 2 (x 1,0) ( ) 1 = 1 1 ( ) = µ 2 (x 1,0) b 2 (1) = b 2 (0) λ 1 (ǫ 1 (0)) µ 1 (x 1,0)+µ 2 (x 1,0) ( ) = 5 1 ( ) = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
94 Updating c 1 j [ ] cj i (k) = ci j (k 1) λ b i (k 1) f(x k θ(k 1)) 2(ǫ k (k 1)) R i=1 µ i(x k,k 1) ( ) x k µ i (x k j cj i (k 1),k 1) (σj i (k 1))2 ( c1 1 (1) = c1 1 (0) 1ǫ b1 (0) f(x 1 ) ( θ(0)) x 1(0) µ 1 (x 1,0)+µ 2 (x 1 µ 1 (x 1 1,0) 1 c1 1(0) ),0) (σ1 1(0))2 c1 1 (1) = 0 ( c2 1 (1) = c2 1 (0) 1ǫ b1 (0) f(x 1 ) ( θ(0)) x 1(0) µ 1 (x 1,0)+µ 2 (x 1 µ 2 (x 1 1,0) 2 c2 1(0) ),0) (σ2 1(0))2 c 1 2 (1) = 2 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
95 Updating c 2 j ( c1 2 (1) = c2 1 (0) 1ǫ b2 (0) f(x 1 ) ( θ(0)) x 1(0) µ 1 (x 1,0)+µ 2 (x 1 µ 2 (x 1 1,0) 1 c1 2(0) ),0) (σ1 2(0))2 c1 1 (1) = ( c2 2 (1) = c2 2 (0) 1ǫ b2 (0) f(x 1 ) ( θ(0)) x 1(0) µ 1 (x 1,0)+µ 2 (x 1 µ 2 (x 1 1,0) 2 c2 2(0) ),0) (σ2 2(0))2 c2 2 (1) = Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
96 Updating σ i j [ ] σj i (k) = σi j (k 1) λ b i (k 1) f(x k θ(k 1)) 3(ǫ k (k 1)) R i=1 µ i(x k,k 1) ( ) (x k j cj i (k 1))2 σ 1 1 (1) = 1 σ 1 2 (1) = 1 µ i (x k,k 1) σ1 2 (1) = σ2(1) 2 = (σj i (k 1))3 Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
97 The End Adriano Cruz (PPGI-UFRJ) Gradient Method 09/ / 41
Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler
Machine Learning and Data Mining Regression Problem (adapted from) Prof. Alexander Ihler Overview Regression Problem Definition and define parameters ϴ. Prediction using ϴ as parameters Measure the error
More informationTwo Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Department of Industrial Engineering and Management Sciences Northwestern University September 15th, 2014
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationLeast-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
More informationA Fuzzy System Approach of Feed Rate Determination for CNC Milling
A Fuzzy System Approach of Determination for CNC Milling Zhibin Miao Department of Mechanical and Electrical Engineering Heilongjiang Institute of Technology Harbin, China e-mail:miaozhibin99@yahoo.com.cn
More informationAbout the NeuroFuzzy Module of the FuzzyTECH5.5 Software
About the NeuroFuzzy Module of the FuzzyTECH5.5 Software Ágnes B. Simon, Dániel Biró College of Nyíregyháza, Sóstói út 31, simona@nyf.hu, bibby@freemail.hu Abstract: Our online edition of the software
More informationMaximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
More informationUNDERSTANDING THE DEPENDENT-SAMPLES t TEST
UNDERSTANDING THE DEPENDENT-SAMPLES t TEST A dependent-samples t test (a.k.a. matched or paired-samples, matched-pairs, samples, or subjects, simple repeated-measures or within-groups, or correlated groups)
More informationCORRELATION ANALYSIS
CORRELATION ANALYSIS Learning Objectives Understand how correlation can be used to demonstrate a relationship between two factors. Know how to perform a correlation analysis and calculate the coefficient
More informationABSTRACT. Keyword double rotary inverted pendulum, fuzzy logic controller, nonlinear system, LQR, MATLAB software 1 PREFACE
DESIGN OF FUZZY LOGIC CONTROLLER FOR DOUBLE ROTARY INVERTED PENDULUM Dyah Arini, Dr.-Ing. Ir. Yul Y. Nazaruddin, M.Sc.DIC, Dr. Ir. M. Rohmanuddin, MT. Physics Engineering Department Institut Teknologi
More informationα α λ α = = λ λ α ψ = = α α α λ λ ψ α = + β = > θ θ β > β β θ θ θ β θ β γ θ β = γ θ > β > γ θ β γ = θ β = θ β = θ β = β θ = β β θ = = = β β θ = + α α α α α = = λ λ λ λ λ λ λ = λ λ α α α α λ ψ + α =
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationTHE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan
More informationSolutions to Homework 5
Solutions to Homework 5 1. Let z = f(x, y) be a twice continously differentiable function of x and y. Let x = r cos θ and y = r sin θ be the equations which transform polar coordinates into rectangular
More informationProject Management Efficiency A Fuzzy Logic Approach
Project Management Efficiency A Fuzzy Logic Approach Vinay Kumar Nassa, Sri Krishan Yadav Abstract Fuzzy logic is a relatively new technique for solving engineering control problems. This technique can
More informationBayesian Classifier for a Gaussian Distribution, Decision Surface Equation, with Application
Iraqi Journal of Statistical Science (18) 2010 p.p. [35-58] Bayesian Classifier for a Gaussian Distribution, Decision Surface Equation, with Application ABSTRACT Nawzad. M. Ahmad * Bayesian decision theory
More informationHow To Solve The Cluster Algorithm
Cluster Algorithms Adriano Cruz adriano@nce.ufrj.br 28 de outubro de 2013 Adriano Cruz adriano@nce.ufrj.br () Cluster Algorithms 28 de outubro de 2013 1 / 80 Summary 1 K-Means Adriano Cruz adriano@nce.ufrj.br
More informationIn this section, we will consider techniques for solving problems of this type.
Constrained optimisation roblems in economics typically involve maximising some quantity, such as utility or profit, subject to a constraint for example income. We shall therefore need techniques for solving
More informationExact Confidence Intervals
Math 541: Statistical Theory II Instructor: Songfeng Zheng Exact Confidence Intervals Confidence intervals provide an alternative to using an estimator ˆθ when we wish to estimate an unknown parameter
More informationConfidence Intervals for Exponential Reliability
Chapter 408 Confidence Intervals for Exponential Reliability Introduction This routine calculates the number of events needed to obtain a specified width of a confidence interval for the reliability (proportion
More informationAC 2012-4561: MATHEMATICAL MODELING AND SIMULATION US- ING LABVIEW AND LABVIEW MATHSCRIPT
AC 2012-4561: MATHEMATICAL MODELING AND SIMULATION US- ING LABVIEW AND LABVIEW MATHSCRIPT Dr. Nikunja Swain, South Carolina State University Nikunja Swain is a professor in the College of Science, Mathematics,
More informationSection 12.6: Directional Derivatives and the Gradient Vector
Section 26: Directional Derivatives and the Gradient Vector Recall that if f is a differentiable function of x and y and z = f(x, y), then the partial derivatives f x (x, y) and f y (x, y) give the rate
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationSECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the
More informationCSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
More informationFigure 1. Experimental setup for the proposed MLR-IPSRR system.
Figure 1. Experimental setup for the proposed MLR-IPSRR system. 99 The hardware included: A Fadal vertical CNC milling machine with multiple tool changing and a 15 HP spindle. A Kistler 9257B type dynamometer
More informationWHERE DOES THE 10% CONDITION COME FROM?
1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay
More informationLecture 8 February 4
ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt
More informationIntroduction to Fuzzy Control
Introduction to Fuzzy Control Marcelo Godoy Simoes Colorado School of Mines Engineering Division 1610 Illinois Street Golden, Colorado 80401-1887 USA Abstract In the last few years the applications of
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationUniversity of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
More informationWhat does the number m in y = mx + b measure? To find out, suppose (x 1, y 1 ) and (x 2, y 2 ) are two points on the graph of y = mx + b.
PRIMARY CONTENT MODULE Algebra - Linear Equations & Inequalities T-37/H-37 What does the number m in y = mx + b measure? To find out, suppose (x 1, y 1 ) and (x 2, y 2 ) are two points on the graph of
More informationFinancial Mathematics and Simulation MATH 6740 1 Spring 2011 Homework 2
Financial Mathematics and Simulation MATH 6740 1 Spring 2011 Homework 2 Due Date: Friday, March 11 at 5:00 PM This homework has 170 points plus 20 bonus points available but, as always, homeworks are graded
More informationFuzzy Candlestick Approach to Trade S&P CNX NIFTY 50 Index using Engulfing Patterns
Fuzzy Candlestick Approach to Trade S&P CNX NIFTY 50 Index using Engulfing Patterns Partha Roy 1, Sanjay Sharma 2 and M. K. Kowar 3 1 Department of Computer Sc. & Engineering 2 Department of Applied Mathematics
More informationMath 2400 - Numerical Analysis Homework #2 Solutions
Math 24 - Numerical Analysis Homework #2 Solutions 1. Implement a bisection root finding method. Your program should accept two points, a tolerance limit and a function for input. It should then output
More informationdegrees of freedom and are able to adapt to the task they are supposed to do [Gupta].
1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very
More informationjorge s. marques image processing
image processing images images: what are they? what is shown in this image? What is this? what is an image images describe the evolution of physical variables (intensity, color, reflectance, condutivity)
More informationTemporal Difference Learning in the Tetris Game
Temporal Difference Learning in the Tetris Game Hans Pirnay, Slava Arabagi February 6, 2009 1 Introduction Learning to play the game Tetris has been a common challenge on a few past machine learning competitions.
More informationZeros of Polynomial Functions
Zeros of Polynomial Functions Objectives: 1.Use the Fundamental Theorem of Algebra to determine the number of zeros of polynomial functions 2.Find rational zeros of polynomial functions 3.Find conjugate
More informationNeural network software tool development: exploring programming language options
INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006
More informationChapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network
Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network Qian Wu, Yahui Wang, Long Zhang and Li Shen Abstract Building electrical system fault diagnosis is the
More informationApplications of improved grey prediction model for power demand forecasting
Energy Conversion and Management 44 (2003) 2241 2249 www.elsevier.com/locate/enconman Applications of improved grey prediction model for power demand forecasting Che-Chiang Hsu a, *, Chia-Yon Chen b a
More informationJAVA FUZZY LOGIC TOOLBOX FOR INDUSTRIAL PROCESS CONTROL
JAVA FUZZY LOGIC TOOLBOX FOR INDUSTRIAL PROCESS CONTROL Bruno Sielly J. Costa, Clauber G. Bezerra, Luiz Affonso H. G. de Oliveira Instituto Federal de Educação Ciência e Tecnologia do Rio Grande do Norte
More informationThe Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem
More informationDERIVATIVES AS MATRICES; CHAIN RULE
DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we
More informationA Game-Theoretical Approach for Designing Market Trading Strategies
A Game-Theoretical Approach for Designing Market Trading Strategies Garrison W. Greenwood and Richard Tymerski Abstract Investors are always looking for good stock market trading strategies to maximize
More informationA Fuzzy-Based Speed Control of DC Motor Using Combined Armature Voltage and Field Current
3rd IFAC International Conference on Intelligent Control and Automation Science. A Fuzzy-Based Speed Control of DC Motor Using Combined Armature Voltage and Field Current A. A. Sadiq* G. A. Bakare* E.
More informationSection 6.1 Joint Distribution Functions
Section 6.1 Joint Distribution Functions We often care about more than one random variable at a time. DEFINITION: For any two random variables X and Y the joint cumulative probability distribution function
More information4 Sums of Random Variables
Sums of a Random Variables 47 4 Sums of Random Variables Many of the variables dealt with in physics can be expressed as a sum of other variables; often the components of the sum are statistically independent.
More informationLecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
More informationMATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...
MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationKnowledge Base and Inference Motor for an Automated Management System for developing Expert Systems and Fuzzy Classifiers
Knowledge Base and Inference Motor for an Automated Management System for developing Expert Systems and Fuzzy Classifiers JESÚS SÁNCHEZ, FRANCKLIN RIVAS, JOSE AGUILAR Postgrado en Ingeniería de Control
More informationCS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
More informationWeek 2: Exponential Functions
Week 2: Exponential Functions Goals: Introduce exponential functions Study the compounded interest and introduce the number e Suggested Textbook Readings: Chapter 4: 4.1, and Chapter 5: 5.1. Practice Problems:
More informationFast Generation of Implied Volatility Surface for Exchange-Traded Stock Options
, March 13-15, 2013, Hong Kong Fast Generation of Implied Volatility Surface for Exchange-Traded Stock Options Nan Zhang and Ka Lok Man Abstract We present an algorithm and its software implementation
More informationPractice problems for Homework 11 - Point Estimation
Practice problems for Homework 11 - Point Estimation 1. (10 marks) Suppose we want to select a random sample of size 5 from the current CS 3341 students. Which of the following strategies is the best:
More informationNonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,
More informationLesson 20. Probability and Cumulative Distribution Functions
Lesson 20 Probability and Cumulative Distribution Functions Recall If p(x) is a density function for some characteristic of a population, then Recall If p(x) is a density function for some characteristic
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More information1 The Brownian bridge construction
The Brownian bridge construction The Brownian bridge construction is a way to build a Brownian motion path by successively adding finer scale detail. This construction leads to a relatively easy proof
More informationA QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS
A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors
More informationCritical points of once continuously differentiable functions are important because they are the only points that can be local maxima or minima.
Lecture 0: Convexity and Optimization We say that if f is a once continuously differentiable function on an interval I, and x is a point in the interior of I that x is a critical point of f if f (x) =
More informationSTT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables
Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables Discrete vs. continuous random variables Examples of continuous distributions o Uniform o Exponential o Normal Recall: A random
More information250325 - METNUMER - Numerical Methods
Coordinating unit: 250 - ETSECCPB - Barcelona School of Civil Engineering Teaching unit: 751 - ECA - Department of Civil and Environmental Engineering Academic year: Degree: 2015 BACHELOR'S DEGREE IN GEOLOGICAL
More informationA New Image Edge Detection Method using Quality-based Clustering. Bijay Neupane Zeyar Aung Wei Lee Woon. Technical Report DNA #2012-01.
A New Image Edge Detection Method using Quality-based Clustering Bijay Neupane Zeyar Aung Wei Lee Woon Technical Report DNA #2012-01 April 2012 Data & Network Analytics Research Group (DNA) Computing and
More information1 Sufficient statistics
1 Sufficient statistics A statistic is a function T = rx 1, X 2,, X n of the random sample X 1, X 2,, X n. Examples are X n = 1 n s 2 = = X i, 1 n 1 the sample mean X i X n 2, the sample variance T 1 =
More informationProblems often have a certain amount of uncertainty, possibly due to: Incompleteness of information about the environment,
Uncertainty Problems often have a certain amount of uncertainty, possibly due to: Incompleteness of information about the environment, E.g., loss of sensory information such as vision Incorrectness in
More informationLecture 2: Universality
CS 710: Complexity Theory 1/21/2010 Lecture 2: Universality Instructor: Dieter van Melkebeek Scribe: Tyson Williams In this lecture, we introduce the notion of a universal machine, develop efficient universal
More informationSummary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
More informationMath 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
More informationDescriptive Statistics. Purpose of descriptive statistics Frequency distributions Measures of central tendency Measures of dispersion
Descriptive Statistics Purpose of descriptive statistics Frequency distributions Measures of central tendency Measures of dispersion Statistics as a Tool for LIS Research Importance of statistics in research
More informationAn Overview of Integer Factoring Algorithms. The Problem
An Overview of Integer Factoring Algorithms Manindra Agrawal IITK / NUS The Problem Given an integer n, find all its prime divisors as efficiently as possible. 1 A Difficult Problem No efficient algorithm
More informationTime Series and Forecasting
Chapter 22 Page 1 Time Series and Forecasting A time series is a sequence of observations of a random variable. Hence, it is a stochastic process. Examples include the monthly demand for a product, the
More informationDynamic intelligent cleaning model of dirty electric load data
Available online at www.sciencedirect.com Energy Conversion and Management 49 (2008) 564 569 www.elsevier.com/locate/enconman Dynamic intelligent cleaning model of dirty electric load data Zhang Xiaoxing
More informationNTC Project: S01-PH10 (formerly I01-P10) 1 Forecasting Women s Apparel Sales Using Mathematical Modeling
1 Forecasting Women s Apparel Sales Using Mathematical Modeling Celia Frank* 1, Balaji Vemulapalli 1, Les M. Sztandera 2, Amar Raheja 3 1 School of Textiles and Materials Technology 2 Computer Information
More informationPoint Biserial Correlation Tests
Chapter 807 Point Biserial Correlation Tests Introduction The point biserial correlation coefficient (ρ in this chapter) is the product-moment correlation calculated between a continuous random variable
More informationNeural Network-Based Tool Breakage Monitoring System for End Milling Operations
Journal of Industrial Technology Volume 6, Number 2 Februrary 2000 to April 2000 www.nait.org Volume 6, Number 2 - February 2000 to April 2000 Neural Network-Based Tool Breakage Monitoring System for End
More informationA New Approach For Estimating Software Effort Using RBFN Network
IJCSNS International Journal of Computer Science and Network Security, VOL.8 No.7, July 008 37 A New Approach For Estimating Software Using RBFN Network Ch. Satyananda Reddy, P. Sankara Rao, KVSVN Raju,
More informationA subjective job scheduler based on a backpropagation neural network
Gopalakrishnan Human-centric Computing and Information Sciences 2013, 3:17 RESEARCH Open Access A subjective job scheduler based on a backpropagation neural network Anilkumar Kothalil Gopalakrishnan Correspondence:
More informationCoefficient of Determination
Coefficient of Determination The coefficient of determination R 2 (or sometimes r 2 ) is another measure of how well the least squares equation ŷ = b 0 + b 1 x performs as a predictor of y. R 2 is computed
More informationClass Meeting # 1: Introduction to PDEs
MATH 18.152 COURSE NOTES - CLASS MEETING # 1 18.152 Introduction to PDEs, Fall 2011 Professor: Jared Speck Class Meeting # 1: Introduction to PDEs 1. What is a PDE? We will be studying functions u = u(x
More informationv w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.
3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with
More informationNonlinear Regression:
Zurich University of Applied Sciences School of Engineering IDP Institute of Data Analysis and Process Design Nonlinear Regression: A Powerful Tool With Considerable Complexity Half-Day : Improved Inference
More informationMetrics on SO(3) and Inverse Kinematics
Mathematical Foundations of Computer Graphics and Vision Metrics on SO(3) and Inverse Kinematics Luca Ballan Institute of Visual Computing Optimization on Manifolds Descent approach d is a ascent direction
More informationMAT12X Intermediate Algebra
MAT12X Intermediate Algebra Workshop I - Exponential Functions LEARNING CENTER Overview Workshop I Exponential Functions of the form y = ab x Properties of the increasing and decreasing exponential functions
More informationLogistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
More informationConfidence Intervals for Cpk
Chapter 297 Confidence Intervals for Cpk Introduction This routine calculates the sample size needed to obtain a specified width of a Cpk confidence interval at a stated confidence level. Cpk is a process
More informationMATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables
MATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables Tony Pourmohamad Department of Mathematics De Anza College Spring 2015 Objectives By the end of this set of slides,
More information4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
More informationDuring the analysis of cash flows we assume that if time is discrete when:
Chapter 5. EVALUATION OF THE RETURN ON INVESTMENT Objectives: To evaluate the yield of cash flows using various methods. To simulate mathematical and real content situations related to the cash flow management
More informationUNIVERSITY OF BOLTON SCHOOL OF ENGINEERING MS SYSTEMS ENGINEERING AND ENGINEERING MANAGEMENT SEMESTER 1 EXAMINATION 2015/2016 INTELLIGENT SYSTEMS
TW72 UNIVERSITY OF BOLTON SCHOOL OF ENGINEERING MS SYSTEMS ENGINEERING AND ENGINEERING MANAGEMENT SEMESTER 1 EXAMINATION 2015/2016 INTELLIGENT SYSTEMS MODULE NO: EEM7010 Date: Monday 11 th January 2016
More informationUsing the Delta Method to Construct Confidence Intervals for Predicted Probabilities, Rates, and Discrete Changes
Using the Delta Method to Construct Confidence Intervals for Predicted Probabilities, Rates, Discrete Changes JunXuJ.ScottLong Indiana University August 22, 2005 The paper provides technical details on
More information12.5: CHI-SQUARE GOODNESS OF FIT TESTS
125: Chi-Square Goodness of Fit Tests CD12-1 125: CHI-SQUARE GOODNESS OF FIT TESTS In this section, the χ 2 distribution is used for testing the goodness of fit of a set of data to a specific probability
More informationA FUZZY LOGIC APPROACH FOR SALES FORECASTING
A FUZZY LOGIC APPROACH FOR SALES FORECASTING ABSTRACT Sales forecasting proved to be very important in marketing where managers need to learn from historical data. Many methods have become available for
More information= δx x + δy y. df ds = dx. ds y + xdy ds. Now multiply by ds to get the form of the equation in terms of differentials: df = y dx + x dy.
ERROR PROPAGATION For sums, differences, products, and quotients, propagation of errors is done as follows. (These formulas can easily be calculated using calculus, using the differential as the associated
More informationMicroeconomic Theory: Basic Math Concepts
Microeconomic Theory: Basic Math Concepts Matt Van Essen University of Alabama Van Essen (U of A) Basic Math Concepts 1 / 66 Basic Math Concepts In this lecture we will review some basic mathematical concepts
More informationFeb 28 Homework Solutions Math 151, Winter 2012. Chapter 6 Problems (pages 287-291)
Feb 8 Homework Solutions Math 5, Winter Chapter 6 Problems (pages 87-9) Problem 6 bin of 5 transistors is known to contain that are defective. The transistors are to be tested, one at a time, until the
More informationLeran Wang and Tom Kazmierski {lw04r,tjk}@ecs.soton.ac.uk
BMAS 2005 VHDL-AMS based genetic optimization of a fuzzy logic controller for automotive active suspension systems Leran Wang and Tom Kazmierski {lw04r,tjk}@ecs.soton.ac.uk Outline Introduction and system
More informationLecture 8. Generating a non-uniform probability distribution
Discrete outcomes Lecture 8 Generating a non-uniform probability distribution Last week we discussed generating a non-uniform probability distribution for the case of finite discrete outcomes. An algorithm
More information