RESCORLA-WAGNER MODEL



Similar documents
Programmed Learning Review

Learning Theories 4- Behaviorism

COMPUTATIONAL MODELS OF CLASSICAL CONDITIONING: A COMPARATIVE STUDY

Behavior Analysis and Strategy Application after Brain Injury: Addressing the long-term behavioral outcomes of brain injury

I d ( r; MPK f, τ) Y < C d +I d +G

Learning. Relatively permanent behavior change that is acquired through experience

Chapter 7 Conditioning and Learning

Understanding Equilibrium in the IS/LM Model

GCSE PSYCHOLOGY UNIT 2 LEARNING REVISION

Agent Simulation of Hull s Drive Theory

The normal approximation to the binomial

Gains From Trade Consumer Surplus Quantifying Welfare Effects Producer Surplus Welfare in Equilibrium. Consumer Surplus and Welfare Measurement

Classical (Pavlovian) Conditioning

Introduction to Learning. Chapter 1

Learning. Any relatively permanent change in behavior brought about by experience or practice. Permanent Experience Practice

Chapter 12 Time-Derivative Models of Pavlovian Reinforcement Richard S. Sutton Andrew G. Barto

Math 120 Final Exam Practice Problems, Form: A

EXPONENTIAL FUNCTIONS

Today. Learning. Learning. What is Learning? The Biological Basis. Hebbian Learning in Neurons

Section 1.4. Difference Equations

For additional information, see the Math Notes boxes in Lesson B.1.3 and B.2.3.

Chapter 5. Chapter 5 Lectures Outline

9/14/2015. Innate behavior. Innate behavior. Stimuli that trigger innate behaviors are called releasers.

Absolute Value Equations and Inequalities

LEARNING. Chapter 6 (Bernstein), pages

MA107 Precalculus Algebra Exam 2 Review Solutions

A. Learning Process through which experience causes permanent change in knowledge or behavior.

The normal approximation to the binomial

1 Review of Least Squares Solutions to Overdetermined Systems

Manual for SOA Exam FM/CAS Exam 2.

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

Classical Conditioning. Classical and Operant Conditioning. Basic effect. Classical Conditioning

A Model for Stimulus Generalization in Pavlovian Conditioning

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

1. Then f has a relative maximum at x = c if f(c) f(x) for all values of x in some

Descriptive Statistics

NPV Versus IRR. W.L. Silber We know that if the cost of capital is 18 percent we reject the project because the NPV

7/17/2014. Applied Behavior Analysis (ABA) Therapy Overview. Applied Behavior Analysis Therapy. Ivan Petrovich Pavlov

An Introduction to Calculus. Jackie Nicholas

A New Interpretation of Information Rate

Objective. Materials. TI-73 Calculator

Why? A central concept in Computer Science. Algorithms are ubiquitous.

A Detailed Price Discrimination Example

Learning from Experience. Definition of Learning. Psychological definition. Pavlov: Classical Conditioning

Microeconomics Sept. 16, 2010 NOTES ON CALCULUS AND UTILITY FUNCTIONS

The Graphical Method: An Example

11 Multivariate Polynomials

Maximizing volume given a surface area constraint

Appendix A: Science Practices for AP Physics 1 and 2

1 Mathematical Models of Cost, Revenue and Profit

Slope-Intercept Equation. Example

CREATING A 12 TONE MATRIX: "My music is not modern, it is merely badly played." A. Schoenberg

Bayesian probability theory

Guide to Using the Ti-nspire for Methods - The simple and the overcomplicated Version 1.5

MAT12X Intermediate Algebra

Circuits 1 M H Miller

Solving simultaneous equations using the inverse matrix

Also, compositions of an exponential function with another function are also referred to as exponential. An example would be f(x) = x.

Okami Study Guide: Chapter 7

Review of Fundamental Mathematics

Correlation key concepts:

1.7 Graphs of Functions

Characterizing Digital Cameras with the Photon Transfer Curve

Exponential Functions. Exponential Functions and Their Graphs. Example 2. Example 1. Example 3. Graphs of Exponential Functions 9/17/2014

Costs of Production and Profit Maximizing Production: 3 examples.

Chapter 6. Commodity Forwards and Futures. Question 6.1. Question 6.2

What are the place values to the left of the decimal point and their associated powers of ten?

Psychology with Mr. Duez UNIT 3 "Learning" LEARNING TARGETS

MATH10040 Chapter 2: Prime and relatively prime numbers

CHAPTER 4 Consumer Choice

Chapter 2 An Introduction to Forwards and Options

8.7 Exponential Growth and Decay

MACROECONOMICS II INFLATION, UNEMPLOYMENT AND THE PHILLIPS CURVE

1 Approximating Set Cover

22 COMPETITIVE MARKETS IN THE LONG-RUN

Monte Carlo simulations and option pricing

Adaptive testing for video quality assessment. Vlado Menkovski, MSc Eindhoven University of Technology, NL

A Brief Explanation of Applied Behavior Analysis. conditioning to identify the contingencies affecting a student s behavior and the functions of the

11. Analysis of Case-control Studies Logistic Regression

CS Course Placement Information

>

16. THE NORMAL APPROXIMATION TO THE BINOMIAL DISTRIBUTION

Graphing calculators Transparencies (optional)

Chapter 4 and 5 solutions

GCSE Psychology Learning

1 Uncertainty and Preferences

Section 4-7 Exponential and Logarithmic Equations. Solving an Exponential Equation. log log 5. log

Linear functions Increasing Linear Functions. Decreasing Linear Functions

8. Average product reaches a maximum when labor equals A) 100 B) 200 C) 300 D) 400

Performance. 13. Climbing Flight

Introduction to the Practice of Statistics Fifth Edition Moore, McCabe

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 2

Risk Analysis and Quantification

COMMUTATIVITY DEGREES OF WREATH PRODUCTS OF FINITE ABELIAN GROUPS

The operations performed to establish Pavlovian conditioned reflexes

16. Recursion. COMP 110 Prasun Dewan 1. Developing a Recursive Solution

Answers to Text Questions and Problems in Chapter 11

Interpretation of Data (IOD) Score Range

Integer Factorization using the Quadratic Sieve

MATH Fundamental Mathematics IV

Transcription:

RESCORLA-WAGNER, LearningSeminar, page 1 RESCORLA-WAGNER MODEL I. HISTORY A. Ever since Pavlov, it was assumed that any CS followed contiguously by any US would result in conditioning. B. Not true: Contingency C. Not true: Taste Aversion Learning D. Not true: Blocking II. FOUNDATION A. The foundation for the model was Kamin's work on blocking. B. Kamin explained blocking by saying that for conditioning to occur, the US must be unexpected or surprising. C. If a US was surprising, the organism will search their memory for possible causes. D. If a US was expected, then an adequate predictor must be already available; no memory search occurs and no new learning takes place. E. "Expectations" are the key. III. THE MODEL A. Rescorla-Wagner Model extended Kamin's notion of expectancy. B. The precise amount of conditioning is dependent upon the amount of surprise. C. Surprise not only determines whether conditioning occurs but also how strong it is. D. Model variables. 1. Strength of association = V 2. Change in associative strength on each trial = V 3. Asymptote = Vmax 4. Strength of conditioning at the beginning of any trial "n" = Vn 5. Surprise = (Vmax - Vn) 6. The change in the strength of the association produced by trial "n" = Vn 7. Constant representing the speed of conditioning = c V n = c(v max - V n )

RESCORLA-WAGNER, LearningSeminar, page 2 V n = c(v max - V n ) E. The learning curve. 1. On the first trial V will increase by a large amount. 2. On successive trials, the increase in V on each trial will get progressively smaller until V approaches a stable value or asymptote. 3. At the beginning of conditioning, Vn will be much less than Vmax, the subject will not expect the US and will be surprised; therefore conditioning will be strong. 4. As conditioning proceeds, the difference between Vn and Vmax will decrease and the US will be less surprising therefore conditioning will be weak. 5. Learning curves are similar -- learning increases over trials but at a declining rate until asymptote is reached. 6. But learning curves do differ -- actual asymptote and the rate at which asymptote is reached differs across experiments dependent upon the choice of CS and US. 7. Model takes actual maximal asymptote into consideration with Vmax 8. Model takes actual speed at which asymptote is reached into consideration with "c"

RESCORLA-WAGNER, LearningSeminar, page 3 F. Problems with mathematical models (e.g., Rescorla-Wagner). 1. V represents the strength of a theoretical association (not a real one). How do we relate measures of overt behavior to associative strength? 2. We need to know the values of c and Vmax before we can predict V. How do we do that? 3. Attempting to figure out real numbers for V, c, Vmax, and other similar constructs had only been attempted once in the entire history of learning theory by Hull (1943) using his mathematico-deductive model. The model was eventually discarded after more than a decade of trying because it was too cumbersome and difficult to understand. 4. Solution = it is not necessary to know the exact values to test the model; use arbitrary values instead. 5. The use of arbitrary values will not allow us to make precise quantitative predictions. For example, we cannot predict how many drops of saliva will occur after 20 pairings of a TONE (CS) with a FOOD (US). 6. The use of arbitrary values will allow us to make qualitative predictions. For example, we can predict whether the amount of saliva will increase or decrease over trials.

RESCORLA-WAGNER, LearningSeminar, page 4 G. Tests of model. 1. Conditioning a. Repeatedly pair a TONE (CS) with FOOD (US). b. Assume Vmax = 1.0 and c = 0.30. Also assume that associative strength at the beginning of trial 1 equals zero since pairing hasn't even begun yet. That is, Vn= V1 = 0.00 c. Plug in your values using the equation V n = c(v max - V n ), add the change in value using the equation V n+1 = V n + V n+1 and fill out the following table. Trial Vn V n = c(v max - V n ); V n+1 = V n + V n+1 1 0.00 V1 = 0.30 (1-0.00) = +0.30; 0.00+0.30 = 0.30 2 0.30 V2 = 0.30 (1-0.30) = +0.21; 0.30+0.21 = 0.51 3 0.51 V3 = 0.30 (1-0.51) = +0.147 0.15; 0.51+0.15 = 0.66 4 0.66 5 6 7 8 9 d. Draw a graph using your values for Vn

RESCORLA-WAGNER, LearningSeminar, page 5 2. Extinction a. After repeated pairings of the CS and US, we know that the increase in level of conditioning reached approaches 0.00 (since we'd be at asymptote no new learning will occur). b. What happens when we no longer pair the CS and US as in extinction? You must assume Vmax= 0.00 (since no new learning can take place). c. Suppose after the 3rd conditioning trial we begin to present the CS alone. Trial Vn V n = c(v max - V n ); V n+1 = V n + V n+1 1 0.00 V1 = 0.30 (1-0.00) = +0.30; 0.00+0.30 = 0.30 2 0.30 V2 = 0.30 (1-0.30) = +0.21; 0.30+0.21 = 0.51 3 0.51 V3 = 0.30 (1-0.51) = +0.147 0.15; 0.51+0.15 = 0.66 Trial Vn V n = c(v max - V n ); V n+1 = V n + V n+1 4 CS alone 0.66 V4 = 0.30 (0.00-0.66) = -0.198 0.20; 0.66-0.20 = 0.46 5 CS alone 0.46 6 CS alone 7 CS alone 8 CS alone 9 CS alone d. Draw a graph using your values.

RESCORLA-WAGNER, LearningSeminar, page 6 3. Blocking a. For an analysis of blocking, we have to take into consideration a compound CS (usually 2 stimuli instead of only one) = NOISE (CSa) and LIGHT (CSb). b. If 2 stimuli are presented together then both can predict the likelihood of the US occurring. c. Rescorla-Wagner Model assumes that if 2 stimuli (a and b) are presented together, the associative strength at the beginning of a trial would be equal to the sum of the strengths of each stimulus present. In other words Vab = Va + Vb d. Assume that CSa has an associative strength of 0.30 and CSb has an associative strength of 0.50. Vab = Va + Vb = 0.30 + 0.50 = 0.80. Therefore, the associative strength of the compound CSa and CSb would be equal to 0.80. e. The amount of conditioning that would take place if both CSa and CSb occur together would be represented by the equation Va = Vb = c(vmax - Vab) where Vab = Va + Vb f. Typical blocking experiment: teach animal to bar press for food pair NOISE (CSa) + SHOCK (US) for many trials pair compound NOISE/LIGHT (CSab) + SHOCK (US) for many trials present NOISE (CSa) while bar pressing --> RESPONSE SUPPRESSION present LIGHT (CSb) while bar pressing --> NO RESPONSE SUPPRESSION g. During the NOISE (CSa) + SHOCK (US) trials, the associative strength of the NOISE would have increased until it reached asymptote. If we assume Vmax for the SHOCK (US) = 1.0 then at the end of the NOISE (CSa) + SHOCK (US) trials, Va = 1.0. h. If the NOISE was now presented with the LIGHT (NOISE/LIGHT CSab) then their combined associative strength could be no more than 1.0 (since we are already at asymptote). g. Given that Vab = Va + Vb, and we know that CSa = 1.0 and CSab = 1.0, what is the value of CSb? h. The Rescorla-Wagner Model predicts accurately that no conditioning will occur to the LIGHT (CSb) since Vb has to equal 0.00.