Matching Law Example

This forum made possible through the generous support of SDN members, donors, and sponsors. Thank you.
Joined
Aug 31, 2011
Messages
2,464
Reaction score
4,261
In another thread (Adult autism assessment), @borne_before requested an example of the matching law (the formula that appears in my account info here). In an effort to not derail that thread more, I'm replying in this new thread. What follows is a long (and- to most of you- boring explanation/example of key ABA principle).

Herrnstein's seminal work on this (https://doi.org/10.1901/jeab.1961.4-267) from 1961 was done with pigeons. Basically, the pigeons could peck a red or white key. Concurrent schedules of reinforcement were set up whereby- for example- pecking the red key would get you would generally result in getting twice as many food pellets as pecking the white key. He found that under such schedules, the pigeons pecked the red key twice as much as they pecked the red key. Filling this into the equation in my signature, you get:

Number peck red Number food pellet for red
--------------------- = ---------------------------------------
Number peck white Number of food pellets for white

Altering the schedules of reinforcement (he used different variable interval schedules for each key) led to alterations in the number of pecks to each different color key that kept the equation true. Note that he controlled for position/location bias (key were right next to each other), as well as for chance that the bird would just learn to alternate between the keys (e.g., there was a minimum time after pecking one key until the other key could "pay off" with a food pellet.

Subsequent studies (c.f. Baum 1974 https://doi.org/10.1901/jeab.1974.22-231 have shown that this "simplified" matching law does not account for all variance in behavior, as is does not also account for bias (e.g. if a lot of effort was needed to switch between keys, the bird would learn that pecking just the red would lead to higher rates), as well as sensitivity to the differences in the scheduled. This led to to revise the formula to the generalized matching law:
1738949724894.png


R=Response parameter (frequency, duration, etc.)
r= Reinforcement parameter
k= bias for choosing one response over the other
a= Sensitivity of the behavior to variations in reinforcement distribution

This generalized law has been shown to account for most of the variance in animal responding und concurrent VI schedules.

For an example with humans, Hoch et. (https://doi.org/10.1901/jaba.2002.35-171) looked at increasing the amount of time that three young boys (ages 9-11) chose to play in an area where a sibling or peer was present vs. in another area by themselves. By altering the magnitude of reinforcement (i.e., duration of access toys) and the quality or reinforcement (access to highly preferred vs. less highly preferred toys), the boys were more likely to choose play with peers/siblings than to play alone. The important thing here is that there was no aversive or time-out from reinforcement necessary- the children could- and sometimes did- choose to play alone. They just chose to play with the other kids more.

In a more "personal" example- if you like both beer and bourbon, but like beer twice as much as you like bourbon, you will choose to drink beer on 4 out of six opportunities and bourbon on 2 out of six (assuming things like they both cost the same and are just as easy to get ahold of).

Members don't see this ad.
 
Only tangentially related, but I find it interesting how hard it is for humans to "manually override" this.

I've administered a variety of choice tasks where stimuli are rewarded at different frequencies. The only reinforcer is feedback (Correct vs Incorrect). A large portion of participants (i.e., the majority) will exhibit explore/exploit behavior that is essentially proportional to the probability of receiving positive feedback even with very explicit instructions to always choose the stimuli that is rewarded most frequently and probabilities that are easy to learn (i.e., 90/10). Spoken to multiple other labs with the same observation. There is something about matching law that seems unusually difficult to override. Haven't had the sample for it yet, but am always curious how those who can differ from those who don't.

Some research ideas for any trainees here in the computational psychiatry space😉
 
Only tangentially related, but I find it interesting how hard it is for humans to "manually override" this.

I've administered a variety of choice tasks where stimuli are rewarded at different frequencies. The only reinforcer is feedback (Correct vs Incorrect). A large portion of participants (i.e., the majority) will exhibit explore/exploit behavior that is essentially proportional to the probability of receiving positive feedback even with very explicit instructions to always choose the stimuli that is rewarded most frequently and probabilities that are easy to learn (i.e., 90/10). Spoken to multiple other labs with the same observation. There is something about matching law that seems unusually difficult to override. Haven't had the sample for it yet, but am always curious how those who can differ from those who don't.

Some research ideas for any trainees here in the computational psychiatry space😉
Begs the question of whether it can actually be overridden? You can alter k (bias) or a (sensitivity) and get changes in the R1/R2 and r1/r2 ratios, but within parameters set by k and a, the match still holds.

As far as being able to "manually override" this stuff, I'd venture to say that any organism that did not have the innate tendency to maximize reinforcement while minimizing effort probably didn't stick around long enough for it's genes to get us organisms that are around today.
 
Members don't see this ad :)
Begs the question of whether it can actually be overridden? You can alter k (bias) or a (sensitivity) and get changes in the R1/R2 and r1/r2 ratios, but within parameters set by k and a, the match still holds.

As far as being able to "manually override" this stuff, I'd venture to say that any organism that did not have the innate tendency to maximize reinforcement while minimizing effort probably didn't stick around long enough for it's genes to get us organisms that are around today.
On the issue of over-riding basic behavioral drives, my family had a dog who had no taste aversion—in fact, it was almost the opposite. The sicker something made her, the more she would seek it out and eat it. Didn’t appear to be any secondary reinforcement, either. As a behaviorist, it was weird as hell.
 
Well, there are an enormous number of innate tendencies necessary for survival we can manually override, particularly in constrained laboratory analogue settings.

Some people - albeit I think its actually a minority in my studies - clearly seem able to manually override it. But again, in an artificial laboratory context. I haven't yet played around to see if inclusion of performance incentives changes behavior (i.e., are the people "overriding" it the people who simply don't find feedback alone meaningfully reinforcing? Distinctly possible in which case r = 0 and the equation is undefined). I suspect that explains at least a subset of the cases.
 
Well, there are an enormous number of innate tendencies necessary for survival we can manually override, particularly in constrained laboratory analogue settings.

Some people - albeit I think its actually a minority in my studies - clearly seem able to manually override it. But again, in an artificial laboratory context. I haven't yet played around to see if inclusion of performance incentives changes behavior (i.e., are the people "overriding" it the people who simply don't find feedback alone meaningfully reinforcing? Distinctly possible in which case r = 0 and the equation is undefined). I suspect that explains at least a subset of the cases.
Matching law only applies in situation with concurrent schedules of reinforcement. If r=0, there's no schedule of reinforcement, so matching law does not apply. Also trickier to do these experiments with humans- our ability to represent and learn from contingencies we haven't experienced (damn you, language and all those derived relationships between stimuli!) Really influences the values of k and a in the generalized law.
 
Last edited:
1) @ClinicalABA What’s that one learning formulae for competing or sequential reinforcement schedules, where the author died of breast cancer? I can’t remember, and you seem like you would know.

2) it would be extremely interesting to get access to the slots data from Vegas. Millions of subjects, untold numbers of trials, likely linked to demographics. There’s no way they don’t have models that exceed anything academic.
 
Only tangentially related, but I find it interesting how hard it is for humans to "manually override" this.

I've administered a variety of choice tasks where stimuli are rewarded at different frequencies. The only reinforcer is feedback (Correct vs Incorrect). A large portion of participants (i.e., the majority) will exhibit explore/exploit behavior that is essentially proportional to the probability of receiving positive feedback even with very explicit instructions to always choose the stimuli that is rewarded most frequently and probabilities that are easy to learn (i.e., 90/10). Spoken to multiple other labs with the same observation. There is something about matching law that seems unusually difficult to override. Haven't had the sample for it yet, but am always curious how those who can differ from those who don't.

Some research ideas for any trainees here in the computational psychiatry space😉
Can you please give me a real world example of this like @ClinicalABA's bourbon beer example?

Sorry, but I just read Surely You're Joking by Richard Faynman and I think a lesson from the book is to make physicals real. Also, I haven't read the other posts yet.
 
To clarify - you are asking for a real world example of the specific instructional manipulation I'm referencing? The task itself? All of the above? It seems a little contrived for a lab task, but I can try - just want to make sure I know what you are looking for before I do it.
 
1) @ClinicalABA What’s that one learning formulae for competing or sequential reinforcement schedules, where the author died of breast cancer? I can’t remember, and you seem like you would know.
I've checked my textbook, done lit searches, etc., and am just not able to find anything to answer this? Any more details you remember?
 
Top