[From Bruce Abbott (970923.2015 EST)]
Bill Powers (970923.1043 MDT) --
Bruce Abbott (970922.2000 EST)
A controlled perception is a variable, like the position of a cursor. It
would not be correct to call the cursor itself the controlled variable if
what is being controlled about the cursor is its position relative to a
target. The cursor is only an object, not a variable. By the same logic,
it would not be correct to call a food pellet -- an object -- the
controlled variable when what is being controlled about the pellet is its
rate of presentation, or its current state of availability (available/not
available). To do so only invites confusion.
The confusion is in the idea that a food pellet is not a variable. All
perceptions are variables in particular states.
A particular state of a variable is not a variable.
What makes the food pellet
what it is is the set of states of all variable attributes associated with
it: size, color, odor, position, shape, weight, and so forth. When you have
specified all the variable attibutes and their values, you have defined the
food pellet as it is perceived. Some of these attributes can be changed by
behavior; some, perceptually, interact (the distance of the food pellet
from the rat affects the intensity of its odor, for example).
Yes, and it is perceptually different if held in the paw, or placed in the
mouth and tasted. None of which has any relevance to the question of
whether the pellet per se, being delivered as a consequence of the action of
a control system, is a controlled variable, or simply a constant being added
to or otherwise affecting the value of the CV.
Perhaps what you have in mind is that the reinforcer is some particular
state of the controlled variable, e.g., cursor on-target or pellet
present. That is, the reinforcer is "what is wanted." But a particular
state of a controlled variable is not the same as a controlled variable,
so in this definition too, a reinforcer (a particular state of a CV)
cannot be a controlled variable. Dr. Marken loses either way.
What Rick has in mind, and what I have in mind, is that there is nothing
reinforcing about the food pellet.
That's O.K. by me, but again irrelevant to the issue. We are asking what is
this thing EABers call a reinforcer, not whether it actually "reinforces"
anything. In particular, we are asking what its role is within a
functioning control system.
The so-called reinforcer is JUST some
attribute of the food pellet, such as its perceived position or taste or
smell, that is under control by the organism.
So there is some attribute of the food pellet (such as its perceived
position) that the rat is controlling by pressing the lever. I agree. But
wait: If the rat's current reference for this attribute is in-the-food-cup,
and the current state of this attribute is in-the-feeder, then the rat will
press the lever in order to bring this attribute of the food pellet to a
perceptual state matching the reference. It presses the lever rather than
doing something else (or nothing at all) because it has learned that this
action will move the pellet into the cup. If the pellet is already in the
food cup, no action is required and we will not observe a lever-press. So
what is motivating ("reinforcing") the lever-press is not the attribute
(where the pellet is), but rather the desire on the part of the rat to
change the pellet's attribute from in-the-feeder to in-the-cup. The rat
presses the lever in order to bring the CV to its reference state, or as
close to that state as one lever-press will produce. In other words, it
presses the lever to reduce (to zero if possible) the currently existing
error between the CV and its reference state.
Thus, a reinforcer must be that which reduces error between what I have
and what I want, or in other words, between the current state of a CV and
its reference value.
This is not a general definition. If anything that reduces the difference
between a CV and its reference value is a reinforcer, then a change in the
reference value toward the perceived value of the CV is a reinforcer. Under
this definition, it makes no difference what reduces the error; there is
nothing in this definition to say that noncontingent food pellets would not
be reinforcing -- they would be just as effective in reducing error as
pellets produced by behavior. Obviously, the reduction in error per se
can't be the only criterion for defining a reinforcer.
Good point, but see my reply to Bruce Gregory on this issue [Bruce Abbott
(970923.1850 EST)]. There must be a particular relationship established
between action, reinforcer, and reduction of error if the reinforcer is to
function as such.
The food pellets consumed change from being merely food pellets to being
reinforcers only when they are produced as a consequence of behavior.
Yes.
Their
error-reducing effects are not what make the difference. If we want to find
the difference, we have to try to see what is different about the situation
when the food pellets are and are not produced by behavior. That is the
only difference that makes any difference.
Incorrect. Both are necessary: the action of the control system must
produce the reinforcer, and the effect of the reinforcer must be to reduce
the error between the CV and its reference state.
If you're trying to define a reinforcer in terms of PCT, you have to find
some unique role that the reinforcer plays that fits your definition under
all circumstances. It is necessary, but not sufficient, that the reinforcer
reduce error. It is necessary, but not sufficient, that the reinforcer be
produced by behavior.
Now you seem to be agreeing with me. In the previous paragraph you said
that the only difference that makes any difference is whether or not the
behavior produces the pellets. You can't have it both ways.
So is it necessary AND sufficient that the reinforcer
be produced by behavior AND reduce error? No, not yet, because there are
other conditions that have to be satisfied.
When you're speaking of Phase 1, food that appears because of behavior will
be followed by more of the behavior that produced it. We will observe a
concurrent increase in the number of behaviors of the required kind and the
number of deliveries of food pellets. So the necessary and sufficient
conditions are met and reinforcement does take place.
Yes, I agree.
During phases 2 and 3, however, we find that while food produced by
behavior does reduce error, any change in the contingency that results in
_more_ food being produced leads to _less_ behavior, and vice versa. So now
the purported reinforcing effect of the food has reversed. Where in Phase 1
an increase in food production went with an increase in behavior, now an
increase in food production goes with a decrease in behavior.
I disagree, because the comparison of rate of responding across schedules,
pellet sizes, etc. is not the appropriate comparison to be making.
Responding continues under _each_ of these new contingencies (rather than
extinguishing) so long as the putative reinforcer continues to be produced
by the operant and continues to reduce the error in the relevant control system.
Thus we have to conclude that food is reinforcing only under phase 1.
However, there are other explanations, such as that when food appears
during phase 1 it does not increase the freqauency with which more of the
same behavior will occur; it simply stops the continuing search for food in
different places.
Certainly. But in either case, the operant will be observed to occur more
often (up to some limit) and other behaviors will be observed to occur less
often (as more time is taken up performing the operant. Even so, I would
expect that the changes expected under these two different scenarios can be
distinguished and thus provide an experimental test betweent them.
At the start of your post, you disputed my claim that reinforcers are not
controlled variables, but rather serve to reduce error between CV and
reference when made contingent on the system's actions. At the end you seem
to be agreeing with my position. Did I misunderstand something, or did you
change your mind?
Regards,
Bruce