[From Bill Powers (951211.1945 MST)]
Bruce Abbott (951211.1830 EST) --
Replying to Mary, you say
Reinforcement as defined in EAB is an empirical fact, not a theory.
An operant is a member of a class of activities having a common
consequence (e.g., depressing a lever to the point of switch-
closure). When this consequence is linked to another (the later is
made "contingent" on the former), the frequency (or some other
aspect, depending on the contingency) of the operant is sometimes
observed to increase as a result. If so, the contingent
consequence of the operant has been demonstrated to reinforce the
operant.
What you seem to be claiming is that depressing a lever to the point of
switch closure starts out as an operant -- that is, this result is
naturally created by a variety of behaviors, and while the detailed
actions may differ, the outcome remains the same, a closure of the
contacts. And THEN, when this operant is linked to a special
consequence, reinforcement is observed to occur, as an increase in the
rate of occurence of the operant act and hence the contact closures.
In order for this description to hold up as an empirical fact, you would
have to see a rat repeatedly depressing a lever by a variety of means,
like using one paw or another, or sitting on it, or nosing it. Then you
could say that you have observed an operant: a class of actions having a
common outcome. It isn't enough to say that in principle there are many
different actions that _could_ produce the same result. You have to
demonstrate that such a variety of behaviors with a common consequence
actually takes place in the behavior of a single rat, and does so prior
to any reinforcements.
I think that what is actually observed during shaping is that normal
behavior seldom produces a lever press. When shaping is done, the
reinforcer is given for any move toward the lever, until finally the rat
accidentally depresses the lever -- upon which, the reinforcer is
immediately given. So we do not first observe the operant, and then make
a reinforcement contingent on it. We do not observe the operant unless
reinforcement is contingent upon it from the start. And in fact, what
usually happens, as I understand it, is that the _first_ action that
depresses the lever is normally used from then on ("superstitious
behavior"). So an operant is not observed at all.
I know you are trying to link the operant with controlled variables. But
this way of doing it misses a principal aspect of a controlled variable:
that when something interferes with the action that normally controls
it, behavior will _change_, it will change _immediately_, and it will
change specifically in the way needed to keep the value of the
controlled variable the same. A lever depression would be a controlled
variable if we did something to interfere with depressing it by one
means, and a different means was immediately used to depress it anyway.
This fits the definition of an operant, but it also brings in the
purposiveness of behavior. The behavior changes to a new method in
precisely the way needed to achieve the same outcome as before.
As I understand the concept of the operant, this is not its intent. The
main idea is that the animal is producing a lot of effects in its
environment, and some of the different acts happen, more or less at
random, to have a similar consequence. So when the common consequence is
made to produce a reinforcer, the operant is reinforced; it is made to
be more likely to happen. However, if something prevented that
particular act from succeeding, the animal would NOT immediately switch
to another degree, direction, or kind of behavior that has the same
consequence. If that happened, we would see the operant as creating an
_intended_ consequence, a concept that has long been rejected by
behaviorists. We would see the lever depression as a controlled
variable.
The concept of the operant was an attempt to get around the appearance
of purpose in behavior. If an animal switched from one means of pressing
a lever to another means, this was not to be taken as showing that the
animal intended that the lever be depressed; that the depression of the
lever was a goal to be achieved by variable means. What Skinner proposed
was that each different means simply happened to have a similar effect,
and that the animal did not _choose_ the means _in order to_ have the
effect. This was an explicitly-stated goal of Skinner's -- to eliminate
language that implies an active role by the organism in selecting the
consequences of its own behavior.
Why would Skinner have specifically wanted to eliminate the concept of
purpose from explanations of behavior? Because he thought it was a
mystical idea, a mentalism. Because he knew of no mechanism by which
purposive behavior could be created. The whole thrust of his concept of
the operant was to do away with the idea of constant ends being achieved
by variable means. He thought up a way in which different actions (the
variable means) could just by chance have a constant end, and saw that
as the answer to James' proposal. Because he thought that this took care
of the problem, he never recognized the cases in which the "variable
means" were _systematically_ varied, in exactly the way required to
achieve the constant end. Under the concept of the operant, if one
action fails to achieve the result, a new trial-and-error period would
have to follow, in which _any_ other action would be tried, not just
actions that have the same consequence. The animal is not permitted to
know that the consequence is what needs to be repeated and to select a
new action that will have the same consequence. That selection must be
left up to the reinforcer, with the animal emitting different responses
at random, not in systematic relationship to one particular consequence.
So if as you say reinforcement is an empirical fact, it is an empirical
fact seen through a narrow-band philosophical filter. It isn't just the
apparent fact of reinforcement that's involved here; it's the
interpretation of the supporting observations that is critical. Skinner
interpreted the variations in action in just the way needed to support
the idea that the reinforcer, not the organism, was producing the
constant consequence. If in fact the organism itself is systematically
selecting whatever action is needed to press the key and get the
reinforcer, then the idea that the reinforcer is a cause rather than an
effect is simply wrong. The airtight case springs a leak.
···
-----------------------------------------------------------------------
Best,
Bill P.