[From Bruce Abbott (2000.12.02.1525 EST)]
Bruce Gregory (2000.1202.1224) --
If operants are strengthened through reinforcement, I have always wondered
why I couldn't develop fantastic pitching control simply by throwing in the
direction of home. No matter how infrequently I got the ball over the plate
low and outside, those operants should be the ones that get strengthened
and the operants leading to wild pitches should be selected against. The
more I throw the more quickly I will develop incredible control. Yet this
doesn't seem to happen. Why?
"Strengthened" means that they come to occur more often relative to variants
that are not reinforced. According to theory, if getting the ball over the
plate low and outside is a reinforcing event for you, then variants that
accomplish this should be selected over those that fail, and the more you
throw while trying to get the ball to cross the plate at that position, the
better you should become at achieving that end. So what prevents you from
developing "incredible control"? There could be several problems -- off the
top of my head I can think of three: (1) You may have uncontrollable muscle
twitches which act as disturbances. The reinforcement process could lead to
the selection of a particular way of throwing the ball, but if the
physiological mechanisms cannot reproduce the act accurately, they cannot be
repeated accurately and throws will go wild. (2) It may be difficult for you
to sense what you were doing when the ball went where you wanted it to go.
The reinforcement mechanism cannot "select" components of behavior for
"strenghening" that cannot be distinguished from other components. (3) You
may not be producing variant operants that permit the most accurate
throwing, so of course they cannot be selected. (Some sort of "response
shaping" may be required to bring these variants into being.)
I'm guessing, of course, so at this point these explanations are only
"just-so" stories in need of empirical testing.
If you think about it, I think you'll see that essentially the same
explanations are still logical possibilities if we substitute an adaptive
control model for the reinforcement model. We know that a set of
hierarchically organized control systems are working to keep muscles
contracting at specified time-varying rates, generating certain time-varying
tensions, attempting to produce the specified motion of the arm and ball
through space and over time despite disturbances such as uncontrolled muscle
twitches, muscle fatigue, wind against the body, and so on. But how does
the system "know" what motions to produce? Generally, from experience.
Errors in ball position as it crosses the plate must somehow lead to a new
selection of reference signal values, gains, and so on -- but which ones,
and by how much should they be changed? If the "right" variations are
eventually tried, error is minimized, or in other words, throwing becomes
accurate, but only if they are tried, and only if, with the "right"
selections, the system so configured is capable of opposing all disturbances
to the extent necessary.
I have a feeling that by now most everyone has drawn the wrong conclusion
about my theoretical preferences, so I'll state them explicitly. I favor
(too weak a word) an adaptive control model over the so-called reinforcement
model. I see what has been labeled as "reinforcement" as the outward
manifestations of an adaptive process that (usually) leads to better
control. The reinforcement explanation simply describes (from the
observer's point of view) changes observed in the behavior of the system as
adaptation takes place. However, I believe that, by and large, the changes
in behavior that reinforcement theory describes as taking place under
specific conditions are real; if true, it follows that any proposed adaptive
control system model must prove capable of changing its behavior as
described in the reinforcement account if it is to be considered a
legitimate model.
This is why I stated at the outset of this discusson my belief that the
reinforcement and control-system views are not necessarily incompatible, at
least if one sticks to the purely descriptive version of reinforcement
theory. The latter amounts to a theoretical (generalized) description of how
the system behaves under various conditions. PCT on the other hand is a
theory of how the system is organized. The latter is the more fundamental
as the behavior of a system can be deduced from its organization, whereas
its organization can only be inferred from its behavior. However, to the
extent that the description of behavior is accurate, a proposed system
organization can be considered valid only if it behaves in conformance with
the description.
Bill and Rick continue to disagree with me on this, and that, I believe, is
the source of any conflict we may have.
Bruce A.