Wholegroup; Optimal control

Hello, all –
Here’s a step in another direction for the addressing. I’ve included
CSGNET in the list, but also added individual addresses for people who
have commented on the CSGnet stuff about neuroscience. So if you get the
message twice, delete the one from CSGNET, then look at the other, and
reply-to-all to it. Eventually I’ll phase out the CSGnet one, because the
cc list doesn’t seem to make it to the CSG list, according to Rick
Marken.
Reading Todorov, I’m coming to realize that my concept of “optimal
control theory” is out of date. Todorov notes the open-loop nature
of the older ideas and sees that “intelligent use of feedback”
is much preferable. He also realizes that there are uncontrolled
dimensions of behavior – he calls them “redundant” – that can
be ignored. And he approaches the idea of hierarchical control, even
going so far as to say that the model-based controllers may exist at
higher levels, but they work by sending “commands” to
neuromuscular negative feedback servos at lower levels. At one point he
proposes that reaching behavior, which is disturbance-resistant, is
actually accomplished by specifying a series of intermediate targets
which are reached one after the other – he comes that close to
seeing that it’s just a smoothly varying reference signal that goes to
the lower systems, but doesn’t quite get there. At least he rejects the
idea of planned and controlled trajectories.

So in the papers I’ve seen, he is working his way slowly toward PCT, and
I suppose would get there eventually.

There are two main ideas that he would have to overcome and discard
first. The first is that sensory delays make fast control impossible
using only negative feedback, and the other is that the signals in the
model have to be horribly noisy. I’d love to see a graduate student do a
real job of tracing where those ideas came from – Lashley is certainly
not the only source. I see Todorov has an “efference copy” in
one diagram, which he says is needed because of sensory delays and noise,
so von Holst has to bear part of the blame (though I don’t think von
Holst spoke of internal models and sensory delays).

All of (what I see as) the more advanced ideas in the Todorov version of
optimal control theory appeared in the first paper by Clark, Macfarland,
and me in 1960. I guess the literature in which we published then and
afterward was simply not read by the people who produced optimal control
theory, starting with the Kalman filter. Of course I didn’t know then
that sensory delays could be so easily compensated by an integrator in
the loop, and I hadn’t developed any testable working models yet (though
I used an analog computer to model control systems). Those things
happened during and after the years between 1960 and 1973 when B:CP came
out.

The other main idea that has to be dropped concerns the noise in the
system. I always thought that was a mistake, simply because when I look
around at my world (I can’t comment on what others see), it’s not buried
in sparkly snow and jumping and wobbling around all over the place. When
I lift a fork to my mouth I hardly ever miss. Tom Bourbon pointed out the
millions upon millions of drivers whizzing past each other in opposite
directions within a couple of feet of a collision, and hardly ever
hitting each other or anything else.

I think the idea that the nervous system is noisy came from looking at
signals in single axons. In the first place, not understanding what the
rapid variations in firing rate represent, it’s easy for an experimenter
to conclude that the variations are random even if they’re completely
systematic. And in the second place, I doubt that any information path in
the brain is served by one lone axon. In the spinal cord I know for a
fact (i.e., a publication says so) that the error signal running from the
spinal cord to a typical muscle is probably a bundle of around a thousand
fibers coming from a thousand motor neurons. Naturally, if you just look
at one of them, there will be some random variations. But their combined
effect is smoothly variable and almost noise-free. With
“recruitment” having its effect, the result is even almost a
linear representation of the stimulus at the source. I think we can
ignore neural noise for the most part, except at very low levels of
signal magnitude at which only a few neurones fire at a time. I’ve never
seen much noise in my tracking data, in which the model’s behavior
accounts for 95% or more of the variance of the real handle
movements.

Finally, most of the apparent noise in behavior comes from disturbances,
which being uncorrelated with each other do give a pretty good imitation
of randomness. Most of the action of a control system is needed to
counteract small (and sometimes large) disturbances. Of course only the
part of the frequency spectrum of the disturbance that lies within the
bandwidth of good control is counteracted.

In the big picture, Todorov’s approach is not wildly different from PCT.
It just hasn’t been developed nearly as far. A good part of it was wasted
effort because of trying to deal with nonexistent problems, but all
theorists including me have been through that. Par for the
course.

Best,

Bill P.

[From Rick Marken (2010.11.01.1400)]

Hello, all --

BP: Here's a step in another direction for the addressing. I've included CSGNET
in the list, but also added individual addresses for people who have
commented on the CSGnet stuff about neuroscience. So if you get the message
twice, delete the one from CSGNET, then look at the other, and reply-to-all
to it. Eventually I'll phase out the CSGnet one, because the cc list doesn't
seem to make it to the CSG list, according to Rick Marken.

RM: Since there is no cc list included when your posts are delivered
via CSGNet replies to your posts will go only to CSGNet. Are you going
to relay them to the cc people or are we just going to have two
parallel conversations, both of which include you, of course, but only
one of which includes the people on CSGNet. Seems a bit confusing. Is
there some reason why everyone on the cc list can't just subscribe to
CSGNet?

BP: Reading Todorov, I'm coming to realize that my concept of "optimal control
theory" is out of date. Todorov notes the open-loop nature of the older
ideas and sees that "intelligent use of feedback" is much preferable. ...

In the big picture, Todorov's approach is not wildly different from PCT.

RM: Well, it is different in terms of that one little point about
control of perception (when they talk about control of feedback I
don't think they get the idea that the feedback is a perceptual
variable and that it can therefore be the output of a computation
like the one you use in the Little Man, where a controlled perception
is the weighted sum of sensory inputs representing both shoulder and
elbow angle). It seems to me that if neurophysiologists understood
that control involves the control of perceptual variables they could
start a new and possibly very productive approach to research; testing
for controlled variables. But in this case the possible controlled
variables could be measured directly as an afferent neural signal. For
example, they could look for afferent signals that do change when a
limb is moved voluntarily (presumably reflecting variance in the
reference input) but that don't change when the limb is voluntarily
held in one position against disturbance. Once such a neuron is found,
then one could make more detailed studies of what set of sensory
inputs that neural signal is a function of.

Actually, this sounds like what I could glean about that Henry Y. and
Bruce A. may be doing using fMRI? Is that right?

Best

Rick

···

On Tue, Nov 1, 2011 at 6:40 AM, Bill Powers <powers_w@frontier.net> wrote:
--
Richard S. Marken PhD
rsmarken@gmail.com
www.mindreadings.com

Hello, all.

From Rick Marken (2010.11.01.1400)]

RM: Since there is no cc list included when your posts are delivered
via CSGNet replies to your posts will go only to CSGNet.

BP: But I usually quote extensively to show what I'm replying to, so the others see even the replies that go only to me, if I comment on them.

RM: Are you going
to relay them to the cc people or are we just going to have two
parallel conversations, both of which include you, of course, but only
one of which includes the people on CSGNet. Seems a bit confusing. Is
there some reason why everyone on the cc list can't just subscribe to
CSGNet?

BP: I wouldn't blame them for not want to be on yet another list. So far, since my starting this discussion, the number of subscribers has increased to 132. It was 130, so two people, count 'em, have subscribed. That sounds like a vote to me. It's not my privilege to ask why people don't subscribe. They don't have to justify themselves to me.

RM: Well, it is different in terms of that one little point about
control of perception (when they talk about control of feedback I
don't think they get the idea that the feedback is a perceptual
variable and that it can therefore be the output of a computation
like the one you use in the Little Man, where a controlled perception
is the weighted sum of sensory inputs representing both shoulder and
elbow angle).

BP: If Todorov just kept going, he would get there eventually. After all, if you look through his diagrams to find a likely place for perception to happen, where would you find it? The feedback signals, of course. That took me a while to figure out, too. He'd need to add some explicit perceptual input functions instead of just drawing arrows from the environment into the control system.

RM: It seems to me that if neurophysiologists understood
that control involves the control of perceptual variables they could
start a new and possibly very productive approach to research; testing
for controlled variables.

BP: They will.

RM: But in this case the possible controlled
variables could be measured directly as an afferent neural signal. For
example, they could look for afferent signals that do change when a
limb is moved voluntarily (presumably reflecting variance in the
reference input) but that don't change when the limb is voluntarily
held in one position against disturbance. Once such a neuron is found,
then one could make more detailed studies of what set of sensory
inputs that neural signal is a function of.

BP: All we're doing is saving them some time. All those things are obvious after you think of them, and if PCT didn't exist, they would then invent it.

RM: Actually, this sounds like what I could glean about that Henry Y. and
Bruce A. may be doing using fMRI? Is that right?

BP: I think their first interest is simply to see if they can identify specific brain volumes associated with specific control tasks. David Goldstein has tried doing that a bit with QEEG -- quantitative EEGs, which looks for correlated activities in different parts of the scalp potentials. No publishable results yet. We'll get all the details from the horse's mouths when the horses are ready to talk.

Best,

Bill P.

I'm copying all this stuff to Steve and the rest of "wholegroup."

Bill

···

At 06:51 PM 11/1/2011 +0000, Warren Mansell wrote:

Also Bill, what does Steve Scott think of your analysis? Can I post it on pctweb? More importantly, wouldn't it be a great opportunity to write a commentary to Nature presenting your analysis as a formal commentary on the Nature paper?
Warren

I have a couple of methodological questions about optimal control in biological systems:

* What is an optimal controller?

* How do you know when you're looking at one?

For engineered systems, an optimal controller is one whose control law was explicitly designed to optimise some performance measure. The problem is, that is a property of how the system was designed, not a property of the control system itself. You may not be able to tell by looking only at the control system that the control law built into it was designed to optimise anything. Even if you can demonstrate post hoc that its control law is the best according to some performance measure, I expect you can find some such measure for every control law that works, and it's like drawing the target after firing the bullet.

Some control systems do their own optimisation, and these can be classed as optimal controllers without consulting the designer. An example would be an automatic pilot of the more sophisticated sort, that explicitly calculates a route between waypoints to minimise some measure of time and fuel used, making the same calculations that the engineer would.

Less clearly an "optimal controller" would be a simple PID controller with extra circuitry to automatically tune its parameters to maximise some performance measure. But it's not clear to me that "optimal control" is a useful way to think of this system. "Adaptive control", certainly, but in the strict sense of the Wikipedia article on optimal control, it isn't.

That's a rather tedious analysis of definitions, but what I draw from it is that "optimal control" may not be a useful way of thinking about biological control systems. It is biologically plausible that organisms will either individually learn or as a species evolve to minimise the physiological costs of their control activities, so one should expect to see adaptive control everywhere, of the sort that is present in the self-tuning PID controller.

But biological systems are not explicitly designed, so the only case where one can really say that optimal control is happening is when the system can be demonstrated to be doing its own optimising, by finding circuitry such as in the example of the automatic pilot, that does those calculations. I am not a neuroscientist, but my understanding is that while the brain may no longer be a competely black box, techniques such as fMRI and recording from neurons leave it still a very murky forest. Demonstrating "optimal control" in that sense may be beyond current capabilities.

It isn't enough just to find that a biological system implements a control law of which some optimality property can be proved, or that a simulation using a designed optimal controller produces similar behaviour. Every controller looks as if it is minimising its error, because keeping the error small is, by definition, what a controller does. All controllers that succeed at the same task will seem to produce the same effect on the variables they are controlling.

···

--
Richard Kennaway, jrk@cmp.uea.ac.uk, http://www.cmp.uea.ac.uk/~jrk/
School of Computing Sciences,
University of East Anglia, Norwich NR4 7TJ, U.K.

Hello, all –

JRK: I have a couple of methodological questions about optimal
control in biological systems:

  • What is an optimal controller?
  • How do you know when you’re looking at one?

BP: Your post covers the territory better than the references I’ve been
looking at.

JRK: For engineered systems, an optimal controller is one whose
control law was explicitly designed to optimise some performance measure.
The problem is, that is a property of how the system was designed, not a
property of the control system itself. You may not be able to tell
by looking only at the control system that the control law built into it
was designed to optimise anything. Even if you can demonstrate post
hoc that its control law is the best according to some performance
measure, I expect you can find some such measure for every control law
that works, and it’s like drawing the target after firing the
bullet.

BP: Yes, and you see claims that this is the “best possible control
system,” which I sincerely doubt that anyone could back up. It’s
“the best according to some performance measure” as you say,
but only for that measure and not for any other, such as achieving the
smallest possible error.

If you look carefully, I think you’ll find that the optimal control
people avoid saying that the control system optimizes itself to satisfy
its own choice of criteria. The criteria are either somehow objective
parts of reality, or they are specified by the design engineer. The same
physical system could be optimized to satisfy many different sets of
optimality criteria.

JRK: Some control systems do their own optimisation, and these can be
classed as optimal controllers without consulting the designer. An
example would be an automatic pilot of the more sophisticated sort, that
explicitly calculates a route between waypoints to minimise some measure
of time and fuel used, making the same calculations that the engineer
would.

BP: In PCT, optimal control is defined in terms of intrinsic variables
and their reference signals: the inherited definition of the set of
“essential variables” as Ashby called them, which are essential
for maintenance of the life support systems and perhaps reproductive
success. But survival does not require “optimal” control, it
just requires good enough control. If it’s good enough to allow the
organism to reproduce sufficiently well to preserve the species, that’s
all Darwinian natural selection can do. It doesn’t require inventing an
iPod.

And in PCT, the strategy for optimization doesn’t require computing
inverse kinematics and dynamics or making models of the world, or
computing the exact action that will have the best possible outcome. It
just requires making random parameter changes (in the E. coli manner, not
just any old way) in response to the existence of too much error signal
in the existing behavioral control systems. No knowledge of the
environment is needed, nor are Hamiltonians
employed.

JRK: Less clearly an “optimal controller” would be a simple
PID controller with extra circuitry to automatically tune its parameters
to maximise some performance measure. But it’s not clear to me that
“optimal control” is a useful way to think of this
system. “Adaptive control”, certainly, but in the strict
sense of the Wikipedia article on optimal control, it
isn’t.

BP: “Adaptive” is somewhat descriptive, but even that carries a
hidden assertion, which is that it is the organism that must adapt itself
to environmental conditions. While that certainly happens, organisms also
do a great deal of adapting the environment to suit their own purposes.
Environments, aside from the living systems in them, can’t control
anything unless they’re little bits of it designed by living systems,
like thermostats.

JRK: That’s a rather tedious analysis of definitions, but what I draw
from it is that “optimal control” may not be a useful way of
thinking about biological control systems. It is biologically
plausible that organisms will either individually learn or as a species
evolve to minimise the physiological costs of their control activities,
so one should expect to see adaptive control everywhere, of the sort that
is present in the self-tuning PID controller.

BP: As drawn in the references we’ve been looking at, a PID controller
senses magnitude, rate of change, and time integral of the controlled
quantity, and acts proportionally to the error. This is a rather special
design and not necessarily the best one for a specific purpose. Look at
our controller for an inverted pendulum (Richard analyzes it in the
appendix of LCS3). It uses only proportional-derivative controllers, the
integral terms being supplied by feedback from the accelerating and
moving masses being controlled. The environment is used as a model of
itself and is incorporated into the control system. A multi-level
controller is used, the lowest level controlling the highest derivative
of the controlled variable. This is probably equivalent to a PID
controller if you add a few more derivatives to it, but design-wise it’s
way simpler than what the optimal PID controllers go through to get
control.

By the way, Richard, wouldn’t it be interesting to make the inverted
pendulum self-reorganizing? Do you think it could learn to erect
itself?

JRK: But biological systems are not explicitly designed, so the only
case where one can really say that optimal control is happening is when
the system can be demonstrated to be doing its own optimising, by finding
circuitry such as in the example of the automatic pilot, that does those
calculations. I am not a neuroscientist, but my understanding is
that while the brain may no longer be a competely black box, techniques
such as fMRI and recording from neurons leave it still a very murky
forest. Demonstrating “optimal control” in that sense may
be beyond current capabilities.
It isn’t enough just to find that a biological system implements a
control law of which some optimality property can be proved, or that a
simulation using a designed optimal controller produces similar
behaviour. Every controller looks as if it is minimising its error,
because keeping the error small is, by definition, what a controller
does. All controllers that succeed at the same task will seem to
produce the same effect on the variables they are
controlling.

BP: If we think of the optimality criterion as just another controlled
variable, it’s pretty clear that no real system actually produces the
best possible control. The best possible controller would keep the system
exactly in the specified reference state with no deviations at all from
the criteria of optimality. As in any control process, we can only judge
the quality of control in terms of how small the remaining error is, and
there will always be some remaining error.

A self-reorganizing PCT model of multi-level, multidimensional control
would adapt itself to achieve the least error it can across all the
control systems that are simultaneously controlling different variables,
from the intrinsic variables to the highest level of learned variables.
There isn’t any “optimum” state other than least
error.

Best,

Bill

···

At 01:12 PM 11/2/2011 +0000, Richard Kennaway wrote:

Hello, Henry –

If you want to see some conceptual confusion, just check out
this
article, published today, on optimal control from Karl Friston,
Richard’s favorite:)
Good God. I have written about six paragraphs here, and this is all
that hasn’t been deleted. I think I just have to decline to comment.
There are so many things wrong with Friston’s ideas that we just have to
deal with them one at a time if they come up in conversation. There’s no
way I can handle this entire tub of ********. That’s an eight-letter
word.

Bill

Friston.pdf (1.15 MB)

···

At 03:03 PM 11/2/2011 -0400, Henry Yin wrote:

Bill

Not sure why you really care, all of life is
a social affair with competing views, some great others not so.

You have been competing with second order
cybernetics for decades and they have a huge following, by the way I have been
singing your praises on the cybercom list for some time now, and nobody has
said anything untoward about PCT, in fact some of the guys there think PCT is
very good stuff.

I think this is actually good thing that people
are beginning to think about control systems and living organisms. The model
and theory with the most robust tests, evidence, and observer corroboration,
will ultimately prevail
and that’s PCT. These guys have very little you have tests, computer
models, observer corroboration, theory.

Don’t worry, if these guys want to
get into control theory, they will be requisitely required to delve into PCT.

All this is good, this is how one turns aware
and attentive minds to alternatives.

Kind regards

Gavin

Hello, Henry

···

At 03:03 PM 11/2/2011 -0400, Henry Yin wrote:

If you want to see some
conceptual confusion, just check out this

article, published today,
on optimal control from Karl Friston,

Richard’s
favorite:)

Good
God. I have written about six paragraphs here, and this is all that hasn’t been
deleted. I think I just have to decline to comment. There are so many things
wrong with Friston’s ideas that we just have to deal with them one at a time if
they come up in conversation. There’s no way I can handle this entire tub of
********. That’s an eight-letter word.

Bill