[Martin Taylor 2103.02.25.23.20]
(Response to
[(nightowl) From Adam Matic, 2013.02.26.0200cet] at the end of this
message).
If I have "information" about something, my uncertainty about it has
been reduced. How is this different from the “common understanding”?
The problem I have is with people who are talking technically, and
treat “information in” something as though it was some kind of
phlogistonic variable waiting to get out.
Motion is change of location. Information is change of uncertainty.
It’s easy to understand that location and motion are related but
different. Nobody uses one word when the other is appropriate. Why
do it for information and uncertainty, which are equally clearly
distinct?
It really is most unfortunate that the two terms have become mixed
up, when Shannon was so clear about their difference. Words
shouldn’t matter, but they do when the same word is used to
represent two different things, in this case a measure and its
derivative or differential. That leads only to confusion. Would you
approve if people used “velocity” indiscriminately to mean
“location” as well as “speed” and perhaps “acceleration”?
That was Shannon's area, which was indeed about communication
systems. But I still don’t see how your idea of “representation”
differs from mine – that to “represent” something, the
representation must change when the something changes and not when
the something doesn’t change.
Are you trying to contradict me? Because if you are, you are not
succeeding.
I fail to see the necessity of something being "correct" in order
for there to be a change in uncertainty about it. Suppose for some
unknown reason you had a bet that the first man you see tomorrow
will be wearing brown shoes. You may believe that the probability of
the first man wearing black is about 0.5, of wearing brown is about
0.25, and of wearing some other colour is about 0.25. You have 1.5
bits of uncertainty about it. Now you conspire with a friend to
arrive at your house very early in the morning wearing brown shoes.
Your probability that the first man you see will have brown shoes is
now about 0.9, and “other” is about 0.1. Your uncertainty is now
just under half a bit. You have gained just over 1 bit of
information. Yet at this point there is no “correct” fact as to
whether the first man you see tomorrow will be wearing brown shoes.
Tomorrow, your uncertainty will change again, probably to near zero
(unless you can’t tell whether the shoes are black shoes brown from
mud or brown shoes blacked with soot), as soon as you see your first
man of the day.
When you are dealing with control, all you need is to be "correct"
enough that your actions do influence your perceptions sufficiently
consistently to allow you to bring them within tolerance bounds of
their reference values. Since we cannot know what is truly “out
there”, the ability to control precisely is our best bet for
correctness. We have no other “ground truth”.
There's a strong conceptual framework. Why do you suggest there
isn’t? Granted, there is disagreement about the philosophical status
of “probability”, but in practice the different ways of approaching
it usually come to the same or very nearly the same result. Once you
have probability, the rest is rock solid.
What would you want? Doesn't the concrete example of Alice's weights
that I gave in response to Adam show where the information theoretic
concepts have a place? Would it not be conceptually helpful to
mention that if you can’t see clearly where something is, it’s going
to difficult to control its location precisely? I can’t really be
very helpful here if I don’t know what would make IT feel
conceptually relevant in your mind. It is highly relevant in mine,
but that’s just me.
"Contain information" is, to me, a meaningless pair of words, for
reasons I stated. “Representation” isn’t meaningless to me, but I am
sure that the meaning I have for the word is not the meaning you
intend to convey when you use it.
OK. Please explain the difference, and what you could possibly mean
by “X represents Y” other than that the state of X is in some way
related to the state of Y. I cannot think of any other way it could
be. You obviously can.
Do you mean Information Theory when you say IF? You used IF several
times, and it brought me up short each time I ran into it.
In the fact that the states of the environment and the internal
states of the zooplankton are variable, and in some respects
co-variable.
OK. Rephrase what you did say, because I thought I "represented" you
quite precisely.
Why? There must be consistency, or there could be no control. There
doesn’t have to be deterministic invariant consistency. Control
systems can deal with variation in loop properties if the variation
is slow enough compared with the speed of control. Control is lost
if the loop properties or disturbances change too fast.
The external signal does not remain the same in that case. The
external signal is only what the sensor systems report. The eyes
don’t get the same light, so the external signal has changed. But
the perception of brightness has a very variable relation to the
amount of light entering the eye. It’s not deterministic at all,
though usually the sign of a change in perceived brightness is the
same as the sign of a change in light intensity.
There's a more important point here. Think about the mantra
“correlation does not imply causation”. There’s a similar mantra for
information: “Positive mutual information does not imply
communication”. Just as X may be correlated with Y because they are
both subject to a common influence, so also may there be positive
mutual information between X and Y because they are subject to a
common influence. In the case of PCT, if control is good, the output
influence on the CEV has a high mutual information with the
disturbance, not because the disturbance communicates with the
output – indeed, the current value of the disturbance has no
influence at all on the current value of the output in any physical
control system. The high mutual information value is because of the
fact that the properties of the control system retain effects of
past values of the disturbance. In the case of the leaky integrator,
it’s just the sum of past influences, with the influence decreasing
into the more distant past. Recent past values of the disturbance do
influence its present value, so output and disturbance have a
positive mutual information (and a negative correlation) because
they are both influenced by the same set of past values.
I suppose I care less about the form of PCT than I do about
scientific correctness. PCT isn’t (or shouldn’t be) a religion in
which something is incorrect if the Pope says it is incorrect. The
test is consistency with all the observations of the world, in
whatever context. If PCT were seen to violate the laws of
thermodynamics, I wouldn’t hesitate to “bastardise” it. But it
doesn’t. In fact, I have argued strongly that PCT is required by
the laws of thermodynamics. But the specific form of HPCT built by
Bill Powers isn’t. Nor are there (or at least nor should there be)
restrictions by fiat against trying different way of analysing PCT
structures.
Yes, I understand it in the same way as I understand the control
processes of the officers of the Spanish Inquisition. However, I
fail to see how any particular method of analysis can threaten the
purity of the religion in this case, unless, as in the case of the
Inquisition, it is the simple act of enquiry that is threatening.
I hope that some day you will see how it is helpful, or that I will
find that it isn’t. Either way, our understanding of science will
have progressed.
···
[Rupert Young 2013.02.25 23.30]
(Martin Taylor 2013.02.24.16.26)
There's a confusing issue of nomenclature in
information theory discussions. I wish people had
kept to Shannon’s usage, which was clear and easily
understood. You quotes from the Wikipedia entry make
“information” mean what Shannon called “entropy” and
I (following Garner) call “uncertainty”. Shannon
used “information” for “change in uncertainty”, and
that’s how I think the term should be used.
If your use of the the word "information"
doesn’t correspond to the common understanding, or to
that used by practitioners of IF than maybe you should
use another word. In your table you have information on
both sides, distinct from uncertainty, yet you define
information in terms of uncertainty. So I hope you’ll
excuse me if I am having trouble understanding what it
means, and its relevance to PCT.
But, if we go with your definition,
“Information gained from a message, which is the change
in uncertainty about the transmitter’s intent as a
consequence of receipt of the message; one bit of
information equates to a doubling of precision”, this
still seems to me to be talking about a communication
system, in that there is a transmitting agent who is
“putting” meaning, intention, within a transmitted
message; that is, it “represents” something.
Furthermore communication systems are designed
to
be that way, to transmit meaning. Natural control
systems are not communication systems, there is no
designer (external agent), there is no intention, or
representation, no meaning (apart from that inferred by
an external observer).
Even if we take information just as "change
in uncertainty", what is there, in natural control
systems, about which we are uncertain? I still infer
from this the concept of something being regarded as
“correct”.
It should be possible to discuss
the validity, or otherwise, of
information theory with respect to
living systems at the conceptual level
without reference to the mathematics;
at least initially, and can be
accepted or rejected as such.
I wish that were true, but so far I have not found
it possible. Maybe someone who has more facility
with language could do it, but whenever I try, I
find there are too many ambiguities left hanging.
Well, maths is pretty meaningless without
any conceptual framework.
If we can't get to the point where we see
that the IF is conceptually relevant to PCT then there’s
no point going further, as far as I am concerned. We
have to deal with our conceptual assumptions first.
The argument, seems to me, to be
the equivalent of the Representation
debate in the Artificial Intelligence
community. That is, that neural
signals “represent” something outside
of their own activation levels. In
other words they “model” or contain
“information” about the external
world.
Red Flag! That pair of words "contain information"
always is a warning that something is going
conceptually off the rails.
What's the difference between "contain
information" and representation, which you support
later?
... However, there
still seems to be an overriding assumption
that neural systems model, or contain
information about, the world, but it is
just an assumption, made from the
viewpoint of complex beings (humans) who
are misguided by their perception of the
world around them that gives them a sense
that metal states represent the world.
I don't see it so much an assumption as a way of
regarding the analysis. If from the state of X you
can deduce something about the state of Y, it seems
harmless to say that in some way X “represents” Y
(and as the maths tell you, vice-versa).
Well, coming from an AI background this is
not at all harmless, but the most fundamental
question. There is a major philosophical difference
between saying X is (causally?) related to Y and X
“represents” Y. The latter, I would say, can lead to an
invalid view of how living systems work, and lead to
invalid research.
Martin, where is "information" in a
simple system like the zooplankton?
Nowhere. Information is never “in” any system.
Ok, so how does, or can, IF be useful in the
context of studying or understanding the zooplankton?
From my viewpoint I do not see how
information is relevant to the
zooplankton. The neural signals do not
represent anything, apart from
themselves; they do not contain
messages or information, about which
there is uncertainty as there is no
“correct” value about which to be
uncertain.
I don't know whether to say "true" or to point out
that those signals have values that depend on what
is happening in the environment outside the
organism. That explicitly does mean that there is
positive mutual information between the environment
and the signals.
Ok, but where does uncertainty come in?
The only reason for the signals
being the values they are is in order
to achieve control so that the
organism can survive within its
environment and pass on its genes.
There is no need for information.
The mind boggles at this. The teleological statement
that the signals somehow manipulate themselves in
order to achieve control, and do this with no
reference to the state of the outside world, seems
really weird.
Who said that, I didn’t?
In fact the beauty, and radical
nature, of PCT is that it shows that
control systems (and any nervous
systems) are successful precisely
without any need for information and
representation.
Obviously, I strongly disagree about "information".
If the perceptual signal did not vary at least
moderately consistently with the state of something
in the environment, how would the output be able
consistently to move the perception in the direction
of decreasing error?
This last point sounds valid at first
glance, but I think there are some subtleties here that
might point to the differences in our viewpoints. Your
point seems to indicate that you see a strong
deterministic link or relationship between the external
and internal signals.
I see them as very independent entities
that may sometimes be related, but other times not. For
example, if you are perceiving lights that are too
bright (your perceptual signal is “related” to the
external light signal), but then you put your hands over
your eyes the perceptual signal changes. The external
signal remains the same, yet there is no longer a
relationship.
The danger, which I think Rick has
alluded to, is if you start
introducing other theories, which are
not really relevant as if they are
then people may start using that
approach to model control systems
which would lead to a completely
incorrect divergent path of research,
as has been the case with AI.
I don't think any paths of research are "incorrect".
They may turn out to be fruitless, but you can’t say
whether that will be the case without trying them.
They may not be the way you personally would choose
to do the research, but someone who did choose that
path would be equally legitimate in calling your
path “incorrect”. I would say you would both be
wrong.
Incorrect, fruitless; potato, potatoe. The issue is
about paradigm creep. If one introduces concepts and
processes that aren’t particularly relevant from
different paradigms then it is more or less inevitable
that we end up doing something that is some bastardised
form of PCT.
Rick may seem sensitive and over-zealous in his
defense of PCT, but I think he is quite right to protect
the “purity” of the theory against being tainted in this
way. I’m sure you can understand that.
Likewise, if it is quicker or helpful in
understanding some aspect of a control system by
using information theory, then why not use it?
Fine, but I am still at a loss to see *
how* it is helpful, or even that the concepts of IF are
relevant to PCT. From what you say it seems that the key
is explaining how the concept of uncertainty applies to
PCT or to the variables in the loop.