TIP: Click on subject to list as thread! ANSI
echo: evolution
to: All
from: Tim Tyler
date: 2004-09-10 17:07:00
subject: Re: Dawkins gives incorre

Guy Hoelzer  wrote or quoted:
> Tim Tyler at tim{at}tt1lock.org wrote on 9/9/04 8:25 AM:
> > Guy Hoelzer  wrote or quoted:
> >> in article chnea4$30dh$1{at}darwin.ediacara.org, Tim Tyler at
tim{at}tt1lock.org

> >>> Shannon's information has proved to be a remarkably
useful construct -
> >>> AFAICS.
> >> 
> >> It has indeed.  Indeed, I often rely on it myself.  However,
my use of the
> >> Shannon/Weaver information measure (-p_i x log p_i,) and most
usage in the
> >> literature utterly ignores the role of perception because it has no
> >> representation in the equation.  In fact, the equation is
solely about the
> >> degree of structure in the data.
> > 
> > p_i is the probability of symbol i arising at that point.
> 
> I have no evidence to offer regarding the actual frequency with which this
> definition of p_i is used, but I do not use it to reference a
"probability"
> per se, and I am certain that many others use it as I do.  We treat it as a
> frequency, which is objectively estimable from data.

In general it is not a frequency.  Probabilities are different from
frequencies.  The frequency of something is an average the number of
times it occurs over a period of time.  Probabilities can be much more 
context sensitive.

The probability of some event occurring can be a function of all the
preceding events, and can include things such as extrapolation - rather 
than having anything to do with symbol frequencies.

One example: the frequency with which "u" is seen in english text
is one thing - but the probability with which it occurs *if* the
last symbol was a letter "q" is something else entirely.

Another example: what comes next:

1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19...?

Extrapolation and probability can be invoked to answer that - but 
symbol frequencies tell you next to nothing - except perhaps than
what comes next might contain a "1".

Probabilities are not really the same as frequencies - and it is 
probabilities - not frequencies - that are used in Shannon's equation.

Probabilities can be estimated from data - but different agents may
have access to different data.  Also, their estimates may depend on
factors such as their intelligence and experience.

There can be no consensus about probabilities - except perhaps by
God - whose probability estimates are either 1.0 or 0.0 - since
presumably he knows what's coming next.

Probabilities are a reflection of the observer's ingorance about
the state of the system - and their uncertainty about its behaviour.

In general - different observers can have different degrees of ignorance.

> In this way, the sensitivity of p_i to individual differences in 
> perception is minimized.
> This transition in the meaning of p_i happened without much notice for many
> (most? The vast majority?) information scientists because it increased the
> extent of objectivity and measurement agreement among scientists, and
> extended the utility of information theory throughout the physical sciences.

Nothing fitting this description ever happened.

> > Different observers will make different estimates of that -
> > based on their past experiences.
> 
> If that were the case, then it would be useless for someone to publish their
> estimates of p_i, except as comparative indices of perception differences,
> because the correct value would be different for each reader.

Whether comparing notes is useful depends on who is measuring what.

If the observers are in similar positions - and have similar levels of
ignorance about the system - comparing notes might result in a good
match.

However, if the observers have wildly different levels of knowledge about 
the system, comparing notes will result in wildly different estimates of
symbol probabilities.

> > An observer who knows what symbol is coming next (because he
> > has seen a very similar message before) will assign different
> > probabilites to the symbols from an observer who is seeing
> > the message for the first time - and both will assign different
> > probabilities from an observer who is not even familiar with
> > the language the message is written in.
> 
> This is a nice description of the (severe IMHO) limitations of the
> "telegraph"-context-laden version of the theory that Shannon
originally
> devised for his telegraph-company employer.  With all of your protestation
> about my lack of fidelity to Shannon's original context, you haven't
> suggested any reasons why treating p_i as a frequency, rather than a
> probability, is problematic.  Can you think of any such problems? [...]

The ones above?

p_i can only be treated as a frequency, *if* the source is something like 
a RNG - where the probability of symbols being emitted is constant - and 
does not depend on the history of the stream or environmental conditions.  

That is certainly not the general case - and it is not the case with many 
common sorts of message streams either.
 
> If not, then don't you think it is worth considering the more extensive 
> version of the theory?

It isn't "more extensive".
-- 
__________
 |im |yler  http://timtyler.org/  tim{at}tt1lock.org  Remove lock to reply.
---
þ RIMEGate(tm)/RGXPost V1.14 at BBSWORLD * Info{at}bbsworld.com

---
 * RIMEGate(tm)V10.2áÿ* RelayNet(tm) NNTP Gateway * MoonDog BBS
 * RgateImp.MoonDog.BBS at 9/10/04 5:07:16 PM
* Origin: MoonDog BBS, Brooklyn,NY, 718 692-2498, 1:278/230 (1:278/230)
SEEN-BY: 633/267 270
@PATH: 278/230 10/345 106/1 2000 633/267

SOURCE: echomail via fidonet.ozzmosis.com

Email questions or comments to sysop@ipingthereforeiam.com
All parts of this website painstakingly hand-crafted in the U.S.A.!
IPTIA BBS/MUD/Terminal/Game Server List, © 2025 IPTIA Consulting™.