TIP: Click on subject to list as thread! ANSI
echo: evolution
to: All
from: Tim Tyler
date: 2004-09-11 21:47:00
subject: Re: Dawkins gives incorre

Guy Hoelzer  wrote or quoted:
> in article chsg65$1hqg$1{at}darwin.ediacara.org, Tim Tyler at tim{at}tt1lock.org:
> > Guy Hoelzer  wrote or quoted:

> >> In this way, the sensitivity of p_i to individual differences in
> >> perception is minimized.
> >> This transition in the meaning of p_i happened without much
notice for many
> >> (most? The vast majority?) information scientists because it
increased the
> >> extent of objectivity and measurement agreement among scientists, and
> >> extended the utility of information theory throughout the
physical sciences.
> > 
> > Nothing fitting this description ever happened.
> 
> Are you arguing that treating p_i as frequency is almost never done, or that
> this practice has not increased in frequency?  Or are you just arguing that
> you don't think it has become sufficiently common to call it a transition?

p_i is /always/ the probability of the i'th symbol arising.

Sometimes the probabilities are determined completly by symbol frequencies 
- but the p_i's are never frequencies.

They always add up to 1.0 - like probabilities do.

> >>> An observer who knows what symbol is coming next (because he
> >>> has seen a very similar message before) will assign different
> >>> probabilites to the symbols from an observer who is seeing
> >>> the message for the first time - and both will assign different
> >>> probabilities from an observer who is not even familiar with
> >>> the language the message is written in.
> >> 
> >> This is a nice description of the (severe IMHO) limitations of the
> >> "telegraph"-context-laden version of the theory
that Shannon originally
> >> devised for his telegraph-company employer.  With all of your
protestation
> >> about my lack of fidelity to Shannon's original context, you haven't
> >> suggested any reasons why treating p_i as a frequency, rather than a
> >> probability, is problematic.  Can you think of any such problems? [...]
> > 
> > The ones above?
> 
> I didn't see any problems suggested in your previous post or in the material
> I snipped above, which was a description of how probabilities and
> frequencies differ.  Your argument then seemed to consist merely of saying
> that Shannon originally meant p_i to be a probability, rather than a
> frequency, to which I already agreed.  None of that addresses my question.

If you used freqencies, it would be equivalent to considering what
new would be learned by an observer with very little brain - whose only 
knowledge of the future consits of measuring symbol frequencies, and 
assuming what has happened before will happen again.

Such observers are surely not common.

> > p_i can only be treated as a frequency, *if* the source is something like
> > a RNG - where the probability of symbols being emitted is constant - and
> > does not depend on the history of the stream or environmental conditions.
> 
> That may be the condition under which a probability and a frequency are
> interchangeable, but it still does not address the issue at hand.  Given the
> differences between probabilities and frequencies, why isn't it better to
> think of p_i as a frequency instead of a probability as Shannon first had in
> mind?

Currently Shannon's information represents the "suprise value" of
a message - an estimate of how unlikely the observer thinks it is
to be received.

I.e.:

``Shannon's entropy measure came to be taken as a measure of the 
  uncertainty about the realization of a random variable. It thus served 
  as a proxy capturing the concept of information contained in a message 
  as opposed to the portion of the message that is strictly determined 
  (hence predictable) by inherent structures.''

 -
http://www.all-science-fair-projects.com/science_fair_projects_encyclopedia/Information_entropy

An estimate of how likely a particular, stupid observer thinks it
is to be received seems likely to be lacking in utility by comparison.

> > That is certainly not the general case - and it is not the case with many
> > common sorts of message streams either.
> >  
> >> If not, then don't you think it is worth considering the more extensive
> >> version of the theory?
> > 
> > It isn't "more extensive".
> 
> It is more absolutely extensive in its potential application and the breadth
> of its explanatory power because it overcomes the limitations of trying to
> approximate identical states among observers before they can agree upon the
> information content in a data set (an observation).

It is equivalent to calculating Shannon information for a rather dumb 
observer who is unable to perform simple logical reasoning :-(

It doesn't completely overcome the supposed "problem" of information being 
subjective  - since agents can still differ on the issue of the frequency 
of source symbols - depending on how much of the stream they have seen.

....and it would mean that the term "information" no longer represented
the "suprise value" of a message to most observers - and that's pretty
much the point of the term.

I don't see the resulting metric as being of much interest.

The term "information" already has a good meaning - and what you are
describing isn't it.
-- 
__________
 |im |yler  http://timtyler.org/  tim{at}tt1lock.org  Remove lock to reply.
---
þ RIMEGate(tm)/RGXPost V1.14 at BBSWORLD * Info{at}bbsworld.com

---
 * RIMEGate(tm)V10.2áÿ* RelayNet(tm) NNTP Gateway * MoonDog BBS
 * RgateImp.MoonDog.BBS at 9/11/04 9:47:06 PM
* Origin: MoonDog BBS, Brooklyn,NY, 718 692-2498, 1:278/230 (1:278/230)
SEEN-BY: 633/267 270
@PATH: 278/230 10/345 106/1 2000 633/267

SOURCE: echomail via fidonet.ozzmosis.com

Email questions or comments to sysop@ipingthereforeiam.com
All parts of this website painstakingly hand-crafted in the U.S.A.!
IPTIA BBS/MUD/Terminal/Game Server List, © 2025 IPTIA Consulting™.