Re: [Fis] Fw: Responses

2014-01-11 Thread Loet Leydesdorff
Having a proper view of physics among the many possible is critical to
placing information theory on a sound basis. I have proposed Logic in
Reality as one way of giving meaning to the statement that energy and
information processes are non-separably related and how they are related.
Are there others?

 

Dear Joseph, 

 

It seems to me that there is at least one alternative: Shannon's
mathematical theory of information. Information is then defined as
content-free. Thermodynamic entropy (physics) is the special case that H is
multiplied by the Boltzmann constant and thus one obtains the dimensionality
of S. (S = kB * H). Information theory, however, can also be used in other
contexts such as economics (Theil, 1972). It does not have a realistic
interpretation such as in your argument.

 

>From such a more differentiated perspective, concepts (e.g., "electrons") do
not "exist", but are meaningful and codified within theories. There can be
sufficient evidence (in physical theorizing) to assume (for the time being)
that the external referents (electrons) exist. The logic is not in reality,
but in the argument, and one cannot jump to the (ontic) conclusion of
existence.

 

Thus, perhaps the sentence "we all agree ." (with you?) is a bit premature. 

 

Best, 

Loet

 

___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis


Re: [Fis] FW: Responses

2014-01-12 Thread Robert E. Ulanowicz
Dear Christophe,

I tried to qualify my use of "meaning", but perhaps I wasn't clear enough.

In my example I wanted to say that I(A;B) is a quantity that can be
considered a "proto-meaning" of B to A. Another way of saying the same
thing is that I(A;B) quantifies A in the context of B.

I should have added that I don't confine my notion of information to the
scenario of communication. I feel that it's discovery in that context was
an historical accident. Rather, like Stan, I  consider information more
generally as constraint, and the information of communication becomes a
subset of the more general attribute.

Hence, anything that is held together by constraints is amenable in one
form or another to the Bayesian forms of Shannon capacity. The real
advantage in doing so is that the complement of the Bayesian information
is made explicit. Vis-a-vis "constraint" this complement becomes
"flexibility". Such flexibility is an apophasis that is missing from most
scientific endeavors, but is essential to our understanding of evolution.

You are probably correct that my terminology is not orthodox and possibly
confusing to some. But I see such inconvenience as a small price to pay
for opening up a new window on quantitative evolutionary theory. I really
want folks to think outside the box of communication theory. What Shannon
started many (such as Terry Deacon) have prematurely cast aside. My
message is that we need to re-evaluate Shannon-type measures in their
Bayesian contexts. The have the potential of becoming very powerful
quantitative tools. (E.g.,
<http://www.cbl.umces.edu/~ulan/pubs/EyesOpen.pdf>.)

Peace,
Bob

>
> Bob,
>
> You seem to implicitly consider all information as being meaningful.
>
> I'm afraid such a position is source of confusion for a scientific
> approach to "information".
>
> As we know, Shannon defined a quantity of information to measure the
> capacity
> of the channel carrying it. He did not address the “meaning” that the
> information may carry (as you write in your paper “Shannon information
> considers the amount of information,
> nominally in bits, but is devoid of semantics” ). Today DP & Telecom
> activities use Shannon type information without referring to its meaning.
> Considering “information” as being meaningful looks to me as potentially
> misleading and source of misunderstandings in our discussions. Information
> can
> be meaningful or meaningless. The meaning comes from the system that
> manages
> the information.
>
> Some research activities explicitly consider information as
> meaningful.data (see http://www.mdpi.org/entropy/papers/e5020125.pdf). I'm
> afraid such a position creates
> some vocabulary problems if we want to keep a scientific background when
> trying
> to understand what is “meaning”.
>
> The meaning of information is not something that exists by itself. The
> meaning
> of information is related to the system that manages the information
> (creates
> it or uses it). And the subject has to be addressed explicitly as systems
> can
> be animals, humans or machines. Focusing on meaning generation by a system
> can bring
> some clarification (see http://www.mdpi.org/entropy/papers/e5020193.pdf).
>
> Hope this helps
>
> Christophe
>
>
>
>
>
> From: lo...@physics.utoronto.ca
> Date: Sat, 11 Jan 2014 10:49:28 -0500
> To: gordana.dodig-crnko...@mdh.se
> CC: fis@listas.unizar.es
> Subject: Re: [Fis] Fw:  Responses
>
> Dear Friends - I have been lurking as far as the discussion of Shannon
> information is concerned. I must confess I am a bit confused by the use of
> the term information associated with what you folks call Shannon
> information. To my way of thinking Shannon produced a theory of signals
> and not information. Signals sometimes contain information and sometimes
> they do not which is exactly what Shannon said of his notion of
> information. Here is what troubles me: A set of random numbers according
> to Shannon has more information than a structured set of numbers like the
> set of even numbers. For me a random set of numbers contains no
> information. Now I am sure you will agree with me that DNA contains
> information and certainly more information than a soup of random organic
> chemicals which seems to contradict Shannon's definition of information. I
> would appreciate what any of you would make of this argument of mine. Here
> is another thought about information that I would love to have some
> comments on. The information in this email post to you folks will appear
> on multiple computers, it might be converted into ink on paper, it might
> be read aloud. The information is not tied to any particular physical
> medium. But my DNA cannot be em

Re: [Fis] Fw: Responses

2014-01-16 Thread Bob Logan
Hi Christophe - I enjoyed your response - full of meaningful information - :-)  
Your point is well taken. I agree what might be meaningful information for one 
agent might be meaningless for another. I can add another example to your list 
of examples which I encountered some time ago. An author whose name I forget 
pointed out that a book written in Urdu is information for a literate Urdu 
speaker but perhaps not for those that cannot read Urdu. According to the 
definitions of Doug MacKay in 1969 and Gregory Bateson in 1973 'information is 
a distinction that makes a difference' and 'information is a difference that 
makes a difference' respectively. Meaningless information does not cut it by 
their definitions as it does not make a difference. Of course one could define 
information as a distinction or a difference that has the potential to make a 
difference for some agent. It seems to me that defining what is information is 
no easy task. My conclusion from our discussion is that depending on how you 
define information context can be an important part of what is information. The 
notion of information is extremely nuanced with multiple meanings and we seem 
to have only one word for it as pointed out by Shannon himself. In the abstract 
to his paper, The Lattice Theory of Information Shannon (1953) wrote, "The word 
"information" has been given many different meanings by various writers in the 
general field of information theory. It is likely that at least a number of 
these will prove sufficiently useful in certain applications to deserve further 
study and permanent recognition. It is hardly to be expected that a single 
concept of information would satisfactorily account for the numerous possible 
applications of this general field. The present note outlines a new approach to 
information theory, which is aimed specifically at the analysis of certain 
communication problems in which there exist a number of information sources 
simultaneously in operation."

MacKay made a distinction between 'selective information' as defined by 
Shannon's formula and 'structural information', which indicates how 'selective 
information' is to be interpreted.

"Structural information must involve semantics and meaning if it is to succeed 
in its role of interpreting selective or Shannon information. Structural 
information is concerned with the effect and impact of the information on the 
mind of the receiver and hence is reflexive. Structural information has a 
relationship to pragmatics as well as semantics where pragmatics tries to 
bridge the explanatory gap between the literal meaning of a sentence and the 
meaning that the speaker or writer intended. Shannon information has no 
particular relation to either semantics or pragmatics. It is only concerned 
with the text of a message and not the intentions of the sender or the possible 
interpretations of the receiver (Logan 2014)."

The above material is from my book What is Information? to be published 
simultaneously as a printed book and an e-book by DEMO press in Toronto. I 
would be happy to share this with you as a PDF, Christophe or any other reader 
that finds the above information meaningful and interesting.

Thank you Christophe for providing me with the opportunity to muse so more 
about the meaning of information especially 'meaningless information'. I quite 
enjoy going down the rabbit hole.  all the best Bob
__

Robert K. Logan
Prof. Emeritus - Physics - U. of Toronto 
Chief Scientist - sLab at OCAD
http://utoronto.academia.edu/RobertKLogan
www.physics.utoronto.ca/Members/logan


On 2014-01-16, at 5:52 AM, Christophe wrote:

> Dear Bob, 
> Thanks for your answer. 
> So for you, information is always meaningful.
> Such a statement is surprising when many examples can display cases of 
> meaningless information. 
> The well known Chinese Room Argument: a  sentence written in Chinese  is 
> meaningless to a non Chinese speaking reader.
> A vervet monkey alarm is meaningful information for other vervet monkeys but 
> is meaningless for a passing by dog. 
> Cyphered information is meaningful or meaningless depending if the receiver 
> has the cyphering key or not.
> We can agree that the meaning of information does not exist by itself but is 
> a result of an interpretation by a system. The interpretations of given 
> information can deliver different meanings and "no meaning" is a possible 
> outcome.  Looking at downgrading to signals the intrerpretation of 
> information is surprising.
> Now, regarding information theory, I still understand it in a scientific 
> background as it is part of mathematics and computing. Things can be 
> different indeed if you look at applications of IT (to linguistics, 
> psychology, …). 
> But the key point may be that disregarding the possibility for meaningless 
> information shuts a road in analyzing the possibilities for computers to 
> understand us. A lot is still to be done in this area and

Re: [Fis] Fw: Responses

2014-01-21 Thread Robert E. Ulanowicz

> The reason of being of information, whatever its content or quantity, is
> to be used by an agent (biological or artificial).

Dear Christophe,

In making this restriction you are limiting the domain of information to
communication and excluding all information that inheres in structure
per-se. John Collier has called the latter manifestation "enformation",
and the calculus of IT is quite effective in quantifying its extent.
Perhaps John would like to comment?

Cheers,
Bob U.


___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis


Re: [Fis] Fw: Responses

2014-01-22 Thread Loet Leydesdorff
Dear colleagues, 

This discussion and reading the beautiful book of Bob Logan entitled "What
is information?" (shortly forthcoming) made me go back to reading MacKay
(1969) once more. I cannot find "the distinction that makes a difference" as
it is quoted by Floridi (2005) -- and thereafter repeated by many -- so that
I think that the honour goes to Bateson (1973) for "a difference which makes
a difference." MacKay, however, makes the point, for example, on p. 136 that
the sentence "S is a source of information" is incomplete. "It must always
be completed (even if sometimes implicitly) in the form 'S is a source of
information to receiver R'." Two sentences later he calls this "significant
information" that must be "capable of embodying and abiding by an agreed
code or symbolic calculus."

Elsewhere, he distinguishes this substantive concept of information from
"amounts of information" that can be measured using (Shannon's) information
theory. It seems to me that any discourse (physics, biology, psychology,
sociology, etc.) can be further "informed" specifically in terms that are
defined within and relevant to the specific discourse. This accords with the
intuitive sense of "information" as meaningful information: meaningful for a
discourse.

Shannon's definition of information is counter-intuitive, but it provides us
with a calculus that has major advantages. Katherine Hayles suggested that
the two concepts can be compared with the discussion of whether a glass is
half-full or half-empty. A Chinese colleague (Wu Yishan) once told me that
in Chinese one has two words: "sjin sji" and "tsin bao" which correspond
respectively to Shannon's and Bateson's definitions of information.

A substantive definition of information (e.g., as a distinction that makes a
difference for a receiver) requires the specification of the concept in a
theory about the receiving system. This definition is therefore a priori
system-specific; for example, for some of us this system is physics; for
others it is biological discourse. At this level, one can again abstract
from the substance and use Shannon's IT as entropy statistics. Sometimes,
this allows us to explore the use of algorithms developed in one field
(e.g., biology) in another (e.g., sociology). Concepts such as autopoiesis
or auto-catalysis have carried these functions.

For example, in the context of Ascendency Theory, Bob Ulanowicz showed how
one can use the mutual information in three dimensions as an indicator of
systemness. I use that as a systems indicator when operationalizing "the
triple helix of university-industry-government relations". Such translations
of metaphors are always in need of further elaboration because the
theoretical context changes and thus the specification of what the
information means. However, the advantage to be able to measure in bits
(nats or dits) frees us from the philosophical confusion about what
information "is". 

In my opinion, "information" can only be defined within a discourse. The
mathematical definition of Shannon has specific functions which enable us to
combine with different discourses (among which, specifically physics since S
= k(B)*H). "H", however, is dimensionless and defined as the expected
information content of a message *before* it is received. It is yet to be
provided with meaning. One could consider this meaninglesness as the
specific difference of a mathematical concept of information. (Perhaps, it
is easier to use "uncertainty" for this mathematical concept.)

Best wishes,
Loet

-Original Message-----
From: fis-boun...@listas.unizar.es [mailto:fis-boun...@listas.unizar.es] On
Behalf Of Robert E. Ulanowicz
Sent: Tuesday, January 21, 2014 8:45 PM
To: Christophe
Cc: fis@listas.unizar.es
Subject: Re: [Fis] Fw: Responses


> The reason of being of information, whatever its content or quantity, 
> is to be used by an agent (biological or artificial).

Dear Christophe,

In making this restriction you are limiting the domain of information to
communication and excluding all information that inheres in structure
per-se. John Collier has called the latter manifestation "enformation", and
the calculus of IT is quite effective in quantifying its extent.
Perhaps John would like to comment?

Cheers,
Bob U.


___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis

___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis


Re: [Fis] Fw: Responses

2014-01-22 Thread John Collier


At 09:45 PM 2014-01-21, Robert E. Ulanowicz wrote:
> The reason of being of
information, whatever its content or quantity, is
> to be used by an agent (biological or artificial).
Dear Christophe,
In making this restriction you are limiting the domain of information
to
communication and excluding all information that inheres in
structure
per-se. John Collier has called the latter manifestation
"enformation",
and the calculus of IT is quite effective in quantifying its extent.
Perhaps John would like to comment?
I developed this concept in order to reply to Jeff Wicken's complaint
that Brooks and Wiley did not distinguish properly between the complement
of entropy and structural information, but I used it in print to discuss,
in the context of cognitive science and especially John Perry's use of
information (see Barwise and Perry Situations and Attitudes and
his What is information?, as well as Dretske's book on information and
perception) what the world must be like in order to make sense of
information coming from the world into our brains. The article can be
found at
Intrinsic
Information (1990) In P. P. Hanson (ed) Information, Language and
Cognition: Vancouver Studies in Cognitive Science, Vol. 1 (originally
University of British Columbia Press, now Oxford University Press, 1990):
390-409. Details about information are there, but the gist of it is that
can be measured, is unique, and depends on time scale to distinguish it
from informational entropy in information systems. The uniqueness
hypothesis was developed very carefully in my former student, Scott
Muller's PhD thesis, published as Asymmetry: The Foundation of
Information (The Frontiers Collection) by Springer in 2007.
I am rather busy now at a conference, or else I would say more
here.
John




Professor John
Collier
colli...@ukzn.ac.za
Philosophy and Ethics, University of KwaZulu-Natal, Durban 4041 South
Africa
T: +27 (31) 260 3248 / 260 2292   F:
+27 (31) 260 3031

Http://web.ncf.ca/collier



___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis


Re: [Fis] Fw: Responses

2014-01-22 Thread Christophe
Dear Bob U,
If your are talking about resident information, as available for usage, I take 
it as being part of information that can be used by the agent.
Let me go through John's paper (thanks John). 
Best
Christophe
 
> Date: Tue, 21 Jan 2014 14:45:15 -0500
> Subject: Re: [Fis] Fw:  Responses
> From: u...@umces.edu
> To: christophe.men...@hotmail.fr
> CC: lo...@physics.utoronto.ca; fis@listas.unizar.es
> 
> 
> > The reason of being of information, whatever its content or quantity, is
> > to be used by an agent (biological or artificial).
> 
> Dear Christophe,
> 
> In making this restriction you are limiting the domain of information to
> communication and excluding all information that inheres in structure
> per-se. John Collier has called the latter manifestation "enformation",
> and the calculus of IT is quite effective in quantifying its extent.
> Perhaps John would like to comment?
> 
> Cheers,
> Bob U.
> 
> 
  ___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis