Cari Tutti,
ho scritto più volte le stesse cose per cui sono d'accordo con Voi,
specialmente con gli ultimi intervenuti. E dato che sono un forestiero
rispetto alle Vostre discipline, ma non uno straniero dell'armonia del
sapere o del sapere dell'armonia, questo è una bella cosa. Auguri di buon
Natale e per il nuovo anno.
Francesco

2016-12-24 7:45 GMT+01:00 Loet Leydesdorff <l...@leydesdorff.net>:

> Dear Terrence and colleagues,
>
>
>
> I agree that we should not be fundamentalistic about “information”. For
> example, one can also use “uncertainty” as an alternative word to
> Shannon-type “information”. One can also make distinctions other than
> semantic/syntactic/pragmatic, such as biological information, etc.
>
>
>
> Nevertheless, what makes this list to a common platform, in my opinion, is
> our interest in the differences and similarities in the background of these
> different notions of information. In my opinion, the status of Shannon’s
> mathematical theory of information is different  from special theories of
> information (e.g., biological ones) since the formal theory enables us to
> translate between these latter theories. The translations are heuristically
> important: they enable us to import metaphors from other backgrounds (e.g.,
> auto-catalysis).
>
>
>
> For example, one of us communicated with me why I was completely wrong,
> and made the argument with reference to Kullback-Leibler divergence between
> two probability distributions. Since we probably will not have “a general
> theory” of information, the apparatus in which information is formally and
> operationally defined—Bar-Hillel once called it “information calculus”—can
> carry this interdisciplinary function with precision and rigor. Otherwise,
> we can only be respectful of each other’s research traditions. J
>
>
>
> I wish you all a splendid 2017,
>
> Loet
>
>
> ------------------------------
>
> Loet Leydesdorff
>
> Professor, University of Amsterdam
> Amsterdam School of Communication Research (ASCoR)
>
> l...@leydesdorff.net ; http://www.leydesdorff.net/
> Associate Faculty, SPRU, <http://www.sussex.ac.uk/spru/>University of
> Sussex;
>
> Guest Professor Zhejiang Univ. <http://www.zju.edu.cn/english/>,
> Hangzhou; Visiting Professor, ISTIC,
> <http://www.istic.ac.cn/Eng/brief_en.html>Beijing;
>
> Visiting Professor, Birkbeck <http://www.bbk.ac.uk/>, University of
> London;
>
> http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en
>
>
>
> *From:* Fis [mailto:fis-boun...@listas.unizar.es] *On Behalf Of *Terrence
> W. DEACON
> *Sent:* Thursday, December 22, 2016 5:33 AM
> *To:* fis
>
> *Subject:* Re: [Fis] What is information? and What is life?
>
>
>
> Against information fundamentalism
>
>
>
> Rather than fighting over THE definition of information, I suggest that we
> stand back from the polemics for a moment and recognize that the term is
> being used in often quite incompatible ways in different domains, and that
> there may be value in paying attention to the advantages and costs of each.
> To ignore these differences, to fail to explore the links and dependencies
> between them, and to be indifferent to the different use values gained or
> sacrificed by each, I believe that we end up undermining the very
> enterprise we claim to be promoting.
>
>
>
> We currently lack broadly accepted terms to unambiguously distinguish
> these divergent uses and, even worse, we lack a theoretical framework for
> understanding their relationships to one another.
>
> So provisionally I would argue that we at least need to distinguish three
> hierarchically related uses of the concept:
>
>
>
> 1. Physical information: Information as intrinsically measurable medium
> properties with respect to their capacity to support 2 or 3 irrespective of
> any specific instantiation of 2 or 3.
>
>
>
> 2. Referential information: information as a non-intrinsic relation to
> something other than medium properties (1) that a given medium can provide
> (i.e. reference or content) irrespective of any specific instantiation of 3.
>
>
>
> 3. Normative information: Information as the use value provided by a given
> referential relation (2) with respect to an end-directed dynamic that is
> susceptible to contextual factors that are not directly accessible (i.e.
> functional value or significance).
>
>
>
> Unfortunately, because of the history of using the same term in an
> unmodified way in each relevant domain irrespective of the others there are
> often pointless arguments of a purely definitional nature.
>
>
>
> In linguistic theory an analogous three-part hierarchic partitioning of
> theory IS widely accepted.
>
>
>
> 1. syntax
>
> 2. semantics
>
> 3. pragmatics
>
>
>
> Thus by analogy some have proposed the distinction between
>
>
>
> 1. syntactic information (aka Shannon)
>
> 2. semantic information (aka meaning)
>
> 3. pragmatic information (aka useful information)
>
>
>
> This has also often been applied to the philosophy of information (e.g.
> see The Stanford Dictionary of Philosophy entry for ‘information’).
> Unfortunately, the language-centric framing of this distinction can be
> somewhat misleading. The metaphoric extension of the terms ‘syntax’ and
> ‘semantics’ to apply to iconic (e.g. pictorial) or indexical (e.g.
> correlational) forms of communication exerts a subtle procrustean influence
> that obscures their naturalistic and nondigital features. This language
> bias is also often introduced with the term ‘meaning’ because of its
> linguistic connotations (i.e. does a sneeze have a meaning? Not in any
> standard sense. But it provides information “about” the state of person who
> sneezed.)
>
>
>
> So as a first rough terminological distinction I propose using
>
>
>
> 1. physical information (or perhaps information1)
>
> 2. referential information (information2)
>
> 3. normative information (information3)
>
>
>
> to avoid definitional equivocation and the loss of referential clarity.
>
>
>
> I would argue that we use the term ‘information’ in a prescinded way in
> both 1 and 2. That is, considered from the perspective of a potential
> interpretation (3) we can bracket consideration of any particular
> interpretation to assess the possible relational properties that are
> available to provide reference (2); and we can bracket both 3 and 2 to only
> consider the medium/signal properties minimally available for 2 and 3
> irrespective of using them for these purposes.*
>
>
>
> Although 2 and 3 are not quantifiable in the same sense that 1 is, neither
> are they unconstrained or merely subjective. The possible referential
> content of a given medium or sign vehicle is constrained by the physical
> properties of the medium and its relationship to its physical context.
> Normative information captures the way that referential content can be
> correct or incorrect, accurate or inaccurate, useful or useless, etc.,
> depending on the requirements of the interpretive system and its relation
> to the context. In both cases there are specific unambiguously identifiable
> constraints on reference and normative value.
>
>
>
> There has been a prejudice in favor of 1 because of the (mistaken) view
> that 2 and three are in some deep sense nonphysical and subjective.
> Consistent with this view, there have been many efforts to find a way to
> reduce 2 and 3 to some expression of 1. Although it is often remarked that
> introducing non reduced concepts of referential content (2) and normative
> evaluation (3) into the theory of information risks introducing non
> quantifiable (and by assumption non scientific) attributes, I think that
> this is more a prejudice than a principle that has been rigorously
> demonstrated. Even if there is currently no widely accepted non
> reductionistic formalization of reference and significance within the
> information sciences this is not evidence that it cannot be achieved. One
> thing is clear, however, until we find a way to use the term ‘information’
> in a way that does not privilege one of these uses over the others and
> unequivocally distinguishes each and their relationships to one another,
> the debates we engage in on this forum will remain interminable.
>
>
>
> So I suggest that we commence a discussion of how best to accomplish this
> terminological brush-clearing before further debating the relevance of
> information to physics, logic, biology, or art. I apologize if this is
> already accepted as “solved” by some readers, and would be glad to receive
> and share your different taxonomies and learn of how they are justified.
>
>
>
> — Terry
>
>
>
> * Stan Salthe might organize them in a subsumptive hierarchy.
>
>
>
>
>
>
>
>
>
> On Tue, Dec 20, 2016 at 4:19 PM, Mark Johnson <johnsonm...@gmail.com>
> wrote:
>
> Dear all,
>
>
>
> It's important that one should remain practical. Shannon's formulae are
> practical. The correspondence with certain tenets of cybernetics such as
> Ashby's Law, or Maturana's "Structural Coupling" presents Shannon as a
> window for exploring *relations* empirically. This I understand to be Bob
> Ulanowicz's focus too. I think Ashby's epistemology which accompanied his
> championing of Shannon (and which seems to me to be quite radical) is worth
> a much deeper exploration (it was eclipsed by second-order cybernetics in
> the early 70s). Klaus Krippendorff wrote an excellent paper about this
> here: http://repository.upenn.edu/cgi/viewcontent.cgi?
> article=1245&context=asc_papers
>
>
>
> Information theory is counting - but it provides a way of measuring
> relations, which I think marks it out as distinct from other statistical
> techniques such as variance. It also provides the basis for questioning
> what we actually mean by counting in the first place: you might call it
> "critical counting". For example, Ashby makes the comment about "analogy"
> (a key concept if we are to say that one thing is the same class as another
> when we count them together)... (apologies because I can't find the
> reference to this right now, but will send if anyone is interested):
>
>
>
> "The principle of analogy is founded upon the assumption that a degree of
> likeness between two objects in respect of their known qualities is some
> reason for expecting a degree of likeness between them in respect of their
> unknown qualities also, and that the probability with which unascertained
> similarities are to be expected depends upon the amount of likeness already
> known."
>
>
>
> Also, just to correct a possible misconception: I don't think counting
> leads to populism. Econometrics has led to populism. Some of the greatest
> economists of the 20th century saw the problem - this is why Keynes wrote a
> book on probability, and Hayek wrote extensively criticising mathematical
> modelling. In the end, I'm afraid, it's an American problem which goes back
> to McCarthy and the distrust of criticality in the social sciences in
> favour of positivist mathematical "objectivist" approaches. Those schools
> in the US which championed mathematical approaches (Chicago, etc) got all
> the funding, controlled the journals, whilst others were starved. The
> legacy from the 1950s is still with us: it's still very hard to get an
> economics paper published unless it's got crazy equations in it. In the
> end, it's just bad theory - and bad mathematics.
>
>
>
> We could well see a similar thing happen with climate science in the next
> four years.
>
>
>
> Best wishes,
>
>
>
> Mark
>
>
>
>
>
>
>
>
>
> On 20 December 2016 at 19:55, Bob Logan <lo...@physics.utoronto.ca> wrote:
>
> Loet - thanks for the mention of our (Kauffman, Logan et al) definition
> our definition of information which is a qualitative description of
> information. As to whether one can measure information with our
> description, my response is no but I am not sure that one can measure
> information at all. What units would one use to measure information? *E* =
> mc 2 contains a lot of information but the amount of information depends
> on context. A McLuhan one-liner such as 'the medium is the message' also
> contains a lot of information even though it is only 5 words or 26
> characters long.
>
>
>
> Hopefully I have provided some information but how much information is
> impossible to measure.
>
>
>
> Bob
>
>
>
>
>
>
>
> ______________________
>
>
>
> Robert K. Logan
>
> Prof. Emeritus - Physics - U. of Toronto
>
> Fellow University of St. Michael's College
>
> Chief Scientist - sLab at OCAD
>
> http://utoronto.academia.edu/RobertKLogan
>
> www.researchgate.net/profile/Robert_Logan5/publications
>
> https://www.physics.utoronto.ca/people/homepages/logan/
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> On Dec 20, 2016, at 3:26 AM, Loet Leydesdorff <l...@leydesdorff.net>
> wrote:
>
>
>
> Dear colleagues,
>
>
>
> A distribution contains uncertainty that can be measured in terms of bits
> of information.
>
> Alternatively: the expected information content *H *of a probability
> distribution is .
>
> *H* is further defined as probabilistic entropy using Gibb’s formulation
> of the entropy .
>
>
>
> This definition of information is an operational definition. In my
> opinion, we do not need an essentialistic definition by answering the
> question of “what is information?” As the discussion on this list
> demonstrates, one does not easily agree on an essential answer; one can
> answer the question “how is information defined?” Information is not
> “something out there” which “exists” otherwise than as our construct.
>
>
>
> Using essentialistic definitions, the discussion tends not to move
> forward. For example, Stuart Kauffman’s and Bob Logan’s (2007) definition
> of information “as natural selection assembling the very constraints on the
> release of energy that then constitutes work and the propagation of
> organization.” I asked several times what this means and how one can
> measure this information. Hitherto, I only obtained the answer that
> colleagues who disagree with me will be cited. J Another answer was that
> “counting” may lead to populism. J
>
>
>
> Best,
>
> Loet
>
>
> ------------------------------
>
> Loet Leydesdorff
>
> Professor, University of Amsterdam
> Amsterdam School of Communication Research (ASCoR)
>
> l...@leydesdorff.net  <l...@leydesdorff.net>; http://www.leydesdorff.net/
> Associate Faculty, SPRU,  <http://www.sussex.ac.uk/spru/>University of
> Sussex;
>
> Guest Professor Zhejiang Univ. <http://www.zju.edu.cn/english/>,
> Hangzhou; Visiting Professor, ISTIC,
> <http://www.istic.ac.cn/Eng/brief_en.html>Beijing;
>
> Visiting Professor, Birkbeck <http://www.bbk.ac.uk/>, University of
> London;
>
> http://scholar.google.com/citations?user=ych9gNYAAAAJ&hl=en
>
>
>
> *From:* Dick Stoute [mailto:dick.sto...@gmail.com <dick.sto...@gmail.com>]
>
> *Sent:* Monday, December 19, 2016 12:48 PM
> *To:* l...@leydesdorff.net
> *Cc:* James Peters; u...@umces.edu; Alex Hankey; FIS Webinar
> *Subject:* Re: [Fis] What is information? and What is life?
>
>
>
> List,
>
>
>
> Please allow me to respond to Loet about the definition of information
> stated below.
>
>
>
> 1. the definition of information as uncertainty is counter-intuitive
> ("bizarre"); (p. 27)
>
>
>
> I agree.  I struggled with this definition for a long time before
> realising that Shannon was really discussing "amount of information" or the
> number of bits needed to convey a message.  He was looking for a formula
> that would provide an accurate estimate of the number of bits needed to
> convey a message and realised that the amount of information (number of
> bits) needed to convey a message was dependent on the "amount" of
> uncertainty that had to be eliminated and so he equated these.
>
>
>
> It makes sense to do this, but we must distinguish between "amount of
> information" and "information".  For example, we can measure amount of
> water in liters, but this does not tell us what water is and likewise the
> measure we use for "amount of information" does not tell us what
> information is. We can, for example equate the amount of water needed to
> fill a container with the volume of the container, but we should not think
> that water is therefore identical to an empty volume.  Similarly we should
> not think that information is identical to uncertainty.
>
>
>
> By equating the number of bits needed to convey a message with the "amount
> of uncertainty" that has to be eliminated Shannon, in effect, equated
> opposites so that he could get an estimate of the number of bits needed to
> eliminate the uncertainty.  We should not therefore consider that this
> equation establishes what information is.
>
>
>
> Dick
>
>
>
>
>
> On 18 December 2016 at 15:05, Loet Leydesdorff <l...@leydesdorff.net>
> wrote:
>
> Dear James and colleagues,
>
>
>
> Weaver (1949) made two major remarks about his coauthor (Shannon)'s
> contribution:
>
>
>
> 1. the definition of information as uncertainty is counter-intuitive
> ("bizarre"); (p. 27)
>
> 2. "In particular, information must not be confused with meaning." (p. 8)
>
>
>
> The definition of information as relevant for a system of reference
> confuses information with "meaningful information" and thus sacrifices the
> surplus value of Shannon's counter-intuitive definition.
>
>
>
> information observer
>
>
>
> that integrates interactive processes such as
>
>
>
> physical interactions such photons stimulating the retina of the eye,
> human-machine interactions (this is the level that Shannon lives on),
> biological interaction such body temperature relative to touch ice or heat
> source, social interaction such as this forum started by Pedro, economic
> interaction such as the stock market, ... [Lerner, page 1].
>
>
>
> We are in need of a theory of meaning. Otherwise, one cannot measure
> meaningful information. In a previous series of communications we discussed
> redundancy from this perspective.
>
>
>
> Lerner introduces mathematical expectation E[Sap] (difference between of a
> priory entropy [sic] and a posteriori entropy), which is distinguished from
> the notion of relative information Iap (Learner, page 7).
>
>
>
> ) expresses in bits of information the information generated when the a
> priori distribution is turned into the a posteriori one . This follows
> within the Shannon framework without needing an observer. I use this
> equation, for example, in my 1995-book *The Challenge of Scientometrics* 
> (Chapters
> 8 and 9), with a reference to Theil (1972). The relative information is
> defined as the *H*/*H*(max).
>
>
>
> I agree that the intuitive notion of information is derived from the Latin
> “in-formare” (Varela, 1979). But most of us do no longer use “force” and
> “mass” in the intuitive (Aristotelian) sense. J The proliferation of the
> meanings of information if confused with “meaningful information” is
> indicative for an “index sui et falsi”, in my opinion. The repetitive
> discussion lames the progression at this list. It is “like asking whether a
> glass is half empty or half full” (Hayles, 1990, p. 59).
>
>
>
> This act of forming forming an information process results in the
> construction of an observer that is the owner [holder] of information.
>
>
>
> The system of reference is then no longer the message, but the observer
> who provides meaning to the information (uncertainty). I agree that this is
> a selection process, but the variation first has to be specified
> independently (before it can be selected.
>
>
>
> And Lerner introduces the threshold between objective and subjective
> observes (page 27).   This leads to a consideration selection and
> cooperation that includes entanglement.
>
>
>
> I don’t see a direct relation between information and entanglement. An
> observer can be entangled.
>
>
>
> Best,
>
> Loet
>
>
>
> PS. Pedro: Let me assume that this is my second posting in the week which
> ends tonight. L.
>
>
>
>
> _______________________________________________
> Fis mailing list
> Fis@listas.unizar.es
> http://listas.unizar.es/cgi-bin/mailman/listinfo/fis
>
>
>
>
>
> --
>
>
> 4 Austin Dr. Prior Park St. James, Barbados BB23004
> Tel:   246-421-8855 <(246)%20421-8855>
> Cell:  246-243-5938 <(246)%20243-5938>
>
> _______________________________________________
> Fis mailing list
> Fis@listas.unizar.es
> http://listas.unizar.es/cgi-bin/mailman/listinfo/fis
>
>
>
>
> _______________________________________________
> Fis mailing list
> Fis@listas.unizar.es
> http://listas.unizar.es/cgi-bin/mailman/listinfo/fis
>
>
>
>
>
> --
>
> Dr. Mark William Johnson
> Institute of Learning and Teaching
>
> Faculty of Health and Life Sciences
>
> University of Liverpool
>
>
>
> Phone: 07786 064505
> Email: johnsonm...@gmail.com
> Blog: http://dailyimprovisation.blogspot.com
>
>
> _______________________________________________
> Fis mailing list
> Fis@listas.unizar.es
> http://listas.unizar.es/cgi-bin/mailman/listinfo/fis
>
>
>
>
>
> --
>
> Professor Terrence W. Deacon
> University of California, Berkeley
>
> _______________________________________________
> Fis mailing list
> Fis@listas.unizar.es
> http://listas.unizar.es/cgi-bin/mailman/listinfo/fis
>
>
_______________________________________________
Fis mailing list
Fis@listas.unizar.es
http://listas.unizar.es/cgi-bin/mailman/listinfo/fis

Reply via email to