Dear Mark and Colleagues (for which latter I attach Mark's message),

Thank you, Mark, for reminding us of your discussion, especially, on p. 104 
"energy is a kind of information". As we have discussed previously, however, 
adding that "information is a kind of energy" does not resolve the problem 
completely:

1. You yourself define information as 'multi-faceted', but this begs the 
question of which facet energy is a kind.

2. We may agree that, if they are not identical, energy and information always 
accompany one another and may have emerged together from some as yet 
incompletely defined substrate. However, they may not be, do not have to be and 
for me are not at the same ontological level, and energy is primary being less 
abstract.

3. Even the sophisticated mathematics of your Theory of Information does not 
capture, any more than any other mathematical theory, the complex qualitative 
properties of information processes for which no algorithm can be written. On 
the other hand, these processes do follow the logical laws for the evolution of 
energetic processes. To state this in other terms, whatever the entire complex 
set of properties of information, one of them is that it reflects the 
underlying duality of a universe constituted by physical matter-energy.

4. As a further example of the importance of ontological priority, one that is 
consistent with 'informational thinking', I mention the theory of Michael 
Graziano, Consciousness and the Social Brain, Oxford, 2013*. According to this, 
consciousness does not result in attention but is a consequence of the 
reflexive processing of information obtained through attention. The 'hard 
problem' of an isolated, subjective experience falls out. This is consistent 
with the principles of my logic which I will not repeat here.

Comments welcome,

Joseph

*Reviewed in SCIENCE, 345:6193, p. 147, 11 July 2014
                            ----- Original Message ----- 
  From: Burgin, Mark 
  To: Joseph Brenner 
  Sent: Friday, August 01, 2014 9:19 PM
  Subject: Re: [Fis] Krassimir's Information Quadruple and GIT. Quintuples?


  Dear Joseph and Colleagues,
  An answer to "the perhaps badly posed question of whether information or 
energy is more fundamental" is given in the book M.Burgin, Theory of 
information. The answer is a little bit unexpected:
  Energy is a kind of information and information is a kind of energy.
  It's a pity that very few researchers read books with advanced theories based 
on sophisticated mathematics.

   Sincerely,
  Mark Burgin




  On 7/31/2014 2:40 AM, Joseph Brenner wrote:

    Dear Krassimir and Colleagues,

    I have followed this discussion with interest but not total agreement. As I 
have commented to Krassimir previously, I feel that his system, based on 
symbols as outlined in his paper, is too static to capture the dynamics of 
complex information processes and their value (valence). It suffers from the 
same problems as that of Peirce and of set-theoretic approaches, namely, a 
certain arbitrariness in the selection and number of independent elements and 
their grounding in nature (or rather absence of grounding).

    If you will permit a naïve but well-intentioned question, why not have a 
theory whose elements are quintuples? Would this not be a 'better', more 
complete theory? This opens the possibility of an infinite regress, but that is 
the point I am trying to make: the form of the theory is, to a certain extent, 
defining its content. 

    The /development/ of any GIT should, from the beginning I think, recognize 
the existence in real time, so to speak, of any new suggestions in the context 
of other recent contributions of a different form, such as those of Luhn, 
Hofkirchner, Marijuan, Deacon, Dodig-Crnkovic, Wu and so on. Several of these 
already permit a more directed discussion of the perhaps badly posed question 
of whether information or energy is more fundamental. Otherwise, all that work 
will need to be done at the end. In any case, the GIT itself, to the extent 
that it could be desirable and useful, would also have to have some dynamics 
capable of accepting theories of different forms. 20th Century physics sought 
only identities throughout nature and the balance is now being somewhat 
restored. I think keeping the diversity of theories of information in mind is 
the most worthwhile strategy. 

    One of the values of Krassimir's approach is that it recognizes the 
existence of some of these more complex questions that need to be answered. I 
simply suggest that process language and a recognition of dynamic interactions 
(e.g., between 'internal' and 'external') could be part of the strategy.

    Best wishes,

    Joseph




      ----- Original Message ----- 
      From: Krassimir Markov 
      To: Jerry LR Chandler ; FIS ; Pridi Siregar 
      Sent: Saturday, July 26, 2014 10:42 AM
      Subject: [Fis] Information quadruple


      Dear Jerry, Pridi, and Colleagues,

      Thank you for the nice comments!

      To answer to questions I have to present next step from the GIT (General 
Information Theory) we are developing.

      Let remember in words (below "Infos" is abbreviation from "Information 
Subject", it is an intelligent natural or artificial agent (system)):

      Information is quadruple (Source, Recipient, Evidence, Infos) or formally 
i = (s, r, e, I) 

      The nest step is to define elements of the quadruple:

      s and r are structured sets;
      e is a mapping from s in r which preserves (all or partial) structure of 
s and resolves any information expectation of I

      I expect new questions:
      - what is an "intelligent agent"
      - what is "information expectation"
      - ...

      If it is interesting, answers to these questions may be given in further 
letters.

      ***

      Now I want to make some comments to letters received (their full texts 
are given below my answers).

      Pridi: "information cannot be viewed in any absolute sense but as 
internal representations of "external patterns""
      Kr.:  Yes, the "reflection" is a property of Matter, "information" is a 
reflection for which the information quadruple exists. But information is not 
"internal representations of "external patterns" ". It is result from resolving 
the subjective information expectation which is process of comparing of 
internal and external patterns. I know, this will cause new questions 

      Pridi: In this framework then, it seems that "information" cannot be 
conceptualized without reference to the both "something out there" and the 
"internal structures" of the receptor/cognitive system. 
      Kr.: Yes.

      Pridi: How can we really quantify meaningful (semantic) information ... ?
      Kr.: By distance between "external patterns" and "information 
expectation" (sorry to be not clear but it is long text for further letters).

      Pridi: All "objective" measures (entropy, negentropy,...) are actually 
totally dependant of I1 and I2 and can never be considered as "absolute". 
      Kr.: Yes, but the world humanity is an Infos and its information 
expectations we assume as "absolute".

      Pridi: ... some researchers that posit that "information" may be more 
fundamental than the fundamental physical (mass, time, space, amps). 
      Kr.: Yes, there are other paradigms which are useful in some cases, but 
in our paradigm "information" is not fundamental but "reflection" is the 
fundamental.

      Pridi: ... no "absolute truth" (whatever this means) is really gained. 
"Only" a richer more complete (subjective but coherent) world-view .
      Kr.: Yes.

      Jerry: ... assertion of a quadruple of symbols is rather close to the 
philosophy of C S Peirce (hereafter "CSP") 
      Kr.: Our paradigm is nor opposite to what science has explored till now. 
All already investigated information theories (Shannon,Peirce, etc) have to be 
a part or intersection of a new GIT.

      Jerry: ... moves these 'definitions' of individual symbols into the 
subjective realm. (CSP's notion of "interpretation?)
      Different researchers have the freedom to interpret the evidence as they 
choose, including the relationships to engineering terms such as "bandwidth".
      Kr.: Yes. But not only researches, everybody has such freedom. Because of 
this there exist advertising processes ... but for this we have to talk in 
further letters. 

      Jerry: Pridi's post appropriately recognizes the tension between 
objective scientific theories and subjective judgments about evidence by 
different  individuals with different professional backgrounds and different 
symbolic processing powers. 
      Kr.: Yes, there will be tension if we assume world as plane structure. 
But it is hierarchical one and what is assumed as "subjective" at one level is 
assumed as "objective" for the low levels.

      Jerry: ... to show that these definitions of symbols motivate a coherent 
symbol system that can be used to transfer information contained in the signal 
from symbolic representations of entities. It may work for engineering 
purposes, but is it extendable to life?
      Kr.: The goal of work on GIT is to create a coherent symbol system which 
is equal valid for life creatures and artificial agents.

      Jerry: ... this requires the use of multiple symbol systems and multiple 
forms of logic in order to gain the functionality of transfer of "in-form" 
between individuals or machines.
      Kr.: Yes, at least on three levels - Information, Infos, Inforaction 
(Information interaction)

      Jerry: Anybody have any suggestions on how this quadruple of symbols can 
be formalized into a quantitative coherent form of communication?
      Kr.: A step toward this I give above in the beginning of this letter but 
it is very long journey ...

      Thank you for creative discussion!
      Friendly regards
      Krassimir





      -----Original Message----- 
      From: Jerry LR Chandler 
      Sent: Wednesday, July 23, 2014 8:57 PM 
      To: FIS 
      Cc: Krassimir Markov ; Pridi Siregar 
      Subject: Re: [Fis] Re to Pridi: infinite bandwith and finite 
informationcontent CS Peirce and Chemical Nomenclature 

      Pridi, Krassimir,  List:

      (In order to place this comment in context, and for reference, I have 
copied Krassimir's "definition" of information below. My comments follow the 
excellent post of Pridi.)

      > In physical world there exist only reflections but not information. 
      > 
      > Information " i " is the quadruple: 
      > i = (s, r, e, I) 
      > where 
      > s is a source entity, which is reflected in r 
      > r is the entity in which reflection of s exists 
      > e is an evidence for the subject I which proofs for him and only for 
him that the reflection in r reflects just s , i.e. the evidence proofs for the 
subject what the reflection reflects . 
      > I is information subject who has possibility to make decisions in 
accordance with some goals - human, animal, bacteria, artificial intelligent 
system, etc. 
      > 
      > In other words, information is a reflection, but not every reflection 
is information - only reflections for which the quadruple above exist are 
assumed as information by the corresponded subjects. 
      > 
      > For different I , information may be different because of subjects' 
finite memory and reflection possibilities. 
      > Because of this, a physical event with an infinite bandwidth may have 
finite information content (for concrete information subject) . 
      On Jul 23, 2014, at 6:45 AM, Pridi Siregar wrote:

      > Dear Krassimir,
      > 
      > Thank you for your explanation. It does give me a better understanding 
of how information (beyond Shannon) can be formalized! However, a closer look 
at the formalism and its semantic does raise new questions:
      > 
      > From the definition you have given, it appears that information cannot 
be viewed in any absolute sense but as internal representations of "external 
patterns" whose meaning depends on the subject capturing/interpreting/storing 
the said patterns. In this framework then, it seems that "information" cannot 
be conceptualized without reference to the both "something out there" and the 
"internal structures" of the receptor/cognitive system. 
      > 
      > In other words the concept of "information" lies within some 
"subjective" (albeit rational) realm. I'm sure that I'm stating the obvious for 
most of FIS members but a question arised upon reading your formalism: How can 
we really quantify meaningful (semantic) information beyond Shannon (that 
disregards semantics) and his purely statistical framework? Or beyond 
Boltzmann's entropy/Information based on micro-macro states ratios?
      > 
      > When we formalize i = (s, r, e, I) there is  a "meta-level" 
formalisation that is only apparent since even (s,r) reflect our own (human) 
subjective world-view. We could actually write (I1(s), I1(r), e, I2) where I1 
and I2 are two distinct cognitive systems and both of which lie at the OBJECT 
level of the formalizing agent which is NEITHER I1 or I2. All "objective" 
measures (entropy, negentropy,...) are actually totally dependant of I1 and I2 
and can never be considered as "absolute". 
      > 
      > 
      > This leads me to a second question (sorry for the lengthy message): 
there are some researchers that posit that "information" may be more 
fundamental than the fundamental physical (mass, time, space, amps). This 
appears (and perhaps only appears) to be at the opposite end of the 
above-mentioned view. Indeed, in this framework some kind of "universal" or 
"absolute" notions must be accepted as true.
      > 
      > One apparent way out would be to demonstrate that information somehow 
logically entails the fundemantal physical entities while accepting that we are 
still within a human-centered  world view. And thus no "absolute truth" 
(whatever this means) is really gained. "Only" a richer more complete 
(subjective but coherent) world-view .
      > 
      > Am I making anys sense? Any thoughts?
      > 
      > Best
      > 
      > Pridi         
      > 

      Pridi's comment concur with many of my views wrt the concept of 
information. 

      Krassimir's assertion of a quadruple of symbols is rather close to the 
philosophy of C S Peirce (hereafter "CSP") in one context.

      S as symbol represents an external source of signal, that which is 
independent of the individual mind and being.  This is analogous to CSP's term 
"sinsign".

      R is a thing itself.  That is, R generates S.

      E as evidence is a vague term which infers an observer (2nd Order 
Cybernetics?) that both receives and evaluates the signal (S) from the thing 
(R).  CSP categorizes evidence as icon, index or symbol with respect to the 
entity of observation.

      I  as Krassimirian information is a personal judgment about the evidence. 
 (Correspondence with CSP's notion of "argument" is conceivable.) 

      Krassimir's assertion that: 
      > For different I , information may be different because of subjects' 
finite memory and reflection possibilities. 
      > Because of this, a physical event with an infinite bandwidth may have 
finite information content (for concrete information subject) . 


      moves these 'definitions' of individual symbols into the subjective 
realm. (CSP's notion of "interpretation?)
      Different researchers have the freedom to interpret the evidence as they 
choose, including the relationships to engineering terms such as "bandwidth".


      Pridi's post appropriately recognizes the tension between objective 
scientific theories and subjective judgments about evidence by different  
individuals with different professional backgrounds and different symbolic 
processing powers.

      The challenge for Krassimirian information, it appears to me, is to show 
that these definitions of symbols motivate a coherent symbol system that can be 
used to transfer information contained in the signal from symbolic 
representations of entities. It may work for engineering purposes, but is it 
extendable to life?

      (For me, of course, this requires the use of multiple symbol systems and 
multiple forms of logic in order to gain the functionality of transfer of 
"in-form" between individuals or machines.)

      Pridi writes:
      >  How can we really quantify meaningful (semantic) information beyond 
Shannon (that disregards semantics) and his purely statistical framework?

      One aspect of this conundrum was solved by chemists over the past to two 
centuries by developing a unique symbol system that is restricted by physical 
constraints, yet functions as an exact mode of communication. 

      Chemical notation, as symbol system, along with mathematics and data, 
achieves this end purpose (entelechy) of communication, for some entities, such 
as the meaning of an "atomic number" as a relational term and hence the meaning 
of a particular integer as both quantity and quality. 

      This requires a dyadic mathematics and synductive logic for sublations.


      Pridi writes:

      > It does give me a better understanding of how information (beyond 
Shannon) can be formalized! 

      Can you communicate how this "better understanding...   ... foramlized"  
works?  

      It is not readily apparent to me how Krassimirian information can be 
formalized.

      Anybody have any suggestions on how this quadruple of symbols can be 
formalized into a quantitative coherent form of communication?

      Cheers

      Jerry 








--------------------------------------------------------------------------
      _______________________________________________
      Fis mailing list
      Fis@listas.unizar.es
      http://listas.unizar.es/cgi-bin/mailman/listinfo/fis


     

_______________________________________________
Fis mailing list
Fis@listas.unizar.es
http://listas.unizar.es/cgi-bin/mailman/listinfo/fis

_______________________________________________
Fis mailing list
Fis@listas.unizar.es
http://listas.unizar.es/cgi-bin/mailman/listinfo/fis

Reply via email to