Hah... well never mind... I thought about this a bunch more this
morning and it's complicated; I'll ruminate more and come to an actual
considered conclusion before polluting mailing lists with more of my
half-baked thoughts on the topic ;D ...

I've been re-reading this nice old paper on the foundations of the Second Law..

http://necsi.edu/projects/baranger/cce.pdf

-- ben

On Wed, Nov 26, 2014 at 12:51 AM, Ben Goertzel <[email protected]> wrote:
> I.e., what would be a case where:
>
> -- X, Y and Z have positive multivariate mutual information
>
> -- Y and Z are common causes of X
>
> -- Y and Z are *not* correlated, i.e. they don't have positive binary
> mutual information
>
> ?
>
>
> On Wed, Nov 26, 2014 at 12:46 AM, Ben Goertzel <[email protected]> wrote:
>> Martin,
>>
>> Hmmm -- This is "explaining away" as discussed in the Bayes Nets
>> literature, right?
>>
>> https://en.wikipedia.org/wiki/Multivariate_mutual_information#Example_of_Negative_Multivariate_mutual_information
>>
>> The examples on that wiki page are ones where common cause corresponds
>> to positive MMI, and common effect corresponds to negative MMI.  I
>> wonder how generally that holds?
>>
>> ben
>>
>> On Wed, Nov 26, 2014 at 12:35 AM, martin biehl <[email protected]> wrote:
>>> "Having a common effect does not induce correlation between events,
>>> while having a common cause does."
>>>
>>> is possibly not always true, take the infamous XOR gate, here if the output
>>> (effect) is known to be 1 then this implies a correlation between the two
>>> inputs i.e. they must be opposites.
>>>
>>> is this a counterexample?
>>>
>>> On Tue, Nov 25, 2014 at 11:53 AM, Ben Goertzel <[email protected]> wrote:
>>>>
>>>> Hmmm...
>>>>
>>>> Having thought about this more, while I was indeed traveling backwards
>>>> in time when I wrote the previous email, it's not too relevant anyhow
>>>> because the Second Law only holds globally, and in complex systems
>>>> there are many subsystems that are behaving anti-entropically.  So I'm
>>>> no sure one can use the law of entropy increase to draw conclusions
>>>> about local causality.
>>>>
>>>> However, I was thinking about section 6.3.2 of
>>>>
>>>> http://cqi.inf.usi.ch/qic/94_Lloyd.pdf
>>>>
>>>> where Seth Lloyd observes that
>>>>
>>>> "Having a common effect does not induce correlation between events,
>>>> while having a common cause does."
>>>>
>>>> I.e.
>>>>
>>>> -- In the case of two causes with a common effect ... there is an
>>>> increase of information from past to future (the probability spread
>>>> across two causes is now concentrated on a single effect).   There no
>>>> correlation in the past (between the causes).   This is the opposite
>>>> direction of the Second Law of Thermodynamics.
>>>>
>>>> -- In the case of two effects with a common cause ...  there is a
>>>> decrease of information from past to future (the probability
>>>> concentrated in one cause is now spread across two effects).   There
>>>> is correlation in the future (between the effects).  This is in the
>>>> direction of the Second Law of Thermodynamics.
>>>>
>>>> ...
>>>>
>>>> I.e. in many cases the direction of causal influence may be
>>>> identifiable as the direction of increasing correlation....   I'm not
>>>> sure exactly what are the limits of this conclusion though.
>>>>
>>>> ...
>>>>
>>>> Soo --   What if one has two sets of variables, S and T, and there is
>>>> significant mutual information between the values of S and the values
>>>> of T, as evaluated across different cases...?   So, suppose we have
>>>> both
>>>>
>>>> S --> T
>>>>
>>>> and
>>>>
>>>> T --> S
>>>>
>>>> in a sense....    But, if there is significantly more correlation
>>>> among the variables within T, than among the variables within S, then
>>>> we can say that it's more likely that T is the effect and S is the
>>>> cause...
>>>>
>>>> The asymmetry used to identify causation is then one of correlation
>>>> rather than of temporality directly...
>>>>
>>>> This may be a way of heuristically inferring causality from
>>>> non-temporal data, if one has a sufficient ensemble of data samples...
>>>>
>>>> -- Ben
>>>>
>>>>
>>>> On Tue, Nov 25, 2014 at 1:46 PM, Ben Goertzel <[email protected]> wrote:
>>>> >
>>>> > Hmm, maybe you're right , maybe I was traveling backwards in time when I
>>>> > wrote that ...
>>>> >
>>>> > (More later)
>>>> >
>>>> > On Tuesday, November 25, 2014, martin biehl <[email protected]> wrote:
>>>> >>
>>>> >> hm, sounds interesting, but I don't get it either. If entropy
>>>> >> increases,
>>>> >> the uncertainty of the state increases and information (about the
>>>> >> state)
>>>> >> decreases as you say, but why would the past then contain more
>>>> >> information
>>>> >> about the future than vice versa? Let X be the past, Y be the future,
>>>> >> then
>>>> >> as mutual information is symmetric:
>>>> >> H(X) - H(X|Y) = H(Y) - H(Y|X)
>>>> >> now H(Y) > H(X) because of entropy increase.
>>>> >> then
>>>> >> H(Y|X) > H(X|Y)
>>>> >> and the future should be more uncertain given the past than vice versa.
>>>> >> Where did this go wrong?
>>>> >>
>>>> >>
>>>> >> On Tue, Nov 25, 2014 at 2:13 AM, Ben Goertzel via AGI <[email protected]>
>>>> >> wrote:
>>>> >>>
>>>> >>> Information is negentropy, so increase of entropy implies decrease of
>>>> >>> information...
>>>> >>>
>>>> >>> Acquiring information about a system is associated with entropy
>>>> >>> production...
>>>> >>>
>>>> >>> On Tue, Nov 25, 2014 at 9:59 AM, Aaron Nitzkin <[email protected]>
>>>> >>> wrote:
>>>> >>> > Sorry, I must be a little confused -- probably thinking from the
>>>> >>> > wrong
>>>> >>> > perspective . . . I would think that there is more information
>>>> >>> > in the future about the past than vice versa, because we know more
>>>> >>> > about the
>>>> >>> > past than we do about the future, and also, doesn't
>>>> >>> > increase in entropy imply increase in information (because it
>>>> >>> > requries
>>>> >>> > more
>>>> >>> > information to specify the configuration of a system
>>>> >>> > with higher entropy than the same system with lower entropy?)
>>>> >>> >
>>>> >>> > On Tue, Nov 25, 2014 at 8:27 AM, Ben Goertzel <[email protected]>
>>>> >>> > wrote:
>>>> >>> >>
>>>> >>> >> In the early part of the paper, the author clarifies that while he
>>>> >>> >> assumes "temporal precedence as an aspect of causality" for
>>>> >>> >> simplicity, actually his approach would work with any other
>>>> >>> >> systematic
>>>> >>> >> way of assigning asymmetric directions to relationships between
>>>> >>> >> events
>>>> >>> >>
>>>> >>> >> I have been thinking a lot about how to infer causality from
>>>> >>> >> non-time-series data (e.g. categorial gene expression data), and
>>>> >>> >> this
>>>> >>> >> is a case where looking at some other sort of asymmetry than
>>>> >>> >> temporal
>>>> >>> >> precedence (but that may generally correlated with temporal
>>>> >>> >> precedence) seems to make sense.   E.g. I've been thinking about
>>>> >>> >> looking at informational asymmetry: If one has P(A = a | B=b), one
>>>> >>> >> can
>>>> >>> >> look at whether the distribution for A gives more information about
>>>> >>> >> the distribution for B, or vice versa.   This informational
>>>> >>> >> asymmetry
>>>> >>> >> can be used similarly to temporal asymmetry in defining causality.
>>>> >>> >> Furthermore, it on the average is going to correlate with temporal
>>>> >>> >> asymmetry, because the past tends to contain more information about
>>>> >>> >> the future than vice versa (due to entropy increase, roughly
>>>> >>> >> speaking... but there's more story here...)
>>>> >>> >>
>>>> >>> >> -- Ben
>>>> >>> >>
>>>> >>> >>
>>>> >>> >> On Tue, Nov 25, 2014 at 5:34 AM, Michael van der Gulik
>>>> >>> >> <[email protected]> wrote:
>>>> >>> >> > "Chapter 1. Quantum mechanics... "
>>>> >>> >> >
>>>> >>> >> > It's a nice article; I'll add it to my reading list. Prediction
>>>> >>> >> > involves
>>>> >>> >> > working out what causes what, so it's pretty fundamental.
>>>> >>> >> >
>>>> >>> >> > I have a question. Causation in my mind seems to always involve
>>>> >>> >> > time,
>>>> >>> >> > and I
>>>> >>> >> > suspect it's impossible to have causation without including
>>>> >>> >> > timing.
>>>> >>> >> > So...
>>>> >>> >> >
>>>> >>> >> > Is it possible for a cause to happen at exactly the same moment
>>>> >>> >> > as
>>>> >>> >> > its
>>>> >>> >> > effect?
>>>> >>> >> >
>>>> >>> >> > Is it possible for a cause to happen after its effect?
>>>> >>> >> >
>>>> >>> >> > One instance I'm trying to get my head around is when an
>>>> >>> >> > intelligence
>>>> >>> >> > anticipates a cause (which is an event in the future), which
>>>> >>> >> > results
>>>> >>> >> > in
>>>> >>> >> > the
>>>> >>> >> > intelligence acting such that the effect occurs before the cause.
>>>> >>> >> > Perhaps
>>>> >>> >> > the anticipation itself is the causal event.
>>>> >>> >> >
>>>> >>> >> > Regards,
>>>> >>> >> > Michael.
>>>> >>> >> >
>>>> >>> >> >
>>>> >>> >> > On Sun, Nov 23, 2014 at 7:17 AM, Ben Goertzel <[email protected]>
>>>> >>> >> > wrote:
>>>> >>> >> >>
>>>> >>> >> >> I just happened across this 2011 paper on the probabilistic
>>>> >>> >> >> foundation
>>>> >>> >> >> of causality,
>>>> >>> >> >>
>>>> >>> >> >> http://philsci-archive.pitt.edu/9729/1/Website_Version_2.pdf
>>>> >>> >> >>
>>>> >>> >> >> which seems to carefully clarify a bunch of issues that remain
>>>> >>> >> >> dangling in prior discussions of the topic
>>>> >>> >> >>
>>>> >>> >> >> It seems to give a good characterization of what it means for "P
>>>> >>> >> >> to
>>>> >>> >> >> appear to cause Q, based on the knowledge-base of observer O"
>>>> >>> >> >>
>>>> >>> >> >> --
>>>> >>> >> >> Ben Goertzel, PhD
>>>> >>> >> >> http://goertzel.org
>>>> >>> >> >>
>>>> >>> >> >> "The reasonable man adapts himself to the world: the
>>>> >>> >> >> unreasonable
>>>> >>> >> >> one
>>>> >>> >> >> persists in trying to adapt the world to himself. Therefore all
>>>> >>> >> >> progress depends on the unreasonable man." -- George Bernard
>>>> >>> >> >> Shaw
>>>> >>> >> >>
>>>> >>> >> >> --
>>>> >>> >> >> You received this message because you are subscribed to the
>>>> >>> >> >> Google
>>>> >>> >> >> Groups
>>>> >>> >> >> "Artificial General Intelligence" group.
>>>> >>> >> >> To unsubscribe from this group and stop receiving emails from
>>>> >>> >> >> it,
>>>> >>> >> >> send
>>>> >>> >> >> an
>>>> >>> >> >> email to
>>>> >>> >> >> [email protected].
>>>> >>> >> >> For more options, visit https://groups.google.com/d/optout.
>>>> >>> >> >
>>>> >>> >> >
>>>> >>> >> >
>>>> >>> >> >
>>>> >>> >> > --
>>>> >>> >> > http://gulik.pbwiki.com/
>>>> >>> >> >
>>>> >>> >> > --
>>>> >>> >> > You received this message because you are subscribed to the
>>>> >>> >> > Google
>>>> >>> >> > Groups
>>>> >>> >> > "Artificial General Intelligence" group.
>>>> >>> >> > To unsubscribe from this group and stop receiving emails from it,
>>>> >>> >> > send
>>>> >>> >> > an
>>>> >>> >> > email to
>>>> >>> >> > [email protected].
>>>> >>> >> > For more options, visit https://groups.google.com/d/optout.
>>>> >>> >>
>>>> >>> >>
>>>> >>> >>
>>>> >>> >> --
>>>> >>> >> Ben Goertzel, PhD
>>>> >>> >> http://goertzel.org
>>>> >>> >>
>>>> >>> >> "The reasonable man adapts himself to the world: the unreasonable
>>>> >>> >> one
>>>> >>> >> persists in trying to adapt the world to himself. Therefore all
>>>> >>> >> progress depends on the unreasonable man." -- George Bernard Shaw
>>>> >>> >>
>>>> >>> >> --
>>>> >>> >> You received this message because you are subscribed to the Google
>>>> >>> >> Groups
>>>> >>> >> "opencog" group.
>>>> >>> >> To unsubscribe from this group and stop receiving emails from it,
>>>> >>> >> send
>>>> >>> >> an
>>>> >>> >> email to [email protected].
>>>> >>> >> To post to this group, send email to [email protected].
>>>> >>> >> Visit this group at http://groups.google.com/group/opencog.
>>>> >>> >> For more options, visit https://groups.google.com/d/optout.
>>>> >>> >
>>>> >>> >
>>>> >>>
>>>> >>>
>>>> >>>
>>>> >>> --
>>>> >>> Ben Goertzel, PhD
>>>> >>> http://goertzel.org
>>>> >>>
>>>> >>> "The reasonable man adapts himself to the world: the unreasonable one
>>>> >>> persists in trying to adapt the world to himself. Therefore all
>>>> >>> progress depends on the unreasonable man." -- George Bernard Shaw
>>>> >>>
>>>> >>>
>>>> >>> -------------------------------------------
>>>> >>> AGI
>>>> >>> Archives: https://www.listbox.com/member/archive/303/=now
>>>> >>> RSS Feed:
>>>> >>> https://www.listbox.com/member/archive/rss/303/10872673-8f99760d
>>>> >>> Modify Your Subscription:
>>>> >>>
>>>> >>> https://www.listbox.com/member/?&;
>>>> >>> Powered by Listbox: http://www.listbox.com
>>>> >>
>>>> >>
>>>> >
>>>> >
>>>> > --
>>>> > Ben Goertzel, PhD
>>>> > http://goertzel.org
>>>> >
>>>> > "The reasonable man adapts himself to the world: the unreasonable one
>>>> > persists in trying to adapt the world to himself. Therefore all progress
>>>> > depends on the unreasonable man." -- George Bernard Shaw
>>>> >
>>>>
>>>>
>>>>
>>>> --
>>>> Ben Goertzel, PhD
>>>> http://goertzel.org
>>>>
>>>> "The reasonable man adapts himself to the world: the unreasonable one
>>>> persists in trying to adapt the world to himself. Therefore all
>>>> progress depends on the unreasonable man." -- George Bernard Shaw
>>>
>>>
>>
>>
>>
>> --
>> Ben Goertzel, PhD
>> http://goertzel.org
>>
>> "The reasonable man adapts himself to the world: the unreasonable one
>> persists in trying to adapt the world to himself. Therefore all
>> progress depends on the unreasonable man." -- George Bernard Shaw
>
>
>
> --
> Ben Goertzel, PhD
> http://goertzel.org
>
> "The reasonable man adapts himself to the world: the unreasonable one
> persists in trying to adapt the world to himself. Therefore all
> progress depends on the unreasonable man." -- George Bernard Shaw



-- 
Ben Goertzel, PhD
http://goertzel.org

"The reasonable man adapts himself to the world: the unreasonable one
persists in trying to adapt the world to himself. Therefore all
progress depends on the unreasonable man." -- George Bernard Shaw


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to