I.e., what would be a case where:

-- X, Y and Z have positive multivariate mutual information

-- Y and Z are common causes of X

-- Y and Z are *not* correlated, i.e. they don't have positive binary
mutual information

?


On Wed, Nov 26, 2014 at 12:46 AM, Ben Goertzel <[email protected]> wrote:
> Martin,
>
> Hmmm -- This is "explaining away" as discussed in the Bayes Nets
> literature, right?
>
> https://en.wikipedia.org/wiki/Multivariate_mutual_information#Example_of_Negative_Multivariate_mutual_information
>
> The examples on that wiki page are ones where common cause corresponds
> to positive MMI, and common effect corresponds to negative MMI.  I
> wonder how generally that holds?
>
> ben
>
> On Wed, Nov 26, 2014 at 12:35 AM, martin biehl <[email protected]> wrote:
>> "Having a common effect does not induce correlation between events,
>> while having a common cause does."
>>
>> is possibly not always true, take the infamous XOR gate, here if the output
>> (effect) is known to be 1 then this implies a correlation between the two
>> inputs i.e. they must be opposites.
>>
>> is this a counterexample?
>>
>> On Tue, Nov 25, 2014 at 11:53 AM, Ben Goertzel <[email protected]> wrote:
>>>
>>> Hmmm...
>>>
>>> Having thought about this more, while I was indeed traveling backwards
>>> in time when I wrote the previous email, it's not too relevant anyhow
>>> because the Second Law only holds globally, and in complex systems
>>> there are many subsystems that are behaving anti-entropically.  So I'm
>>> no sure one can use the law of entropy increase to draw conclusions
>>> about local causality.
>>>
>>> However, I was thinking about section 6.3.2 of
>>>
>>> http://cqi.inf.usi.ch/qic/94_Lloyd.pdf
>>>
>>> where Seth Lloyd observes that
>>>
>>> "Having a common effect does not induce correlation between events,
>>> while having a common cause does."
>>>
>>> I.e.
>>>
>>> -- In the case of two causes with a common effect ... there is an
>>> increase of information from past to future (the probability spread
>>> across two causes is now concentrated on a single effect).   There no
>>> correlation in the past (between the causes).   This is the opposite
>>> direction of the Second Law of Thermodynamics.
>>>
>>> -- In the case of two effects with a common cause ...  there is a
>>> decrease of information from past to future (the probability
>>> concentrated in one cause is now spread across two effects).   There
>>> is correlation in the future (between the effects).  This is in the
>>> direction of the Second Law of Thermodynamics.
>>>
>>> ...
>>>
>>> I.e. in many cases the direction of causal influence may be
>>> identifiable as the direction of increasing correlation....   I'm not
>>> sure exactly what are the limits of this conclusion though.
>>>
>>> ...
>>>
>>> Soo --   What if one has two sets of variables, S and T, and there is
>>> significant mutual information between the values of S and the values
>>> of T, as evaluated across different cases...?   So, suppose we have
>>> both
>>>
>>> S --> T
>>>
>>> and
>>>
>>> T --> S
>>>
>>> in a sense....    But, if there is significantly more correlation
>>> among the variables within T, than among the variables within S, then
>>> we can say that it's more likely that T is the effect and S is the
>>> cause...
>>>
>>> The asymmetry used to identify causation is then one of correlation
>>> rather than of temporality directly...
>>>
>>> This may be a way of heuristically inferring causality from
>>> non-temporal data, if one has a sufficient ensemble of data samples...
>>>
>>> -- Ben
>>>
>>>
>>> On Tue, Nov 25, 2014 at 1:46 PM, Ben Goertzel <[email protected]> wrote:
>>> >
>>> > Hmm, maybe you're right , maybe I was traveling backwards in time when I
>>> > wrote that ...
>>> >
>>> > (More later)
>>> >
>>> > On Tuesday, November 25, 2014, martin biehl <[email protected]> wrote:
>>> >>
>>> >> hm, sounds interesting, but I don't get it either. If entropy
>>> >> increases,
>>> >> the uncertainty of the state increases and information (about the
>>> >> state)
>>> >> decreases as you say, but why would the past then contain more
>>> >> information
>>> >> about the future than vice versa? Let X be the past, Y be the future,
>>> >> then
>>> >> as mutual information is symmetric:
>>> >> H(X) - H(X|Y) = H(Y) - H(Y|X)
>>> >> now H(Y) > H(X) because of entropy increase.
>>> >> then
>>> >> H(Y|X) > H(X|Y)
>>> >> and the future should be more uncertain given the past than vice versa.
>>> >> Where did this go wrong?
>>> >>
>>> >>
>>> >> On Tue, Nov 25, 2014 at 2:13 AM, Ben Goertzel via AGI <[email protected]>
>>> >> wrote:
>>> >>>
>>> >>> Information is negentropy, so increase of entropy implies decrease of
>>> >>> information...
>>> >>>
>>> >>> Acquiring information about a system is associated with entropy
>>> >>> production...
>>> >>>
>>> >>> On Tue, Nov 25, 2014 at 9:59 AM, Aaron Nitzkin <[email protected]>
>>> >>> wrote:
>>> >>> > Sorry, I must be a little confused -- probably thinking from the
>>> >>> > wrong
>>> >>> > perspective . . . I would think that there is more information
>>> >>> > in the future about the past than vice versa, because we know more
>>> >>> > about the
>>> >>> > past than we do about the future, and also, doesn't
>>> >>> > increase in entropy imply increase in information (because it
>>> >>> > requries
>>> >>> > more
>>> >>> > information to specify the configuration of a system
>>> >>> > with higher entropy than the same system with lower entropy?)
>>> >>> >
>>> >>> > On Tue, Nov 25, 2014 at 8:27 AM, Ben Goertzel <[email protected]>
>>> >>> > wrote:
>>> >>> >>
>>> >>> >> In the early part of the paper, the author clarifies that while he
>>> >>> >> assumes "temporal precedence as an aspect of causality" for
>>> >>> >> simplicity, actually his approach would work with any other
>>> >>> >> systematic
>>> >>> >> way of assigning asymmetric directions to relationships between
>>> >>> >> events
>>> >>> >>
>>> >>> >> I have been thinking a lot about how to infer causality from
>>> >>> >> non-time-series data (e.g. categorial gene expression data), and
>>> >>> >> this
>>> >>> >> is a case where looking at some other sort of asymmetry than
>>> >>> >> temporal
>>> >>> >> precedence (but that may generally correlated with temporal
>>> >>> >> precedence) seems to make sense.   E.g. I've been thinking about
>>> >>> >> looking at informational asymmetry: If one has P(A = a | B=b), one
>>> >>> >> can
>>> >>> >> look at whether the distribution for A gives more information about
>>> >>> >> the distribution for B, or vice versa.   This informational
>>> >>> >> asymmetry
>>> >>> >> can be used similarly to temporal asymmetry in defining causality.
>>> >>> >> Furthermore, it on the average is going to correlate with temporal
>>> >>> >> asymmetry, because the past tends to contain more information about
>>> >>> >> the future than vice versa (due to entropy increase, roughly
>>> >>> >> speaking... but there's more story here...)
>>> >>> >>
>>> >>> >> -- Ben
>>> >>> >>
>>> >>> >>
>>> >>> >> On Tue, Nov 25, 2014 at 5:34 AM, Michael van der Gulik
>>> >>> >> <[email protected]> wrote:
>>> >>> >> > "Chapter 1. Quantum mechanics... "
>>> >>> >> >
>>> >>> >> > It's a nice article; I'll add it to my reading list. Prediction
>>> >>> >> > involves
>>> >>> >> > working out what causes what, so it's pretty fundamental.
>>> >>> >> >
>>> >>> >> > I have a question. Causation in my mind seems to always involve
>>> >>> >> > time,
>>> >>> >> > and I
>>> >>> >> > suspect it's impossible to have causation without including
>>> >>> >> > timing.
>>> >>> >> > So...
>>> >>> >> >
>>> >>> >> > Is it possible for a cause to happen at exactly the same moment
>>> >>> >> > as
>>> >>> >> > its
>>> >>> >> > effect?
>>> >>> >> >
>>> >>> >> > Is it possible for a cause to happen after its effect?
>>> >>> >> >
>>> >>> >> > One instance I'm trying to get my head around is when an
>>> >>> >> > intelligence
>>> >>> >> > anticipates a cause (which is an event in the future), which
>>> >>> >> > results
>>> >>> >> > in
>>> >>> >> > the
>>> >>> >> > intelligence acting such that the effect occurs before the cause.
>>> >>> >> > Perhaps
>>> >>> >> > the anticipation itself is the causal event.
>>> >>> >> >
>>> >>> >> > Regards,
>>> >>> >> > Michael.
>>> >>> >> >
>>> >>> >> >
>>> >>> >> > On Sun, Nov 23, 2014 at 7:17 AM, Ben Goertzel <[email protected]>
>>> >>> >> > wrote:
>>> >>> >> >>
>>> >>> >> >> I just happened across this 2011 paper on the probabilistic
>>> >>> >> >> foundation
>>> >>> >> >> of causality,
>>> >>> >> >>
>>> >>> >> >> http://philsci-archive.pitt.edu/9729/1/Website_Version_2.pdf
>>> >>> >> >>
>>> >>> >> >> which seems to carefully clarify a bunch of issues that remain
>>> >>> >> >> dangling in prior discussions of the topic
>>> >>> >> >>
>>> >>> >> >> It seems to give a good characterization of what it means for "P
>>> >>> >> >> to
>>> >>> >> >> appear to cause Q, based on the knowledge-base of observer O"
>>> >>> >> >>
>>> >>> >> >> --
>>> >>> >> >> Ben Goertzel, PhD
>>> >>> >> >> http://goertzel.org
>>> >>> >> >>
>>> >>> >> >> "The reasonable man adapts himself to the world: the
>>> >>> >> >> unreasonable
>>> >>> >> >> one
>>> >>> >> >> persists in trying to adapt the world to himself. Therefore all
>>> >>> >> >> progress depends on the unreasonable man." -- George Bernard
>>> >>> >> >> Shaw
>>> >>> >> >>
>>> >>> >> >> --
>>> >>> >> >> You received this message because you are subscribed to the
>>> >>> >> >> Google
>>> >>> >> >> Groups
>>> >>> >> >> "Artificial General Intelligence" group.
>>> >>> >> >> To unsubscribe from this group and stop receiving emails from
>>> >>> >> >> it,
>>> >>> >> >> send
>>> >>> >> >> an
>>> >>> >> >> email to
>>> >>> >> >> [email protected].
>>> >>> >> >> For more options, visit https://groups.google.com/d/optout.
>>> >>> >> >
>>> >>> >> >
>>> >>> >> >
>>> >>> >> >
>>> >>> >> > --
>>> >>> >> > http://gulik.pbwiki.com/
>>> >>> >> >
>>> >>> >> > --
>>> >>> >> > You received this message because you are subscribed to the
>>> >>> >> > Google
>>> >>> >> > Groups
>>> >>> >> > "Artificial General Intelligence" group.
>>> >>> >> > To unsubscribe from this group and stop receiving emails from it,
>>> >>> >> > send
>>> >>> >> > an
>>> >>> >> > email to
>>> >>> >> > [email protected].
>>> >>> >> > For more options, visit https://groups.google.com/d/optout.
>>> >>> >>
>>> >>> >>
>>> >>> >>
>>> >>> >> --
>>> >>> >> Ben Goertzel, PhD
>>> >>> >> http://goertzel.org
>>> >>> >>
>>> >>> >> "The reasonable man adapts himself to the world: the unreasonable
>>> >>> >> one
>>> >>> >> persists in trying to adapt the world to himself. Therefore all
>>> >>> >> progress depends on the unreasonable man." -- George Bernard Shaw
>>> >>> >>
>>> >>> >> --
>>> >>> >> You received this message because you are subscribed to the Google
>>> >>> >> Groups
>>> >>> >> "opencog" group.
>>> >>> >> To unsubscribe from this group and stop receiving emails from it,
>>> >>> >> send
>>> >>> >> an
>>> >>> >> email to [email protected].
>>> >>> >> To post to this group, send email to [email protected].
>>> >>> >> Visit this group at http://groups.google.com/group/opencog.
>>> >>> >> For more options, visit https://groups.google.com/d/optout.
>>> >>> >
>>> >>> >
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> Ben Goertzel, PhD
>>> >>> http://goertzel.org
>>> >>>
>>> >>> "The reasonable man adapts himself to the world: the unreasonable one
>>> >>> persists in trying to adapt the world to himself. Therefore all
>>> >>> progress depends on the unreasonable man." -- George Bernard Shaw
>>> >>>
>>> >>>
>>> >>> -------------------------------------------
>>> >>> AGI
>>> >>> Archives: https://www.listbox.com/member/archive/303/=now
>>> >>> RSS Feed:
>>> >>> https://www.listbox.com/member/archive/rss/303/10872673-8f99760d
>>> >>> Modify Your Subscription:
>>> >>>
>>> >>> https://www.listbox.com/member/?&;
>>> >>> Powered by Listbox: http://www.listbox.com
>>> >>
>>> >>
>>> >
>>> >
>>> > --
>>> > Ben Goertzel, PhD
>>> > http://goertzel.org
>>> >
>>> > "The reasonable man adapts himself to the world: the unreasonable one
>>> > persists in trying to adapt the world to himself. Therefore all progress
>>> > depends on the unreasonable man." -- George Bernard Shaw
>>> >
>>>
>>>
>>>
>>> --
>>> Ben Goertzel, PhD
>>> http://goertzel.org
>>>
>>> "The reasonable man adapts himself to the world: the unreasonable one
>>> persists in trying to adapt the world to himself. Therefore all
>>> progress depends on the unreasonable man." -- George Bernard Shaw
>>
>>
>
>
>
> --
> Ben Goertzel, PhD
> http://goertzel.org
>
> "The reasonable man adapts himself to the world: the unreasonable one
> persists in trying to adapt the world to himself. Therefore all
> progress depends on the unreasonable man." -- George Bernard Shaw



-- 
Ben Goertzel, PhD
http://goertzel.org

"The reasonable man adapts himself to the world: the unreasonable one
persists in trying to adapt the world to himself. Therefore all
progress depends on the unreasonable man." -- George Bernard Shaw


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to