I.e., what would be a case where: -- X, Y and Z have positive multivariate mutual information
-- Y and Z are common causes of X -- Y and Z are *not* correlated, i.e. they don't have positive binary mutual information ? On Wed, Nov 26, 2014 at 12:46 AM, Ben Goertzel <[email protected]> wrote: > Martin, > > Hmmm -- This is "explaining away" as discussed in the Bayes Nets > literature, right? > > https://en.wikipedia.org/wiki/Multivariate_mutual_information#Example_of_Negative_Multivariate_mutual_information > > The examples on that wiki page are ones where common cause corresponds > to positive MMI, and common effect corresponds to negative MMI. I > wonder how generally that holds? > > ben > > On Wed, Nov 26, 2014 at 12:35 AM, martin biehl <[email protected]> wrote: >> "Having a common effect does not induce correlation between events, >> while having a common cause does." >> >> is possibly not always true, take the infamous XOR gate, here if the output >> (effect) is known to be 1 then this implies a correlation between the two >> inputs i.e. they must be opposites. >> >> is this a counterexample? >> >> On Tue, Nov 25, 2014 at 11:53 AM, Ben Goertzel <[email protected]> wrote: >>> >>> Hmmm... >>> >>> Having thought about this more, while I was indeed traveling backwards >>> in time when I wrote the previous email, it's not too relevant anyhow >>> because the Second Law only holds globally, and in complex systems >>> there are many subsystems that are behaving anti-entropically. So I'm >>> no sure one can use the law of entropy increase to draw conclusions >>> about local causality. >>> >>> However, I was thinking about section 6.3.2 of >>> >>> http://cqi.inf.usi.ch/qic/94_Lloyd.pdf >>> >>> where Seth Lloyd observes that >>> >>> "Having a common effect does not induce correlation between events, >>> while having a common cause does." >>> >>> I.e. >>> >>> -- In the case of two causes with a common effect ... there is an >>> increase of information from past to future (the probability spread >>> across two causes is now concentrated on a single effect). There no >>> correlation in the past (between the causes). This is the opposite >>> direction of the Second Law of Thermodynamics. >>> >>> -- In the case of two effects with a common cause ... there is a >>> decrease of information from past to future (the probability >>> concentrated in one cause is now spread across two effects). There >>> is correlation in the future (between the effects). This is in the >>> direction of the Second Law of Thermodynamics. >>> >>> ... >>> >>> I.e. in many cases the direction of causal influence may be >>> identifiable as the direction of increasing correlation.... I'm not >>> sure exactly what are the limits of this conclusion though. >>> >>> ... >>> >>> Soo -- What if one has two sets of variables, S and T, and there is >>> significant mutual information between the values of S and the values >>> of T, as evaluated across different cases...? So, suppose we have >>> both >>> >>> S --> T >>> >>> and >>> >>> T --> S >>> >>> in a sense.... But, if there is significantly more correlation >>> among the variables within T, than among the variables within S, then >>> we can say that it's more likely that T is the effect and S is the >>> cause... >>> >>> The asymmetry used to identify causation is then one of correlation >>> rather than of temporality directly... >>> >>> This may be a way of heuristically inferring causality from >>> non-temporal data, if one has a sufficient ensemble of data samples... >>> >>> -- Ben >>> >>> >>> On Tue, Nov 25, 2014 at 1:46 PM, Ben Goertzel <[email protected]> wrote: >>> > >>> > Hmm, maybe you're right , maybe I was traveling backwards in time when I >>> > wrote that ... >>> > >>> > (More later) >>> > >>> > On Tuesday, November 25, 2014, martin biehl <[email protected]> wrote: >>> >> >>> >> hm, sounds interesting, but I don't get it either. If entropy >>> >> increases, >>> >> the uncertainty of the state increases and information (about the >>> >> state) >>> >> decreases as you say, but why would the past then contain more >>> >> information >>> >> about the future than vice versa? Let X be the past, Y be the future, >>> >> then >>> >> as mutual information is symmetric: >>> >> H(X) - H(X|Y) = H(Y) - H(Y|X) >>> >> now H(Y) > H(X) because of entropy increase. >>> >> then >>> >> H(Y|X) > H(X|Y) >>> >> and the future should be more uncertain given the past than vice versa. >>> >> Where did this go wrong? >>> >> >>> >> >>> >> On Tue, Nov 25, 2014 at 2:13 AM, Ben Goertzel via AGI <[email protected]> >>> >> wrote: >>> >>> >>> >>> Information is negentropy, so increase of entropy implies decrease of >>> >>> information... >>> >>> >>> >>> Acquiring information about a system is associated with entropy >>> >>> production... >>> >>> >>> >>> On Tue, Nov 25, 2014 at 9:59 AM, Aaron Nitzkin <[email protected]> >>> >>> wrote: >>> >>> > Sorry, I must be a little confused -- probably thinking from the >>> >>> > wrong >>> >>> > perspective . . . I would think that there is more information >>> >>> > in the future about the past than vice versa, because we know more >>> >>> > about the >>> >>> > past than we do about the future, and also, doesn't >>> >>> > increase in entropy imply increase in information (because it >>> >>> > requries >>> >>> > more >>> >>> > information to specify the configuration of a system >>> >>> > with higher entropy than the same system with lower entropy?) >>> >>> > >>> >>> > On Tue, Nov 25, 2014 at 8:27 AM, Ben Goertzel <[email protected]> >>> >>> > wrote: >>> >>> >> >>> >>> >> In the early part of the paper, the author clarifies that while he >>> >>> >> assumes "temporal precedence as an aspect of causality" for >>> >>> >> simplicity, actually his approach would work with any other >>> >>> >> systematic >>> >>> >> way of assigning asymmetric directions to relationships between >>> >>> >> events >>> >>> >> >>> >>> >> I have been thinking a lot about how to infer causality from >>> >>> >> non-time-series data (e.g. categorial gene expression data), and >>> >>> >> this >>> >>> >> is a case where looking at some other sort of asymmetry than >>> >>> >> temporal >>> >>> >> precedence (but that may generally correlated with temporal >>> >>> >> precedence) seems to make sense. E.g. I've been thinking about >>> >>> >> looking at informational asymmetry: If one has P(A = a | B=b), one >>> >>> >> can >>> >>> >> look at whether the distribution for A gives more information about >>> >>> >> the distribution for B, or vice versa. This informational >>> >>> >> asymmetry >>> >>> >> can be used similarly to temporal asymmetry in defining causality. >>> >>> >> Furthermore, it on the average is going to correlate with temporal >>> >>> >> asymmetry, because the past tends to contain more information about >>> >>> >> the future than vice versa (due to entropy increase, roughly >>> >>> >> speaking... but there's more story here...) >>> >>> >> >>> >>> >> -- Ben >>> >>> >> >>> >>> >> >>> >>> >> On Tue, Nov 25, 2014 at 5:34 AM, Michael van der Gulik >>> >>> >> <[email protected]> wrote: >>> >>> >> > "Chapter 1. Quantum mechanics... " >>> >>> >> > >>> >>> >> > It's a nice article; I'll add it to my reading list. Prediction >>> >>> >> > involves >>> >>> >> > working out what causes what, so it's pretty fundamental. >>> >>> >> > >>> >>> >> > I have a question. Causation in my mind seems to always involve >>> >>> >> > time, >>> >>> >> > and I >>> >>> >> > suspect it's impossible to have causation without including >>> >>> >> > timing. >>> >>> >> > So... >>> >>> >> > >>> >>> >> > Is it possible for a cause to happen at exactly the same moment >>> >>> >> > as >>> >>> >> > its >>> >>> >> > effect? >>> >>> >> > >>> >>> >> > Is it possible for a cause to happen after its effect? >>> >>> >> > >>> >>> >> > One instance I'm trying to get my head around is when an >>> >>> >> > intelligence >>> >>> >> > anticipates a cause (which is an event in the future), which >>> >>> >> > results >>> >>> >> > in >>> >>> >> > the >>> >>> >> > intelligence acting such that the effect occurs before the cause. >>> >>> >> > Perhaps >>> >>> >> > the anticipation itself is the causal event. >>> >>> >> > >>> >>> >> > Regards, >>> >>> >> > Michael. >>> >>> >> > >>> >>> >> > >>> >>> >> > On Sun, Nov 23, 2014 at 7:17 AM, Ben Goertzel <[email protected]> >>> >>> >> > wrote: >>> >>> >> >> >>> >>> >> >> I just happened across this 2011 paper on the probabilistic >>> >>> >> >> foundation >>> >>> >> >> of causality, >>> >>> >> >> >>> >>> >> >> http://philsci-archive.pitt.edu/9729/1/Website_Version_2.pdf >>> >>> >> >> >>> >>> >> >> which seems to carefully clarify a bunch of issues that remain >>> >>> >> >> dangling in prior discussions of the topic >>> >>> >> >> >>> >>> >> >> It seems to give a good characterization of what it means for "P >>> >>> >> >> to >>> >>> >> >> appear to cause Q, based on the knowledge-base of observer O" >>> >>> >> >> >>> >>> >> >> -- >>> >>> >> >> Ben Goertzel, PhD >>> >>> >> >> http://goertzel.org >>> >>> >> >> >>> >>> >> >> "The reasonable man adapts himself to the world: the >>> >>> >> >> unreasonable >>> >>> >> >> one >>> >>> >> >> persists in trying to adapt the world to himself. Therefore all >>> >>> >> >> progress depends on the unreasonable man." -- George Bernard >>> >>> >> >> Shaw >>> >>> >> >> >>> >>> >> >> -- >>> >>> >> >> You received this message because you are subscribed to the >>> >>> >> >> Google >>> >>> >> >> Groups >>> >>> >> >> "Artificial General Intelligence" group. >>> >>> >> >> To unsubscribe from this group and stop receiving emails from >>> >>> >> >> it, >>> >>> >> >> send >>> >>> >> >> an >>> >>> >> >> email to >>> >>> >> >> [email protected]. >>> >>> >> >> For more options, visit https://groups.google.com/d/optout. >>> >>> >> > >>> >>> >> > >>> >>> >> > >>> >>> >> > >>> >>> >> > -- >>> >>> >> > http://gulik.pbwiki.com/ >>> >>> >> > >>> >>> >> > -- >>> >>> >> > You received this message because you are subscribed to the >>> >>> >> > Google >>> >>> >> > Groups >>> >>> >> > "Artificial General Intelligence" group. >>> >>> >> > To unsubscribe from this group and stop receiving emails from it, >>> >>> >> > send >>> >>> >> > an >>> >>> >> > email to >>> >>> >> > [email protected]. >>> >>> >> > For more options, visit https://groups.google.com/d/optout. >>> >>> >> >>> >>> >> >>> >>> >> >>> >>> >> -- >>> >>> >> Ben Goertzel, PhD >>> >>> >> http://goertzel.org >>> >>> >> >>> >>> >> "The reasonable man adapts himself to the world: the unreasonable >>> >>> >> one >>> >>> >> persists in trying to adapt the world to himself. Therefore all >>> >>> >> progress depends on the unreasonable man." -- George Bernard Shaw >>> >>> >> >>> >>> >> -- >>> >>> >> You received this message because you are subscribed to the Google >>> >>> >> Groups >>> >>> >> "opencog" group. >>> >>> >> To unsubscribe from this group and stop receiving emails from it, >>> >>> >> send >>> >>> >> an >>> >>> >> email to [email protected]. >>> >>> >> To post to this group, send email to [email protected]. >>> >>> >> Visit this group at http://groups.google.com/group/opencog. >>> >>> >> For more options, visit https://groups.google.com/d/optout. >>> >>> > >>> >>> > >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> Ben Goertzel, PhD >>> >>> http://goertzel.org >>> >>> >>> >>> "The reasonable man adapts himself to the world: the unreasonable one >>> >>> persists in trying to adapt the world to himself. Therefore all >>> >>> progress depends on the unreasonable man." -- George Bernard Shaw >>> >>> >>> >>> >>> >>> ------------------------------------------- >>> >>> AGI >>> >>> Archives: https://www.listbox.com/member/archive/303/=now >>> >>> RSS Feed: >>> >>> https://www.listbox.com/member/archive/rss/303/10872673-8f99760d >>> >>> Modify Your Subscription: >>> >>> >>> >>> https://www.listbox.com/member/?& >>> >>> Powered by Listbox: http://www.listbox.com >>> >> >>> >> >>> > >>> > >>> > -- >>> > Ben Goertzel, PhD >>> > http://goertzel.org >>> > >>> > "The reasonable man adapts himself to the world: the unreasonable one >>> > persists in trying to adapt the world to himself. Therefore all progress >>> > depends on the unreasonable man." -- George Bernard Shaw >>> > >>> >>> >>> >>> -- >>> Ben Goertzel, PhD >>> http://goertzel.org >>> >>> "The reasonable man adapts himself to the world: the unreasonable one >>> persists in trying to adapt the world to himself. Therefore all >>> progress depends on the unreasonable man." -- George Bernard Shaw >> >> > > > > -- > Ben Goertzel, PhD > http://goertzel.org > > "The reasonable man adapts himself to the world: the unreasonable one > persists in trying to adapt the world to himself. Therefore all > progress depends on the unreasonable man." -- George Bernard Shaw -- Ben Goertzel, PhD http://goertzel.org "The reasonable man adapts himself to the world: the unreasonable one persists in trying to adapt the world to himself. Therefore all progress depends on the unreasonable man." -- George Bernard Shaw ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
