for fans of lectures, Janzing and Schoelkopf (who are among the author's of
the paper Ben just linked and are among the most established people in the
field afaik):
https://www.youtube.com/watch?v=KsbftkwZTq4

On Tue, Jan 13, 2015 at 5:53 PM, Ben Goertzel via AGI <[email protected]>
wrote:

> See Figure 1 in
>
> http://arxiv.org/abs/1407.2256
>
> Distinguishing between cases a) and b) is useful in narrowing down
> causality
>
> But yeah, most interesting causal inference involves cases where the
> agent can't actually carry out the exact manipulations in question.
> So then IMO causal inference involves analogical inference, whether
> the source of the analogy is some related case where the agent could
> carry out a direct manipulation...
>
> ben g
>
> On Tue, Jan 13, 2015 at 11:48 AM, Aaron Hosford <[email protected]>
> wrote:
> >> Are you familiar with "causal networks" as developed by Judea Pearl,
> >> etc.?
> >
> >
> > I have heard the name thrown around but that's the extent of it. I'll
> have
> > to read up on this.
> >
> >> But once the causal network theory has done its
> >> job, one is left with a set of events that are *potentially*
> >> consistently labeled causal -- one still needs some other kind of
> >> insight to tell what should really be considered causal or not
> >
> >
> > That's a bit vague. Can you put your finger on precisely which types of
> > event pairs would mistakenly be considered potentially causal by this
> > approach? Based on what you have said so far, I would think it would
> just be
> > a matter of formulating an experiment. The agent could either test the
> > ability to control one variable (the effect) with another (the cause) by
> > controlling the other, or if the cause itself could not be controlled by
> the
> > agent then whether it is "truly" a cause is a moot point because the
> > information is inaccessible and consequentially irrelevant to the agent's
> > internal model.
> >
> > On Tue, Jan 13, 2015 at 11:34 AM, Ben Goertzel <[email protected]> wrote:
> >>
> >> Hi Aaron,
> >>
> >> Are you familiar with "causal networks" as developed by Judea Pearl,
> >> etc.?  The basic theory thereof involves precisely this type of
> >> manipulation that you're talking about....
> >>
> >> However, I've never been fully satisfied with causal network theory.
> >> Ultimately my conclusion is that, given a network of probabilistically
> >> related events, causal network theory helps rule out which
> >> relationships are almost surely NOT causal but rather consequential
> >> from other events.  But once the causal network theory has done its
> >> job, one is left with a set of events that are *potentially*
> >> consistently labeled causal -- one still needs some other kind of
> >> insight to tell what should really be considered causal or not
> >>
> >>
> >> -- Ben
> >>
> >> On Tue, Jan 13, 2015 at 11:23 AM, Aaron Hosford <[email protected]>
> >> wrote:
> >> > Ben,
> >> >
> >> > I'm not sure this would be useful within the scope of what you're
> >> > working
> >> > on, but I don't think probabilities are the right basis for
> >> > understanding
> >> > causation. Causation, to me, requires interaction between objects or
> >> > events
> >> > that results in a transfer of information/control. Think of a
> >> > simulation. If
> >> > changing the simulation in some way (i.e. modifying the state of an
> >> > object
> >> > or introducing/removing an object or event) changes whether a
> predicate
> >> > of
> >> > the simulation's (future) history is satisfied, then that change can
> be
> >> > said
> >> > to be the cause of the predicate's (or its negation's) satisfaction.
> In
> >> > other words, you can control one locus of the model indirectly by
> >> > modifying
> >> > some other locus of it. (It makes sense that human beings would be
> >> > preoccupied with flow of control, given our comparative expertise at
> >> > controlling our environments relative to other species.) This notion
> of
> >> > causality is a purely deterministic one, which is in agreement with
> most
> >> > people's intuitive understanding of causality. Probabilities are
> >> > introduced
> >> > only when we take into account potential uncertainties, inaccuracies,
> or
> >> > imprecisions that may be present in the model used for simulation;
> they
> >> > are
> >> > averages or densities of deterministic behaviors over the space of
> >> > possibilities. If you start from this deterministic perspective and
> then
> >> > ask
> >> > yourself how you can derive probabilistic models of such a
> deterministic
> >> > phenomenon under uncertainty, I think the relationships between
> >> > causality,
> >> > probability, correlation, and flow of time will become fairly evident,
> >> > and
> >> > it will be much easier to put together a system that reconstructs
> >> > underlying
> >> > causal relationships from sample data.
> >> >
> >> > On Tue, Nov 25, 2014 at 5:53 AM, Ben Goertzel via AGI <
> [email protected]>
> >> > wrote:
> >> >>
> >> >> Hmmm...
> >> >>
> >> >> Having thought about this more, while I was indeed traveling
> backwards
> >> >> in time when I wrote the previous email, it's not too relevant anyhow
> >> >> because the Second Law only holds globally, and in complex systems
> >> >> there are many subsystems that are behaving anti-entropically.  So
> I'm
> >> >> no sure one can use the law of entropy increase to draw conclusions
> >> >> about local causality.
> >> >>
> >> >> However, I was thinking about section 6.3.2 of
> >> >>
> >> >> http://cqi.inf.usi.ch/qic/94_Lloyd.pdf
> >> >>
> >> >> where Seth Lloyd observes that
> >> >>
> >> >> "Having a common effect does not induce correlation between events,
> >> >> while having a common cause does."
> >> >>
> >> >> I.e.
> >> >>
> >> >> -- In the case of two causes with a common effect ... there is an
> >> >> increase of information from past to future (the probability spread
> >> >> across two causes is now concentrated on a single effect).   There no
> >> >> correlation in the past (between the causes).   This is the opposite
> >> >> direction of the Second Law of Thermodynamics.
> >> >>
> >> >> -- In the case of two effects with a common cause ...  there is a
> >> >> decrease of information from past to future (the probability
> >> >> concentrated in one cause is now spread across two effects).   There
> >> >> is correlation in the future (between the effects).  This is in the
> >> >> direction of the Second Law of Thermodynamics.
> >> >>
> >> >> ...
> >> >>
> >> >> I.e. in many cases the direction of causal influence may be
> >> >> identifiable as the direction of increasing correlation....   I'm not
> >> >> sure exactly what are the limits of this conclusion though.
> >> >>
> >> >> ...
> >> >>
> >> >> Soo --   What if one has two sets of variables, S and T, and there is
> >> >> significant mutual information between the values of S and the values
> >> >> of T, as evaluated across different cases...?   So, suppose we have
> >> >> both
> >> >>
> >> >> S --> T
> >> >>
> >> >> and
> >> >>
> >> >> T --> S
> >> >>
> >> >> in a sense....    But, if there is significantly more correlation
> >> >> among the variables within T, than among the variables within S, then
> >> >> we can say that it's more likely that T is the effect and S is the
> >> >> cause...
> >> >>
> >> >> The asymmetry used to identify causation is then one of correlation
> >> >> rather than of temporality directly...
> >> >>
> >> >> This may be a way of heuristically inferring causality from
> >> >> non-temporal data, if one has a sufficient ensemble of data
> samples...
> >> >>
> >> >> -- Ben
> >> >>
> >> >>
> >> >> On Tue, Nov 25, 2014 at 1:46 PM, Ben Goertzel <[email protected]>
> wrote:
> >> >> >
> >> >> > Hmm, maybe you're right , maybe I was traveling backwards in time
> >> >> > when I
> >> >> > wrote that ...
> >> >> >
> >> >> > (More later)
> >> >> >
> >> >> > On Tuesday, November 25, 2014, martin biehl <[email protected]>
> wrote:
> >> >> >>
> >> >> >> hm, sounds interesting, but I don't get it either. If entropy
> >> >> >> increases,
> >> >> >> the uncertainty of the state increases and information (about the
> >> >> >> state)
> >> >> >> decreases as you say, but why would the past then contain more
> >> >> >> information
> >> >> >> about the future than vice versa? Let X be the past, Y be the
> >> >> >> future,
> >> >> >> then
> >> >> >> as mutual information is symmetric:
> >> >> >> H(X) - H(X|Y) = H(Y) - H(Y|X)
> >> >> >> now H(Y) > H(X) because of entropy increase.
> >> >> >> then
> >> >> >> H(Y|X) > H(X|Y)
> >> >> >> and the future should be more uncertain given the past than vice
> >> >> >> versa.
> >> >> >> Where did this go wrong?
> >> >> >>
> >> >> >>
> >> >> >> On Tue, Nov 25, 2014 at 2:13 AM, Ben Goertzel via AGI
> >> >> >> <[email protected]>
> >> >> >> wrote:
> >> >> >>>
> >> >> >>> Information is negentropy, so increase of entropy implies
> decrease
> >> >> >>> of
> >> >> >>> information...
> >> >> >>>
> >> >> >>> Acquiring information about a system is associated with entropy
> >> >> >>> production...
> >> >> >>>
> >> >> >>> On Tue, Nov 25, 2014 at 9:59 AM, Aaron Nitzkin <
> [email protected]>
> >> >> >>> wrote:
> >> >> >>> > Sorry, I must be a little confused -- probably thinking from
> the
> >> >> >>> > wrong
> >> >> >>> > perspective . . . I would think that there is more information
> >> >> >>> > in the future about the past than vice versa, because we know
> >> >> >>> > more
> >> >> >>> > about the
> >> >> >>> > past than we do about the future, and also, doesn't
> >> >> >>> > increase in entropy imply increase in information (because it
> >> >> >>> > requries
> >> >> >>> > more
> >> >> >>> > information to specify the configuration of a system
> >> >> >>> > with higher entropy than the same system with lower entropy?)
> >> >> >>> >
> >> >> >>> > On Tue, Nov 25, 2014 at 8:27 AM, Ben Goertzel <
> [email protected]>
> >> >> >>> > wrote:
> >> >> >>> >>
> >> >> >>> >> In the early part of the paper, the author clarifies that
> while
> >> >> >>> >> he
> >> >> >>> >> assumes "temporal precedence as an aspect of causality" for
> >> >> >>> >> simplicity, actually his approach would work with any other
> >> >> >>> >> systematic
> >> >> >>> >> way of assigning asymmetric directions to relationships
> between
> >> >> >>> >> events
> >> >> >>> >>
> >> >> >>> >> I have been thinking a lot about how to infer causality from
> >> >> >>> >> non-time-series data (e.g. categorial gene expression data),
> and
> >> >> >>> >> this
> >> >> >>> >> is a case where looking at some other sort of asymmetry than
> >> >> >>> >> temporal
> >> >> >>> >> precedence (but that may generally correlated with temporal
> >> >> >>> >> precedence) seems to make sense.   E.g. I've been thinking
> about
> >> >> >>> >> looking at informational asymmetry: If one has P(A = a | B=b),
> >> >> >>> >> one
> >> >> >>> >> can
> >> >> >>> >> look at whether the distribution for A gives more information
> >> >> >>> >> about
> >> >> >>> >> the distribution for B, or vice versa.   This informational
> >> >> >>> >> asymmetry
> >> >> >>> >> can be used similarly to temporal asymmetry in defining
> >> >> >>> >> causality.
> >> >> >>> >> Furthermore, it on the average is going to correlate with
> >> >> >>> >> temporal
> >> >> >>> >> asymmetry, because the past tends to contain more information
> >> >> >>> >> about
> >> >> >>> >> the future than vice versa (due to entropy increase, roughly
> >> >> >>> >> speaking... but there's more story here...)
> >> >> >>> >>
> >> >> >>> >> -- Ben
> >> >> >>> >>
> >> >> >>> >>
> >> >> >>> >> On Tue, Nov 25, 2014 at 5:34 AM, Michael van der Gulik
> >> >> >>> >> <[email protected]> wrote:
> >> >> >>> >> > "Chapter 1. Quantum mechanics... "
> >> >> >>> >> >
> >> >> >>> >> > It's a nice article; I'll add it to my reading list.
> >> >> >>> >> > Prediction
> >> >> >>> >> > involves
> >> >> >>> >> > working out what causes what, so it's pretty fundamental.
> >> >> >>> >> >
> >> >> >>> >> > I have a question. Causation in my mind seems to always
> >> >> >>> >> > involve
> >> >> >>> >> > time,
> >> >> >>> >> > and I
> >> >> >>> >> > suspect it's impossible to have causation without including
> >> >> >>> >> > timing.
> >> >> >>> >> > So...
> >> >> >>> >> >
> >> >> >>> >> > Is it possible for a cause to happen at exactly the same
> >> >> >>> >> > moment
> >> >> >>> >> > as
> >> >> >>> >> > its
> >> >> >>> >> > effect?
> >> >> >>> >> >
> >> >> >>> >> > Is it possible for a cause to happen after its effect?
> >> >> >>> >> >
> >> >> >>> >> > One instance I'm trying to get my head around is when an
> >> >> >>> >> > intelligence
> >> >> >>> >> > anticipates a cause (which is an event in the future), which
> >> >> >>> >> > results
> >> >> >>> >> > in
> >> >> >>> >> > the
> >> >> >>> >> > intelligence acting such that the effect occurs before the
> >> >> >>> >> > cause.
> >> >> >>> >> > Perhaps
> >> >> >>> >> > the anticipation itself is the causal event.
> >> >> >>> >> >
> >> >> >>> >> > Regards,
> >> >> >>> >> > Michael.
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> > On Sun, Nov 23, 2014 at 7:17 AM, Ben Goertzel
> >> >> >>> >> > <[email protected]>
> >> >> >>> >> > wrote:
> >> >> >>> >> >>
> >> >> >>> >> >> I just happened across this 2011 paper on the probabilistic
> >> >> >>> >> >> foundation
> >> >> >>> >> >> of causality,
> >> >> >>> >> >>
> >> >> >>> >> >>
> http://philsci-archive.pitt.edu/9729/1/Website_Version_2.pdf
> >> >> >>> >> >>
> >> >> >>> >> >> which seems to carefully clarify a bunch of issues that
> >> >> >>> >> >> remain
> >> >> >>> >> >> dangling in prior discussions of the topic
> >> >> >>> >> >>
> >> >> >>> >> >> It seems to give a good characterization of what it means
> for
> >> >> >>> >> >> "P
> >> >> >>> >> >> to
> >> >> >>> >> >> appear to cause Q, based on the knowledge-base of observer
> O"
> >> >> >>> >> >>
> >> >> >>> >> >> --
> >> >> >>> >> >> Ben Goertzel, PhD
> >> >> >>> >> >> http://goertzel.org
> >> >> >>> >> >>
> >> >> >>> >> >> "The reasonable man adapts himself to the world: the
> >> >> >>> >> >> unreasonable
> >> >> >>> >> >> one
> >> >> >>> >> >> persists in trying to adapt the world to himself. Therefore
> >> >> >>> >> >> all
> >> >> >>> >> >> progress depends on the unreasonable man." -- George
> Bernard
> >> >> >>> >> >> Shaw
> >> >> >>> >> >>
> >> >> >>> >> >> --
> >> >> >>> >> >> You received this message because you are subscribed to the
> >> >> >>> >> >> Google
> >> >> >>> >> >> Groups
> >> >> >>> >> >> "Artificial General Intelligence" group.
> >> >> >>> >> >> To unsubscribe from this group and stop receiving emails
> from
> >> >> >>> >> >> it,
> >> >> >>> >> >> send
> >> >> >>> >> >> an
> >> >> >>> >> >> email to
> >> >> >>> >> >>
> [email protected].
> >> >> >>> >> >> For more options, visit https://groups.google.com/d/optout
> .
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> > --
> >> >> >>> >> > http://gulik.pbwiki.com/
> >> >> >>> >> >
> >> >> >>> >> > --
> >> >> >>> >> > You received this message because you are subscribed to the
> >> >> >>> >> > Google
> >> >> >>> >> > Groups
> >> >> >>> >> > "Artificial General Intelligence" group.
> >> >> >>> >> > To unsubscribe from this group and stop receiving emails
> from
> >> >> >>> >> > it,
> >> >> >>> >> > send
> >> >> >>> >> > an
> >> >> >>> >> > email to
> >> >> >>> >> >
> [email protected].
> >> >> >>> >> > For more options, visit https://groups.google.com/d/optout.
> >> >> >>> >>
> >> >> >>> >>
> >> >> >>> >>
> >> >> >>> >> --
> >> >> >>> >> Ben Goertzel, PhD
> >> >> >>> >> http://goertzel.org
> >> >> >>> >>
> >> >> >>> >> "The reasonable man adapts himself to the world: the
> >> >> >>> >> unreasonable
> >> >> >>> >> one
> >> >> >>> >> persists in trying to adapt the world to himself. Therefore
> all
> >> >> >>> >> progress depends on the unreasonable man." -- George Bernard
> >> >> >>> >> Shaw
> >> >> >>> >>
> >> >> >>> >> --
> >> >> >>> >> You received this message because you are subscribed to the
> >> >> >>> >> Google
> >> >> >>> >> Groups
> >> >> >>> >> "opencog" group.
> >> >> >>> >> To unsubscribe from this group and stop receiving emails from
> >> >> >>> >> it,
> >> >> >>> >> send
> >> >> >>> >> an
> >> >> >>> >> email to [email protected].
> >> >> >>> >> To post to this group, send email to [email protected]
> .
> >> >> >>> >> Visit this group at http://groups.google.com/group/opencog.
> >> >> >>> >> For more options, visit https://groups.google.com/d/optout.
> >> >> >>> >
> >> >> >>> >
> >> >> >>>
> >> >> >>>
> >> >> >>>
> >> >> >>> --
> >> >> >>> Ben Goertzel, PhD
> >> >> >>> http://goertzel.org
> >> >> >>>
> >> >> >>> "The reasonable man adapts himself to the world: the unreasonable
> >> >> >>> one
> >> >> >>> persists in trying to adapt the world to himself. Therefore all
> >> >> >>> progress depends on the unreasonable man." -- George Bernard Shaw
> >> >> >>>
> >> >> >>>
> >> >> >>> -------------------------------------------
> >> >> >>> AGI
> >> >> >>> Archives: https://www.listbox.com/member/archive/303/=now
> >> >> >>> RSS Feed:
> >> >> >>> https://www.listbox.com/member/archive/rss/303/10872673-8f99760d
> >> >> >>> Modify Your Subscription:
> >> >> >>> https://www.listbox.com/member/?&;
> >> >> >>> Powered by Listbox: http://www.listbox.com
> >> >> >>
> >> >> >>
> >> >> >
> >> >> >
> >> >> > --
> >> >> > Ben Goertzel, PhD
> >> >> > http://goertzel.org
> >> >> >
> >> >> > "The reasonable man adapts himself to the world: the unreasonable
> one
> >> >> > persists in trying to adapt the world to himself. Therefore all
> >> >> > progress
> >> >> > depends on the unreasonable man." -- George Bernard Shaw
> >> >> >
> >> >>
> >> >>
> >> >>
> >> >> --
> >> >> Ben Goertzel, PhD
> >> >> http://goertzel.org
> >> >>
> >> >> "The reasonable man adapts himself to the world: the unreasonable one
> >> >> persists in trying to adapt the world to himself. Therefore all
> >> >> progress depends on the unreasonable man." -- George Bernard Shaw
> >> >>
> >> >>
> >> >> -------------------------------------------
> >> >> AGI
> >> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> >> RSS Feed:
> >> >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> >> >> Modify Your Subscription:
> >> >>
> >> >> https://www.listbox.com/member/?&;
> >> >> Powered by Listbox: http://www.listbox.com
> >> >
> >> >
> >>
> >>
> >>
> >> --
> >> Ben Goertzel, PhD
> >> http://goertzel.org
> >>
> >> "The reasonable man adapts himself to the world: the unreasonable one
> >> persists in trying to adapt the world to himself. Therefore all
> >> progress depends on the unreasonable man." -- George Bernard Shaw
> >
> >
>
>
>
> --
> Ben Goertzel, PhD
> http://goertzel.org
>
> "The reasonable man adapts himself to the world: the unreasonable one
> persists in trying to adapt the world to himself. Therefore all
> progress depends on the unreasonable man." -- George Bernard Shaw
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/10872673-8f99760d
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to