It's not always a problem in principle, but I'd need to think about the
specific case more carefully...




On Mon, Sep 22, 2008 at 4:56 PM, Abram Demski <[EMAIL PROTECTED]> wrote:

> Possibly, but how would you mix infinite-order probabilities with
> regular probabilities?
>
> -Abram
>
> On Mon, Sep 22, 2008 at 1:28 PM, Ben Goertzel <[EMAIL PROTECTED]> wrote:
> >
> > The {A} statements are consistent with NARS, but the existing NARS
> inference
> > rules don't use these statements...
> >
> > A related train of thought has occurred to me...
> >
> > In PLN we explicitly have both intensional and extensional inheritance
> links
> > (though with semantics nonidentical to that used in NARS, and
> fundamentally
> > probabilistic in nature) ... so the "probabilistic quasi-NARS" logic
> you're
> > describing could potentially be used as a sort of "NARS on top of PLN"
> ...
> >
> > I'm not sure how useful such a thing is, but it might be interesting...
> >
> > ben
> >
> >
> >
> > On Mon, Sep 22, 2008 at 12:18 PM, Abram Demski <[EMAIL PROTECTED]>
> > wrote:
> >>
> >> Sure, but it is a consistent extension; {A}-statements have a strongly
> >> NARS-like semantics, so we know they won't just mess everything up.
> >>
> >> On Mon, Sep 22, 2008 at 11:31 AM, Ben Goertzel <[EMAIL PROTECTED]>
> wrote:
> >> >
> >> > Of course ... but then you are not doing NARS inference anymore...
> >> >
> >> > On Mon, Sep 22, 2008 at 8:25 AM, Abram Demski <[EMAIL PROTECTED]>
> >> > wrote:
> >> >>
> >> >> It would be possible to get what you want in the setting, by allowing
> >> >> some probabilistic manipulations not done in NARS. The node
> >> >> probability you want in this case could be simulated by talking about
> >> >> the probability distribution of sentences of the form "X is the
> author
> >> >> of a book". We can give this a low prior probability. Since the
> system
> >> >> manipulates likelihoods, it won't notice; but if we manipulate
> >> >> probabilities, it would.
> >> >>
> >> >> Perhaps a more satisfying answer would be to introduce a new operator
> >> >> into the system, {A}, that simulates the node probability; or more
> >> >> specifically, it represents the average truth-value distribution of
> >> >> statements that have A on one side or the other. So, it has a 'par'
> >> >> value just like inheritance statements do. If there was evidence for
> a
> >> >> low par, there would be an effect in the direction you want. (It
> might
> >> >> be way too small, though?)
> >> >>
> >> >> --Abram
> >> >>
> >> >> On Sun, Sep 21, 2008 at 10:46 PM, Ben Goertzel <[EMAIL PROTECTED]>
> >> >> wrote:
> >> >> >
> >> >> >
> >> >> > On Sun, Sep 21, 2008 at 10:43 PM, Abram Demski
> >> >> > <[EMAIL PROTECTED]>
> >> >> > wrote:
> >> >> >>
> >> >> >> The calculation in which I sum up a bunch of pairs is equivalent
> to
> >> >> >> doing NARS induction + abduction with a final big revision at the
> >> >> >> end
> >> >> >> to combine all the accumulated evidence. But, like I said, I need
> to
> >> >> >> provide a more explicit justification of that calculation...
> >> >> >
> >> >> > As an example inference, consider
> >> >> >
> >> >> > Ben is an author of a book on AGI <tv1>
> >> >> > This dude is an author of a book on AGI <tv2>
> >> >> > |-
> >> >> > This dude is Ben <tv3>
> >> >> >
> >> >> > versus
> >> >> >
> >> >> > Ben is odd <tv1>
> >> >> > This dude is odd <tv2>
> >> >> > |-
> >> >> > This dude is Ben <tv4>
> >> >> >
> >> >> > (Here each of the English statements is a shorthand for a logical
> >> >> > relationship that in the AI systems in question is expressed in a
> >> >> > formal
> >> >> > structure; and the notations like <tv1> indicate uncertain truth
> >> >> > values
> >> >> > attached to logical relationships,  In both NARS and PLN, uncertain
> >> >> > truth
> >> >> > values have multiple components, including a "strength" value that
> >> >> > denotes a
> >> >> > frequency, and other values denoting confidence measures.  However,
> >> >> > the
> >> >> > semantics of the strength values in NARS and PLN are not
> identical.)
> >> >> >
> >> >> > Doing these two inferences in NARS you will get
> >> >> >
> >> >> > tv3.strength = tv4.strength
> >> >> >
> >> >> > whereas in PLN you will not, you will get
> >> >> >
> >> >> > tv3.strength >> tv4.strength
> >> >> >
> >> >> > The difference between the two inference results in the PLN case
> >> >> > results
> >> >> > from the fact that
> >> >> >
> >> >> > P(author of book on AGI) << P(odd)
> >> >> >
> >> >> > and the fact that PLN uses Bayes rule as part of its approach to
> >> >> > these
> >> >> > inferences.
> >> >> >
> >> >> > So, the question is, in your probabilistic variant of NARS, do you
> >> >> > get
> >> >> >
> >> >> > tv3.strength = tv4.strength
> >> >> >
> >> >> > in this case, and if so, why?
> >> >> >
> >> >> > thx
> >> >> > ben
> >> >> > ________________________________
> >> >> > agi | Archives | Modify Your Subscription
> >> >>
> >> >>
> >> >> -------------------------------------------
> >> >> agi
> >> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> >> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> >> >> Modify Your Subscription: https://www.listbox.com/member/?&;
> >> >> Powered by Listbox: http://www.listbox.com
> >> >
> >> >
> >> >
> >> > --
> >> > Ben Goertzel, PhD
> >> > CEO, Novamente LLC and Biomind LLC
> >> > Director of Research, SIAI
> >> > [EMAIL PROTECTED]
> >> >
> >> > "Nothing will ever be attempted if all possible objections must be
> first
> >> > overcome " - Dr Samuel Johnson
> >> >
> >> >
> >> > ________________________________
> >> > agi | Archives | Modify Your Subscription
> >>
> >>
> >> -------------------------------------------
> >> agi
> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> >> Modify Your Subscription: https://www.listbox.com/member/?&;
> >> Powered by Listbox: http://www.listbox.com
> >
> >
> >
> > --
> > Ben Goertzel, PhD
> > CEO, Novamente LLC and Biomind LLC
> > Director of Research, SIAI
> > [EMAIL PROTECTED]
> >
> > "Nothing will ever be attempted if all possible objections must be first
> > overcome " - Dr Samuel Johnson
> >
> >
> > ________________________________
> > agi | Archives | Modify Your Subscription
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

"Nothing will ever be attempted if all possible objections must be first
overcome " - Dr Samuel Johnson



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69
Powered by Listbox: http://www.listbox.com

Reply via email to