Pei,
I wonder if Cox's Assumption 1 could be salvaged by replacing it
with, say, an assumption that
"Among a set of statements supported by equivalent amounts of
evidence, the relative plausibility of an individual
statement may be assessed by a single real number."
Based on this modified assumption, I think a variation on Cox's
arguments could probably be made to work.
Would this modification address your objection?
-- Ben
On Feb 2, 2007, at 4:13 PM, Pei Wang wrote:
> Ben,
>
> To me, not only Assumption 3 is too strong, but also Assumption 1,
> which does assume that a real number is enough for the
"plausibility
> of a statement". For this reason, these assumptions do not even
"holds
> approximately" in the AGI context --- using one number or two
numbers
> makes a huge difference, which I'm sure you know well.
>
> The Halpern vs. Snow debate is largely irrelevant to this issue. I
> mentioned them just to show that Cox's work is well known to the
UAI
> community.
>
> Pei
>
> On 2/2/07, Ben Goertzel <[EMAIL PROTECTED]> wrote:
>>
>> The paper Pei forwarded claims that Cox's arguments don't work for
>> the discrete case, but the attached paper from Snow in 2002 [which
>> will come through if this listserver allows attachments...]
presents
>> a counterargument, suggesting that a variant of Cox's argument
does
>> in fact work for the discrete case.
>>
>> However, my contention is that Cox's assumptions, while
reasonable,
>> are too strong to be viably assumed for a finite-resources AI
system
>> (or a human brain).
>>
>> To see why, look at Assumption 3 in
>>
>> http://en.wikipedia.org/wiki/Cox's_theorem
>>
>> which states basically that
>>
>> "
>> Suppose [A & B] is equivalent to [C & D]. If we acquire new
>> information A and then acquire further new information B, and
update
>> all probabilities each time, the updated probabilities will be the
>> same as if we had first acquired new information C and then
acquired
>> further new information D.
>> "
>>
>> This is not exactly the case in Novamente, nor in the human brain.
>>
>> So one question is: If this assumption holds only to
approximately in
>> an AI system (or other mind), how inaccurate is the ensuing
>> approximation of probabilistic correctness constituted by its
>> judgments? I.e., how wide are the error bars on the conclusion of
>> Cox's Theorem, when its assumptions are approximately varied?
>>
>> -- Ben
>>
>>
>>
>>
>>
>> On Feb 2, 2007, at 2:39 PM, Pei Wang wrote:
>>
>> >> > I don't know of any work explicitly addressing this sort of
>> >> issue, do
>> >> > you?
>> >>
>> >> No, none that address Cox and AI directly, but I suspect one is
>> >> forthcoming perhaps from you. Yes? :)
>> >
>> > There is a literature on Cox and AI. For example,
>> > http://www.cs.cornell.edu/home/halpern/papers/cox1.pdf
>> >
>> > Pei
>> >
>> > -----
>> > This list is sponsored by AGIRI: http://www.agiri.org/email
>> > To unsubscribe or change your options, please go to:
>> > http://v2.listbox.com/member/?list_id=303
>>
>> -----
>> This list is sponsored by AGIRI: http://www.agiri.org/email
>> To unsubscribe or change your options, please go to:
>> http://v2.listbox.com/member/?list_id=303
>>
>>
>>
>>
>
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?list_id=303
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303