On Jan 28, 2008 7:56 AM, Mike Tintner <[EMAIL PROTECTED]> wrote:

> X:Of course this is a variation on "the grounding problem" in AI.  But
> do you think some sort of **absolute** grounding is relevant to
> effective interaction between individual agents (assuming you think
> any such ultimate grounding could even perform a function within a
> limited system), or might it be that systems interact effectively to
> the extent their dynamics are based on **relevant** models, regardless
> of even proximate grounding in any functional sense?
>
> Er.. my body couldn't make any sense of this :). Could you be clearer giving
> examples of the agents/systems  and what you mean by absolute/ proximate
> grounding?

I see that you're talking about interaction between systems considered
to be "minds", and highlighting the question of what is necessary to
form a shared basis for **relevant** interaction.  I agree that a
"mind" without an environment of interaction is meaningless, in the
same way that any statement (or pattern of bits) without context is
meaningless.  However, I would argue that just as context is never
absolute, nor is there ever any need for it to be absolute, indeed for
practical (functional) reasons it can never be absolute, embodiment
need not be absolute, complete, or ultimately grounded.

I use the term "system" to refer as clearly as possible to any
distinct configuration of inter-related objects, with the implication
that the system must be physically realizable, therefore it models
neither infinities or infinitesimals, nor could it model a Cartesian
singularity of Self.

I use the term "agent" to refer as clearly as possible to a system
exhibiting "agency", i.e. behavior recognized as intentional, i.e.
operating on behalf of an entity.  It may be useful here to point out
that recognition of agency inheres in the observer (including the case
of the observer being the agent-system itself), rather than agency
being somehow an objectively measurable property of the system itself.
 Further, the "entity" which is the principal behind any agency is
entirely abstract (independent of any physical instantiation.)
[Understanding this is key to various paradoxes of personal identity.]

I distinguish between "absolute" and "proximate" grounding in regard
to the functional (and information-theoretic) impossibility of a
system modeling it's entire chain of connections to "ultimate
reality", while in actuality any system interacts only with its
proximate environment, just as to "know" an object is not to know what
it "is" but to know its interface.  To presume to know more would be
to presume some privileged mode of knowledge.

So in short, I agree with you that "embodiment" is essential to
meaningful interaction, thus for there to be agency, thus for there to
be a "Self" for the mind to know.  But I extend this and emphasize
that it's not necessary that such "embodiment" be physical, nor that
it be logically grounded in "ultimate reality", but rather, that
interaction is relevant and meaningful to the extent that some
(necessarily partial and arbitrarily distant from "reality") context
is shared.

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=4007604&id_secret=90632239-135dac

Reply via email to