But I do think the nature of conscious qualia, as an abstract system, is
interesting and non-trivial. Each person is their own universe - there is
something more to feelings than just a neuron lighting up, they are part of
an integrated dynamic.


2008/11/2 Brent Meeker <[EMAIL PROTECTED]>

>
> Michael Rosefield wrote:
> > I think there's so many different questions involved in this topic
> > it's going to be hard to sort them out. There's 'what produces our
> > sense of self', 'how can continuity of identity be quantified', 'at
> > what point do differentiated substrates produce different
> > consciousnesses', 'can the nature of consciousness be captured through
> > snapshots of mental activity, or only through a dynamic interpretation
> > taken over a period of time?'... and it's far too late for me to
> > attempt to unravel all that!
> >
> > My feeling, though, is that once you've managed to assign some
> > informational entity as being a conscious mind, then you could track
> > it through time.
> But notice that everything you say about paths and variables and
> measure, apply to any system.  Saying it is a conscious  process doesn't
> change anything.
>
> My guess is that eventually we'll be able to create AI/robots that seem
> as intelligent and conscious as, for example, dogs seem. We'll also be
> able to partially map brains so that we can say that when these neurons
> do this the person is thinking thus and so. Once we have this degree of
> understanding and control, questions about "consciousness" will no
> longer seem relevant.  They'll be like the questions that philosophers
> asked about life before we understood the molecular functions of living
> systems.  They would ask:Where is the life?  Is a virus alive?  How does
> life get passed from parent to child?   The questions won't get
> answered; they'll just be seen as the wrong questions.
>
> Brent
> "One cannot guess the real difficulties of a problem before having
> solved it."
>   --- Carl Ludwig Siegel
>
> > If you tweaked some physical variables, then much like a monte carlo
> > simulation you could see potential paths it could follow. Given enough
> > variables and tweaking, you might be able to fully populate the
> > state-space according to what question we're asking, and it would seem
> > to me to be all about measure theory. Of course, this doesn't say
> > anything yet about any characteristics of the conscious mind itself,
> > which is undoubtedly of importance.
> >
> >
> >
> > 2008/11/2 Brent Meeker <[EMAIL PROTECTED]
> > <mailto:[EMAIL PROTECTED]>>
> >
> >
> >     What are you calling "the process" when you've made two copies of it?
> >
> >     Bretn
> >
> >     Michael Rosefield wrote:
> >     > But, given that they are processes, then by definition they are
> >     > characterised by changing states. If we have some uncertainty
> >     > regarding the exact mechanics of that process, or the external
> >     input,
> >     > then we can draw an extradimensional state-space in which the
> >     degrees
> >     > of uncertainty correspond to new variables. If we can try and place
> >     > bounds on the uncertainty then we can certainly produce a kind of
> >     > probability mapping as to future states of the process.
> >     >
> >     >
> >     > 2008/11/2 Brent Meeker <[EMAIL PROTECTED]
> >     <mailto:[EMAIL PROTECTED]>
> >     > <mailto:[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>>
> >     >
> >     >
> >     >     Kory Heath wrote:
> >     >     > On Oct 31, 2008, at 1:58 PM, Brent Meeker wrote:
> >     >     >
> >     >     >
> >     >     >> I think this problem is misconceived as being about
> >     probability of
> >     >     >> survival.
> >     >     >>
> >     >     >
> >     >     > In the case of simple teleportation, I agree. If I step into
> a
> >     >     > teleporter, am obliterated at one end, and come out the
> >     other end
> >     >     > "changed" - missing a bunch of memories, personality traits,
> >     >     etc., it
> >     >     > doesn't seem quite correct to ask the question, "what's the
> >     >     > probability that that person is me?" It seems more correct
> >     to ask
> >     >     > something like "what percentage of 'me' is that person?"
> >     And in
> >     >     fact,
> >     >     > this is the point I've been trying to make all along - that
> we
> >     >     have to
> >     >     > accept some spectrum of cases between "the collection of
> >     molecules
> >     >     > that came out is 100% me" and "the collection of molecules
> >     that came
> >     >     > out is 0% me".
> >     >     >
> >     >     > The idea of probability enters the picture (or seems to)
> >     when we
> >     >     start
> >     >     > talking about multiple copies. If I step into a teleporter,
> am
> >     >     > obliterated, and out of teleporter A steps a copy that's
> >     100% me and
> >     >     > out of teleporter B steps a copy that's 10% me, what's the
> >     best
> >     >     way to
> >     >     > view this situation? Subjectively, what should I believe
> >     that I'm
> >     >     > about to experience as I step into that teleporter? It's
> >     hard for me
> >     >     > not to think about this situation in terms of probability
> >     - to think
> >     >     > that I'm more likely to find myself at A than B. It's
> >     especially
> >     >     hard
> >     >     > for me not to think in these terms when I consider that,
> >     in the case
> >     >     > when the thing that ends up in teleporter A is 100% me and
> >     the thing
> >     >     > that ends up in teleporter B is 0% me, the answer is
> >     unambiguous: I
> >     >     > should simply believe that I'm going to subjectively
> >     experience
> >     >     ending
> >     >     > up in teleporter A.
> >     >     >
> >     >     > I'm sympathetic to the argument that it's still not
> >     correct to frame
> >     >     > this problem in terms of probability. But I don't
> >     understand how
> >     >     else
> >     >     > to frame it. How do you (Brent) frame the problem?
> >     Subjectively,
> >     >     what
> >     >     > should I expect to experience (or feel that I'm most likely
> to
> >     >     > experience) when I step into a teleporter, and I know that
> >     the thing
> >     >     > that's going to come out Receiver A will be 100% me and
> >     the thing
> >     >     > that's going to come out of Receiver B will be 10% me?
> >     >     >
> >     >     > -- Kory
> >     >     >
> >     >     The way I look at it, there is no "I".  Kory-A and Kory-B
> >     are just two
> >     >     different processes.  We can ask how similar each one is to
> >     the Kory
> >     >     that stepped into the teleporter, but there's no fact of the
> >     matter
> >     >     about which one is *really* Kory.  And there's no sense to the
> >     >     question
> >     >     of what "I should expect to experience" because "I" is
> >     nothing but a
> >     >     process of experiencing anyway.  We could make up some legal
> >     rule
> >     >     (which
> >     >     we would need if there really were teleporters) but it would
> >     have
> >     >     to be
> >     >     based on it's social utility, not ontology.
> >     >
> >     >     Brent
> >     >
> >     >
> >     >
> >     >
> >     > >
> >
> >
> >
> >
> >
> > >
>
>
> >
>

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to [EMAIL PROTECTED]
To unsubscribe from this group, send email to [EMAIL PROTECTED]
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to