On 2009-06-03, at 14:01, Tigro Spottystripes wrote:
> handshaking without haptic feedback and with SL's usual latency  
> wouldn't
> work, I imagine it would be possible to make a handshake work in SL   
> by
> adapting the IK code used to position the hand and arm while selecting
> an object at arm's reach, and then locally on each client play an
> animation over the IK moving both avatar's hands (or the IK target
> itself) up and down (if it was done server side, or just on one client
> and communicated to the other, or each client animating only it's own
> av, there would be the risk of things not matching or lagging too  
> much I
> guess)

Yes, that's something I've often thought of. I think just allowing a  
small amount of scripted IK based on having avatar joints track a prim  
if they were within a certain distance would allow this and more. With  
or without camera tracking... you'd wear a handshake attachment that  
would track the matching prim. Then if one or both users had camera  
tracking code that would be used to drive the handshake, if not the  
handshake would track the avatar's canned animations.

And you could have sit or stand animations with appropriate scripted  
overrides actually put your hands on your hips or your knees no matter  
what your avatar size.

llTrackAnimation(integer joint, key target, float strength);
_______________________________________________
Policies and (un)subscribe information available here:
http://wiki.secondlife.com/wiki/SLDev
Please read the policies before posting to keep unmoderated posting privileges

Reply via email to