On 2009-06-03, at 14:01, Tigro Spottystripes wrote: > handshaking without haptic feedback and with SL's usual latency > wouldn't > work, I imagine it would be possible to make a handshake work in SL > by > adapting the IK code used to position the hand and arm while selecting > an object at arm's reach, and then locally on each client play an > animation over the IK moving both avatar's hands (or the IK target > itself) up and down (if it was done server side, or just on one client > and communicated to the other, or each client animating only it's own > av, there would be the risk of things not matching or lagging too > much I > guess)
Yes, that's something I've often thought of. I think just allowing a small amount of scripted IK based on having avatar joints track a prim if they were within a certain distance would allow this and more. With or without camera tracking... you'd wear a handshake attachment that would track the matching prim. Then if one or both users had camera tracking code that would be used to drive the handshake, if not the handshake would track the avatar's canned animations. And you could have sit or stand animations with appropriate scripted overrides actually put your hands on your hips or your knees no matter what your avatar size. llTrackAnimation(integer joint, key target, float strength); _______________________________________________ Policies and (un)subscribe information available here: http://wiki.secondlife.com/wiki/SLDev Please read the policies before posting to keep unmoderated posting privileges
