Hey,
I was wondering if anyone knows what would be the best solution for building an
application where the camera is headtracked using motion captured data provided
in realtime from a Vicon system in OpenSceneGraph. I can get the position and
rotation of the 'real' camera (cap with infrared mar
Hey all,
Many thanks for your answers.
The Vicon system uses a z-up coordinate system, with millimeter units. It sends
x,y,z coordinates for a given tracked object(in that case the cap you are
wearing) as well as rotation information in axis/angle form. The client I'm
using converts this to thr
Thanks for your answers.
I understand the theory behind this, which means that I will need to set
projection matrices for the left and right eye cameras according to my distance
to the display surface. The problem is, can these cameras be retrieved and have
their projection matrices modified be
3 matches
Mail list logo