From: Dave J. <dav...@go...> - 2011-04-27 15:57:37
|
On 27 April 2011 15:13, doug sanden <hig...@ho...> wrote: > > > So the iPhone gestures are implemented by pushing modelview matrices onto the openGL stack, outside of or in addition to what normally happens in a Vrml rendering. .... > > Or use gestures to control VRML's modelview matrix through the avatar's navigation: WALK, FLY, EXAMINE except using gestures instead of mouse or keyboard input. I would suggest we use method 2 because of a potential disconnect between what the 'user' sees on the screen vs what the avatar 'sees' Dave |