AnyOrbit is a powerful navigation algorithm for 3D virtual environments including virtual reality. This repository contains a Unity project that you can get running straight away.
To get started, try downloading the "AnyOrbit stars demo.zip" file, unzip and load. You can experience a guided tour around the brightest stars visible from Earth.
Next try the "AnyOrbit mouse stars.zip" file. This one gives you control of the orbital center directly with using mouse input. Use the mouse scroll wheel to move the red marker away and towards you and around in your field of view. This marker represents your orbital center of motion, or point of interest (POI). Rotate your head to orbit around the star field. Here is a video of the experience: https://www.youtube.com/watch?v=fmzXvdzG9HQ
The stars are a sample of data of real stars visible from earth, a reference to the data set can be found in the papers written about AnyOrbit.
AnyOrbit may be a very useful navigation tool for 3D environments including for CAD, gaming, sports and e-sports spectating, and data visualisation. In addition to using the mouse to control the point of interest at the center of the orbital camera motion, we have also attempted use of eye-tracking, directed control, in augmented reality, and also in desktop situations, also using eye-tracking for point of interest selection. Videos of some of these implementations can be found at my youtube channel.
desktop environment with tobii eye-tracking
moving to new points of interest using mouse to control POI:
guided by a director
A paper on the technique can be found here (won best poster at SUI 2016!)
A further note has been accepted for publication at ETRA, which details the technical implementation https://www.researchgate.net/profile/Benjamin_Outram2/publication/325637125_Anyorbit_orbital_navigation_in_virtual_environments_with_eye-tracking/links/5b1fdedf0f7e9b0e373ecb52/Anyorbit-orbital-navigation-in-virtual-environments-with-eye-tracking.pdf
More information including notes on the technical implementation can be found at
Thanks to collaborators: Yun Suen Pai, Tanner Person, Kouta Minamizawa, Kai Kunze