1
Calibration and Interaction in Optical See-Through Augmented Reality using Leap Motion Kenneth R Moser Sujan Anreddy J. Edward Swan II Mississippi State University Department of Computer Science and Engineering Figure 1: Use of Leap Motion tracking data in optical see-through and video see-through AR applications. (Top) Interaction and occlusion between a user’s hand and virtual content. (Bottom) Tracking data overlayed onto a user’s hands. ABSTRACT The growing prevalence of hand and gesture tracking technology has led to an increased availability of consumer level devices, such as the Leap Motion controller, and also facilitated the inclusion of similar hardware into forth coming head mounted display of- ferings, including the Microsoft HoloLens and Moverio Pro BT- 2000. In this video, we demonstrate the utility of the Leap Motion for calibrating optical see-through augmented reality systems by employing a variation on Tuceryan and Navab’s Single Point Ac- tive Alignment Method [3]. We also showcase a straightforward method for calibrating the coordinate frame of the Leap Motion to a secondary tracking system by employing absolute orientation al- gorithms [2, 1, 4], allowing us to properly transform and visualize hand and finger tracking data from the user’s viewpoint. Our com- bined display and coordinate frame calibration techniques produce a viable mechanism for not only intuitive interaction with virtual objects but also the creation of natural occlusion between computer e-mail: [email protected] e-mail: [email protected] e-mail: [email protected] generated content and the user themselves. We believe that these techniques will be pivotal in the development of novel consumer applications for next generation head mounted display hardware. Index Terms: H.5.1 [[Information Interfaces and Presentation]: Multimedia Information Systems]: Artificial, augmented, and vir- tual realities— ACKNOWLEDGEMENTS Support for this work is provided in part by NSF awards IIS- 1320909 and IIS-1018413 to J. E. Swan II, and a NASA Mississippi Space Grant Consortium fellowship to K. Moser. REFERENCES [1] K. S. Arun, T. S. Huang, and S. D. Blostein. Least-squares fitting of two 3-d point sets. Pattern Analysis and Machine Intelligence, IEEE Transactions on, (5):698–700, 1987. [2] B. K. Horn. Closed-form solution of absolute orientation using unit quaternions. JOSA A, 4(4):629–642, 1987. [3] M. Tuceryan and N. Navab. Single point active alignment method (SPAAM) for optical see-through HMD calibration for AR. In Pro- ceedings of the IEEE and ACM International Symposium on Augmented Reality, pages 149–158, October 2000. [4] S. Umeyama. Least-squares estimation of transformation parameters between two point patterns. IEEE Transactions on Pattern Analysis & Machine Intelligence, (4):376–380, 1991. 332 IEEE Virtual Reality Conference 2016 19–23 March, Greenville, SC, USA 978-1-5090-0836-0/16/$31.00 ©2016 IEEE

Calibration and Interaction in Optical See-Through ...€¦ · Calibration and Interaction in Optical See-Through ... hand and gesture tracking technology ... in Optical See-Through

  • Upload
    lyhuong

  • View
    213

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Calibration and Interaction in Optical See-Through ...€¦ · Calibration and Interaction in Optical See-Through ... hand and gesture tracking technology ... in Optical See-Through

Calibration and Interaction in Optical See-Through Augmented Realityusing Leap Motion

Kenneth R Moser∗ Sujan Anreddy† J. Edward Swan II‡

Mississippi State UniversityDepartment of Computer Science and Engineering

Figure 1: Use of Leap Motion tracking data in optical see-through and video see-through AR applications. (Top) Interaction and occlusionbetween a user’s hand and virtual content. (Bottom) Tracking data overlayed onto a user’s hands.

ABSTRACT

The growing prevalence of hand and gesture tracking technologyhas led to an increased availability of consumer level devices, suchas the Leap Motion controller, and also facilitated the inclusionof similar hardware into forth coming head mounted display of-ferings, including the Microsoft HoloLens and Moverio Pro BT-2000. In this video, we demonstrate the utility of the Leap Motionfor calibrating optical see-through augmented reality systems byemploying a variation on Tuceryan and Navab’s Single Point Ac-tive Alignment Method [3]. We also showcase a straightforwardmethod for calibrating the coordinate frame of the Leap Motion toa secondary tracking system by employing absolute orientation al-gorithms [2, 1, 4], allowing us to properly transform and visualizehand and finger tracking data from the user’s viewpoint. Our com-bined display and coordinate frame calibration techniques producea viable mechanism for not only intuitive interaction with virtualobjects but also the creation of natural occlusion between computer

∗e-mail: [email protected]†e-mail: [email protected]‡e-mail: [email protected]

generated content and the user themselves. We believe that thesetechniques will be pivotal in the development of novel consumerapplications for next generation head mounted display hardware.

Index Terms: H.5.1 [[Information Interfaces and Presentation]:Multimedia Information Systems]: Artificial, augmented, and vir-tual realities—

ACKNOWLEDGEMENTS

Support for this work is provided in part by NSF awards IIS-1320909 and IIS-1018413 to J. E. Swan II, and a NASAMississippiSpace Grant Consortium fellowship to K. Moser.

REFERENCES

[1] K. S. Arun, T. S. Huang, and S. D. Blostein. Least-squares fitting of

two 3-d point sets. Pattern Analysis and Machine Intelligence, IEEETransactions on, (5):698–700, 1987.

[2] B. K. Horn. Closed-form solution of absolute orientation using unit

quaternions. JOSA A, 4(4):629–642, 1987.[3] M. Tuceryan and N. Navab. Single point active alignment method

(SPAAM) for optical see-through HMD calibration for AR. In Pro-ceedings of the IEEE and ACM International Symposium on AugmentedReality, pages 149–158, October 2000.

[4] S. Umeyama. Least-squares estimation of transformation parameters

between two point patterns. IEEE Transactions on Pattern Analysis &Machine Intelligence, (4):376–380, 1991.

332

IEEE Virtual Reality Conference 201619–23 March, Greenville, SC, USA978-1-5090-0836-0/16/$31.00 ©2016 IEEE