
MOTION TRACKING WITH REFLECTIONS 
3D pointing device with self-calibrating mirror system 
Shinichi Fukushige, Hiromasa Suzuki 
Department of Precision Enginieering, The University of Tokyo,7-3-1 Hongo Bunkyo, Tokyo, Japan 
Keywords:  Interactive pointing device, 3D input, reflections, motion tracking. 
Abstract:  We propose a system that uses a camera and a mirror to input behaviour of a pointer in 3D space. Using 
direct and reflection images of the pointer obtained from single directional camera input, the system 
computes the 3D positions and the normal vector of the mirror simultaneously. Although the system can 
only input the ‘‘relative positions’’ of the pointer, in terms of 3D locations without scale factor, calibration 
of the mirror orientation is not needed. Thus, the system presents a very simple and inexpensive way of 
implementing an interaction device. 
1 INTRODUCTION 
Input devices for processing 3-dimensional (3D) 
computer-generated models are divided into two 
types - those with a 2-dimensional (2D) interface 
and those with a 3D interface. 2D input devices, 
such as mice, tablets, and touch monitors, are used 
more widely than 3D input devices because of their 
simplicity and easy-use. However, because the target 
is in 3D space, 2D-base-input equipment needs 
several constrains and restrictions of pointer 
movements (Sugishita, 1996) (Zeleznik, 1996) 
(Branco, 1994). Preconditions, for translating 2D 
input operations into 3D, often hinder the intuitive 
input operations of designers.  
Therefore, recently, various devices have been 
developed which can directly indicate the position 
on 3D space. 
Currently, however, 3D input devices are not 
widely used among general users and are not used as 
general-purpose tools due to their costs and 
complexity, requiring special sensors for treating 
magnetism, ultrasonic waves  and laser, or having 
complex structures, such as joint or wire 
mechanisms or stereo camera systems (Kenneth, 
1994) (Sato, 2000) (Smith, 1995) (Turban, 1992).  
Stereovision is commonly used to calculate 3D 
positioning of a pointer by implementing images 
from more than one single camera (Faugeras, 1993) 
(Yonemoto, 2002) (Xu, 1996) (Longuet-Higgins, 
1981). However, processing multiple video images 
in real time thus requires large amount of CPU 
resources or special hardware. Furthermore, these 
methods involve synchronization and complex 
computations that usually require an initial 
calibration phase. Since multiple cameras must be 
placed at separated positions to ensure full 3D 
restoration accuracy, it is difficult to miniaturize 
such systems.  
We would like to provide a simple 3D pointing 
device that users can handle easily and with a feeling 
of familiarity. This paper proposes a system for 
assuming the 3D motion of a pointer in real time by 
inputting a single video image of the pointer tip with 
a mirror reflection. Conventionally, in order to 
determine an object's 3D positioning from a single 
view, the shape and size of the object or multiple 
markers on it should be recognized simultaneously. 
And the restoration accuracy of them are low in the 
direction of the optical axis. 
 The proposal method is different from the 
method of Lane et al (Lane, 2001), which also uses a 
mirror reflections and estimates the ‘‘absolute’’ 3D 
positions. This method needs manual calibration and 
must divide the 3D space into a mirror reflection 
area and an inputting area. 
We propose using a mirror system with self 
calibration which estimates the relative 3D positions 
of the pointer. ‘‘Relative positions’’ mean that the 
restored x, y, z coordinates of the pointer include the 
same unknown parameter regarded as a scale factor. 
However, in the 3D pointing usage, the scale 
factor can be set freely by a user, because the fine 
428
Fukushige S. and Suzuki H. (2006).
MOTION TRACKING WITH REFLECTIONS - 3D pointing device with self-calibrating mirror system.
In Proceedings of the First International Conference on Computer Vision Theory and Applications, pages 428-434
DOI: 10.5220/0001377404280434
Copyright
c
 SciTePress