SlideShare a Scribd company logo
1 of 9
Download to read offline
IPT & EGVE Workshop (2005)
R. Blach, E. Kjems (Editors)




     Tangible 3D: Hand Gesture Interaction for Immersive 3D
                           Modeling

                              Hyosun Kim1,2 , Georgia Albuquerque1 , Sven Havemann1,2 , Dieter W. Fellner1,2

                                                1 Institut für ComputerGraphik, TU Braunschweig, Germany
                                       2 Institut   für ComputerGraphik und WissensVisualisierung, TU Graz, Austria




        Abstract

        Most of all interaction tasks relevant for a general three-dimensional virtual environment can be supported by
        6DOF control and grab/select input. Obviously a very efficient method is direct manipulation with bare hands,
        like in real environment. This paper shows the possibility to perform non-trivial tasks using only a few well-known
        hand gestures, so that almost no training is necessary to interact with 3D-softwares. Using this gesture interaction
        we have built an immersive 3D modeling system with 3D model representation based on a mesh library, which is
        optimized not only for real-time rendering but also accommodates for changes of both vertex positions and mesh
        connectivity in real-time. For performing the gesture interaction, the user’s hand is marked with just four fingertip-
        thimbles made of inexpensive material as simple as white paper. Within our scenario, the recognized hand gestures
        are used to select, create, manipulate and deform the meshes in a spontaneous and intuitive way. All modeling
        tasks are performed wirelessly through a camera/vision tracking method for the head and hand interaction.
        Categories and Subject Descriptors (according to ACM CCS): I.3.6 [Computer Graphics]: Interaction Techniques




1. Introduction                                                                      Bare hands are the ultimate form of three-dimensional
                                                                                  human-computer interface (3D-HCI). The vision guiding the
                                                                                  work in this paper is the possibility to perform non-trivial
                                                                                  tasks using only a few well-known gestures, so that almost
                                                                                  no training is necessary to interact with 3D-software. Today,
                                                                                  interfaces of this type still seem to be science-fiction, as they
                                                                                  appear for instance in the recent Matrix and Minority Report.
                                                                                  In both movies, computers are controlled by pointing with
                                                                                  bare fingers at 2D interfaces floating in 3D. This has inspired
                                                                                  us to look for what is possible with today’s technology to get
                                                                                  as close as possible to such an intuitive an un-instrumented
                                                                                  way for operating computers.

                                                                                     A less far-fetched and more concrete usage scenario for
                                                                                  such kind of technique are 3D shopping terminals, located
                                                                                  for instance in shopping malls or in specialized stores. They
                                                                                  may be thought of as a generalization of the familiar selling
                                                                                  machines with a 2D interface, such as ATMs (for drawing
                                                                                  cash money), or terminals for selling railway tickets. Antici-
Figure 1: User interacts with a 3D virtual model by using                         pating the increasing use of 3D technology also in the com-
non-instrumented hand gestures                                                    mercial market, easy to use 3D interfaces and applications
                                                                                  are mandatory.

c The Eurographics Association 2005.
192                         H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling

   This leads to two fields of research that are tightly re-           tesselated, i.e., converted to triangles, in order to display it.
lated: First, to find the right kind of input device technology,       But object representation that require costly pre-processing
and second, to identify suitable interaction metaphors and            before they can be displayed are ruled out when interactive
devise general design guidelines for responsive 3D appli-             update rates are required.
cations. This paper presents our first prototype system that
integrates both of these aspects, along with the the design              Triangle meshes [KBS00] and also point clouds
decisions and experiences we’ve made.                                 [PKKG03] are routinely used for direct modeling of free-
                                                                      form surfaces. They permit arbitrary surface deformations
   Concerning the interface device, our choice was to use             while maintaining the surface connected or even closed
a vision-based tracking approach, based on a pair of cam-             (“water-tight”). Both representations can also work in a
eras with black-light sources. The system has been devel-             multi-resolution fashion, a feature which is indispensable for
oped and set up by ourselves, as described in a previous pa-          adaptive real-time display. They have the drawback, how-
per [KF04]. It is capable of tracking the fingertips of one            ever, of being low-level representations: A distinction be-
hand and also the position of the user’s head. It is used in a        tween different parts of a 3D object on the level of indi-
passive stereo back-projection environment, so that the user          vidual points or triangles is difficult. This requires to main-
can move freely in a space of approximately (2m)3 . Conse-            tain a segmentation of the surface into different patches. A
quently, the user has a stereo view on the scene, which fol-          good compromise are therefore subdivision surfaces, which
lows the head position to optimally adapt to the user’s point         provide free-form manipulation, have the multi-resolution
of view. A typical situation during 3D interaction is shown           property, and are also compatible with polygonal meshes.
in Figure 1.                                                          Incidentally there are some modeling systems which con-
                                                                      tain their own modeling primitives, and allow to create more
   This choice of an interface device was motivated by our
                                                                      complex objects by grouping together simpler ones. The
targeted application scenario. Our main premise is to be as
                                                                      drawback is that they are specifically designed for the re-
minimalistic as possible on the user requirements (see Figure
                                                                      spective applications [KHK∗ 96, HG96, SIS02].
5 (c)): We consider a pair of black gloves with bright white
marks on the fingertips plus a pair of polarized glasses with a           The usefulness of interaction between user and virtual
white frame as being quite acceptable, and maybe even mass            models is the issue of VR based modelers. Debunne et al.
market compatible, both in price and in practical handling.           [DDCB01] and Hua et al. [HQ01] use a haptic feedback in-
The stationary system setup is not prohibitively expensive            terface for surface manipulation. Such devices provide the
either: Basically a pair of projectors and a pair of cameras are      user with the physical feeling of virtual objects, but they gen-
sufficient. Both of these components are currently becoming            erally restrict the available workspace to a desktop environ-
less pricy through the pressure from the mass market.                 ment. VLEGO [KHK∗ 96] and THRED [SG94] use a pair
   Concerning our exemple application, our choice was                 of 6-DOF magnetic trackers (position + orientation data)
the creation and manipulation of three-dimensional objects,           with additional buttons to issue commands, such as to grab
herein referred to as 3D modeling. While this may be an ob-           or release the virtual object. These systems support com-
vious choice for a genuine 3D interface, it has nevertheless          plex modeling tasks by two-handed manipulation, but their
turned out to be quite challenging: Controllability definitely         input devices are unfortunately less intuitive than a glove-
is an issue when working in free 3D, without a table to lean          type input device [NUK98]. Schkolne et al. [SPP01] have
on or to guide the user’s movements. Again the reason was             introduced the Surface Drawing approach for 3D shape cre-
to make no assumptions on the physical constraints of our             ation, which uses a glove-type sensor together with a num-
system, and to gain experiences with ’free 3D’.                       ber of tangible tools: The motion of the hand draws strokes
                                                                      in the 3D modeling workspace, and the thumb activates the
                                                                      creation of a surface. Other tangible interfaces, specifically
2. Related Research                                                   designed for their respective systems, were also found in
                                                                      [VRCW04, SIS02]. However, to our knowledge, bare hands
Interactive 3D modeling is one of the areas where immer-              have rarely appeared in the virtual modeling field, in spite
sive VR can be expected to be most beneficial. There is                of their high functionality and excellent naturalness, as was
a trade-off, however, between ease of manipulation and                noted by Pavlovic [PSH97].
rendering efficiency. One of the crucial ingredients is the
choice of a suitable 3D object representation, which must                Concerning display devices, head-mounted-display
be suited to the target application. See [Kni98] for a dis-           (HMD) can be found more often [KSW00, BO92, FS99]
cussion of the implications related to the components of              than others, such as a virtual workbench [SPP01, SIS02], or
a VR system. Implicit surfaces for instance are an exam-              a CAVE [DBW∗ 00]. This may be due to the high usability
ple of a model representation with intuitive degrees of free-         of the HMDs, which do not require fully equipped display
dom [DBW∗ 00, HQ01, DTG96]. But they tend to have slow                environments. But besides the high price, the drawback
update rates when the model is interactively changed, due to          of HMD devices is that only one person can follow the
the fact that the graphics hardware requires a surface to be          interaction.

                                                                                                          c The Eurographics Association 2005.
H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling                    193




                                                 (a) direct hand manipulation in real environment




                                       (b) hand gestures for 3D interactive modeling in virtual environment
Figure 2: The five static hand gestures used in our system: Pause, point, grab, rotate, scaling. The gestures mimic characteristic
human movements for object manipulation. They are mostly self-explanatory, except for the scaling gesture, which needs to be
learned.



3. Hand Gesture based 3D Modeling                                            users. So to keep things simple for the user, we have limited
                                                                             ourselves to a fixed set of five static gestures which are easy
Focusing on interaction methods for virtual applications
                                                                             to memorize, shown in Figure 2. Another advantage is a re-
in large physical space we find many instances of using
                                                                             duced processing load for gesture recognition, since in most
computer-vision based techniques. The popularity of this ap-
                                                                             cases the identical configuration can be used over a number
proach comes from the fact that the connection between the
                                                                             of frames. A typical scenario of using these gestures is the
user and a computer system is wireless. The unrestrained
                                                                             following:
movement encourages a liberal and intuitive interaction. The
fact that the user can freely move on the other hand implies                 Interaction with virtual objects: Pinch the thumb and an
that the gesture recognition must be very robust. The set of                   another finger after pointing to a target object. The se-
meaningful gestures must therefore be carefully chosen so                      lected objects are manipulated in the same way with the
that they can be both issued easily by the user, and robustly                  hand motions.
detected by the system.
                                                                                • Grab(x, y, z): Close the hand while pinching the thumb
                                                                                  and index finger, then move it.
3.1. Gesture Scenario                                                           • Rotate(x, y, z): Open the hand while pinching the
An interaction technique is a method that allows the user to                      thumb and index finger, and then rotate it.
perform some tasks in the system as a part of the user in-                      • Scale(x, y, z): Open the hand while pinching the thumb
terface. Depending on the application, it supports different                      and middle finger, and then move it.
modes of interaction, each of them can be issued by a dif-                   Interaction with a virtual cursor: Release the pinching
ferent gesture. Since our chosen application field is shape                     states. This is to handle the cursor of application.
modeling, the mapping from gestures to interaction modes
should mimic the way how hands are used for sculpting and                       • Point(x, y, z): Close the hand except for the thumb and
manipulation in the real world. Furthermore, it should sup-                       index finger, which points straight at the target object.
port all necessary means for the basic modeling functions                         This orders a cursor to move to the position indicated
through gestures.                                                                 by the index finger.
                                                                                • Pause(x, y, z): Open the whole hand.
   Hand gestures used for 3D-HCI can be classified into two
categories: dynamic gestures and static gestures [PSH97].                       VR applications have to support four main categories of
Using dynamic gestures generally requires a larger amount                    interaction tasks: navigation, selection, manipulation, sys-
of time for training the interaction system, which should                    tem control [BKL∗ 00]. Table 1 shows the commands sup-
also be personalized for each user to get the best recognition               ported by our hand gestures. System control is achieved
performance. Yet our system should also support occasional                   through a graphical 3D menu, which is operated in the same

c The Eurographics Association 2005.
194                        H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling




                                                                 (a)




                                                                 (b)




                                          (c)                                                                   (d)
Figure 3: Hand gesture interaction: (a) object manipulation and deformation by (b), (b) the sequence of gesture interaction
– pause, point(object), rotate, translate (z-direction), point(object’s vertex), translate(object’s vertex), and rotate, (c) object
creation from a simple primitive by (d), (d) the shape design drawn along the hand movement



  Selection             (Point(x, y, z) or Pause(x, y, z))             6DOF manipulation and selection, our gesture commands
                      + (Grab(x, y, z) or Rotate(x, y, z))             include scaling function as well, which is useful especially
  Manipulation        Translation            Grab(x, y, z)             for shape modeling. It is triggered by pinching the thumb
                                          or Rotation(x, y, z)         and middle finger, while pinching the thumb and index fin-
                       Rotation              Rotate(x, y, z)           ger goes into rotation. The use of different pinching combi-
                    Uniform scaling           Scaling(z)               nations shows the possibility of extension of the gesture set
                                                                       if necessary.
                 Table 1: Interaction tasks
                                                                       3.1.1. Visual feedback
                                                                       To control graphical objects the user ’touches’ them with the
way as object selection (see section 3.3), namely by point-
                                                                       interaction interfaces. Without the physical collision, how-
ing. Navigation is done mainly via head tracking (see sec-
                                                                       ever, it is not easy to perceive the distance between the in-
tion 4.1.3).
                                                                       terface and the targets. To reduce the user’s cognitive load,
   All gestures are associated with a 3D position, which can           our modeler provides a number of different gizmos, proxy
obviously be changed by moving the hand. So the rotate ges-            objects which are occasionally set into the scene as visual
ture can simultaneously be used to translate an object. One            feedback clues. Gizmos are not part of the 3D model, but
crucial point in designing interaction modes is to take into           serve as a visual representative for a particular operation, to
account the limited freedom of movement of the different               indicate a specific mode, or to give information about the
parts of the human body. Especially the hand can only per-             status of the interaction.
form limited rotations, so it is important to allow absolute
                                                                          The great advantage of gizmos is that they also improve
as well as relative motions. Consequently, our system allows
                                                                       the robustness of the interaction tasks: Certain modeling op-
for cumulative rotations by repeating rotate-and-pause ges-
                                                                       erations become much more controllable if they are not di-
tures.
                                                                       rectly issued by gestures, but indirectly: The gizmo is ma-
  In addition to the basic 3D interaction functionalities of           nipulated by gestures, and the gizmo in turn issues the mod-

                                                                                                          c The Eurographics Association 2005.
H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling                     195

eling operation, so that it is executed in a more controlled
way. One simple example is the arrow-shaped model which
is used as a 3D cursor to represent the user’s hand in 3D vir-
tual space. The position of the arrow is determined by the
projection of the user’s hand position in real space to the vir-
tual 3D scene. So it is always in front of the scene, and it
works like the familiar 2D mouse pointer on a 2D desktop.
When there is a collision between the arrow cursor and a
                                                                             Figure 4: Face representation change by seting the sharp-
target object, another gizmo like a control ball or an object
                                                                             ness flags of the edges (from sharp to smooth)
bounding-box appears. For other tasks such as rotating or
twisting objects, a tripod gizmo with three coordinate axes
is drawn, which moves in correspondence to the user’s hand
orientation.                                                                 mesh greatly improved the controllability of surface manip-
                                                                             ulation: Distinct graphical feedback can be given to the user
                                                                             who is going to manipulate either a vertex, an edge, or a face
3.2. Modeling Operations
                                                                             of the control mesh. The mesh as intermediate structure also
Figure 3(a-b) illustrates some deformations of a 3D object                   facilitates the selection of specific parts of the object (sur-
together with the corresponding gestures used in the each                    face navigation), and has proven less tedious than a direct
step. The respective gestures in the hand-motion sequences                   manipulation of the many triangles from the tesselation via
also indicate the required modeling operations for the steps,                gestures.
as they correspond a direct manipulation tool for creating or
                                                                                Our model data are maintained in the Wavefront .obj
modifying 3D shapes.
                                                                             file format, thus the input/output objects can be compatibly
                                                                             shared with other modeling products supporting this popular
3.2.1. Shape Representation                                                  format.
The basis of the modeling operations is a mesh data struc-
ture for interactive 3D design and manipulation, the Com-                    3.2.2. Selection
bined BRep (CBRep). It supports incremental on-line update                   The pointing approach [VdPRHP99] is used for target selec-
and adaptive rendering through a multi-resolution tessela-                   tion. Note that only visible targets in the user’s field of view
tion [HF04].                                                                 can be candidates for selection. But since our system sup-
   The special feature of Combined BReps is that they bridge                 ports head tracking (see section 4.1.3), it allows to change
the gap between polygonal and free-form models. The mesh                     between different views in a natural way, and unselectable
is an ordinary polygonal mesh by default, but any part of                    targets in one view may become selectable in another. This
the mesh can also be used as the control mesh of a Cat-                      approach is advantageous because:
mull/Clark subdivision surface. So a CBRep mesh can have                     - In most cases, an occluded object can be selected without
both smooth free-form and polygonal parts, with crease                          actually removing the objects obstructing it. This possi-
curves establishing a link between them. Both types of                          bility depends of course on the nature of the scene, and
geometry are integrated in a single data structure by associ-                   works best with scenes that are not too ’opaque’.
ating a boolean sharpness flag with each edge, to distinguish                 - The target must be searched only in 2D, not in 3D. This
between sharp and smooth edges. In regions with only sharp                      enables the use of fast screen-space selection methods,
edges, BRep faces are triangulated, just as with standard                       which accelerates the selection of very distant objects.
polygonal meshes, whereas in smooth regions, the tessela-
tion is created using recursive subdivision. By just chang-                     Smooth parts in our model are represented in Cat-
ing the sharpness flag, it is possible to switch between both                 mull/Clark subdivision surfaces. The individual triangles of
representations: Any face that contains at least one smooth                  the tesselation usually cover only a few pixels, so there are
edge is treated as a Catmull/Clark surface (see Figure 4).                   obvious difficulties to accurately intersect a specific trian-
Combined BReps are optimized not only for real-time ren-                     gle by a ray shot from the tracked head position through the
dering, but they can also accommodate for changes of both                    tracked position of the pointing finger. The solution is to re-
the vertex positions and the mesh connectivity in real-time,                 strict the selection to the vertices, edges, and faces of the
by incrementally updating the changed parts of the tessela-                  control mesh, which are much easier to uniquely identify –
tion.                                                                        both for the user and for the system.

   The availability of CBReps was the basis of the plan to
                                                                             3.2.3. Navigation
realize a 3D modeling application which made use of our vi-
sion based tracking system as an intuitive input device. One                 The size of the work space in front of the projection screen
important lesson learned was that the existence of a control                 is limited by the viewing angle of the cameras used to track

c The Eurographics Association 2005.
196                          H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling




                (a)                                (b)                               (c)                                (d)
Figure 5: System environment: (a) back-projection screen, (b) two projectors for passive stereo, (c) user interfaces – white
paper-rimmed polarized glasses and fingertip markers (d) a black light



the user’s motion. In particular, it is not possible to walk           3.3. Changing the modeling mode
completely around a 3D object, so it must be possible for the
                                                                       The modeler needs to support various useful operations in
user to navigate within the scene. In interaction mode, we
                                                                       terms of the system productivity. Using many kinds of ges-
therefore provide the usual possibilities of translation and
                                                                       tures for each operation however can easily overloaded the
zooming supported by head tracking (see section 4.1.3), as
                                                                       user with a large amount of the gesture learning. Providing
well as the rotation around a particular target (’orbit’). In
                                                                       a graphical menu can be the suitable solution for changing
addition to this, objects can be moved in the scene. Target
                                                                       the operation modes, instead of developing new gestures or
selection is done as usual, by pointing.
                                                                       adding second interfaces. This is attributable to the fact that
                                                                       users in general already have rich experiences and skills in
3.2.4. Global vs. Local Deformations
                                                                       using menus and windows. The menu is placed on the boder
By moving individual vertices and faces of the mesh, the               of the scene, typically considered as the ’dead space’ of the
user has fine-grained local control about the shape. But the            screen. The user can choose the operation by using the hand
shape is not only defined by the position of the vertices, but          gestures, like selecting a model object.
also by the connectivity of the mesh – as the case with all sur-
face types based on subdivision. So to allow for controlled
changes of mesh connectivity, we have added the Euler op-              4. System Setup
erators to the set of manipulation modes. Specifically, the
                                                                       The hardware of our modeling environment consists of a
user can
                                                                       passive stereo back-projection display with two projectors,
- insert an edge to split a face in two,                               one PC to run the vision tracking software, and another PC
- insert a new vertex on an edge,                                      to run the modeling application. Both computers are con-
- and split a vertex in two, connected by a new edge.                  nected via standard ethernet.
   It should equally be possible to smoothly deform larger
parts of the mesh, which involves moving a number of                   4.1. Camera/Vision tracker
mesh vertices. We have therefore adopted the Twister ap-
proach from Llamas et al. [LKG∗ 03] to use it with Com-                The space in front of the back-projection screen (about 2.5m
bined BReps. It works by applying a smooth space warp                  × 2m × 2m) is the work space for interaction. This volume
to all vertices inside a spherical region of influence of ad-           is typically of sufficient size for any kind of modeling ap-
justable radius. So the user first selects a warping-region in-         plication. The cameras are mounted near the upper edge of
cluding several vertices on the control mesh of a model, and           the screen, covering the interaction area but not limiting the
moves it then to a new position, with the hand rotating in or-         movement of the users.
der to tilt, bend, or twist the shape. Since this can have drastic
effects on the global shape, this method is quite sketchy and
                                                                       4.1.1. Interaction environment setup
suitable for simulating the spatial deformation of a rubber-
like material. It is less applicable though for obtaining pre-         The projection display environment usually involves a dim
cise results, and is not suitable for CAD-like construction            light condition to achieve a high-contrast screen projection.
tasks.                                                                 On the other hand the screen emits a variable amount of

                                                                                                           c The Eurographics Association 2005.
H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling                      197

light, so it is not feasible for the vision-based tracking to as-            frame the center of the pair of glasses is computed. The left
sume a static lighting condition. Variable lighting can deteri-              and right eyes are assumed to be 3 cm to the left and to the
orate the quality of the images captured by a camera, which                  right of the center of the glasses. We use linear polarizers
in turn affects the image processing, and thus the results for               for passive stereo, where the light for one eye is polarized
tracking features in the images. To solve these problems we                  perpendicular to the light for the other eye. It means that the
use a ’black-light’ [Rye96] source and white paper markers                   users must keep their head level or they will lose the stereo
on the fingertips: The black light stimulates enough fluores-                  effect as the polarization of their stereo glasses no longer
cent emission from the white paper markers to distinguish                    matches that of the projector filters. So only the position of
the detection features even on the low-contrast camera im-                   the head is needed. These positions establish the two view-
age. Since the black light is barely visible, it hardly influ-                points for the stereo rendering. Figure 6 shows the typical
ences the projection quality, and only the markers are high-                 scenes generated from the tracking of the head.
lighted [KF04]. Using white paper as a marker material has
obvious advantages: it is readily available, of low cost, and
                                                                             4.2. Network Connection
of light weight. Since the black light illuminates a white T-
shirt in the same way it illuminates the white paper, this ap-               The whole system can be roughly divided into two parts:
proach however builds on the assumption that users do not                    the modeling system and the tracking system. They are con-
wear white clothes or reflective accessories.                                 nected via our standard ethernet network. The tracking sys-
                                                                             tem executes the computer-vision algorithms to monitor the
4.1.2. Interface design                                                      gestures of the user’s hand, and the position of the head. This
                                                                             information is sent to the second component, the modeling
In order to let a user’s field of view cover the whole screen,                system, for viewpoint maneuvering and object interaction.
the user generally keeps the adequate distance to the screen.                The transmission of the data from the tracking system to the
Symmetrically, the cameras and the user’s hand are at the                    modeling system is done via TCP/IP packets containing the
same distance, which may reduce the details of the tracking                  description of user motion. To maintain a nearly real-time
features to a limited image resolution. So, if a hand is marked              communication the send-receive procedure works in its own
with too many markers, segmenting each marker will cause                     thread on both machines, so that it is independent from the
problems. Therefore we limit the number of markers to only                   rendering of the scene.
four. While a user performs an interaction, the postures and
positions of the hand will be continuously changed and up-
dated. Though the unrestricted motions are desirable for a                   5. Experimental Result
natural interaction, it often can cause some difficulties such                We have developed the modeling system on an AMD Athlon
that the markers turn away from the camera view. To com-                     XP 2200+ PC with 512 MB of RAM and a GForce4 TI4600
pensate this we experimented with different marker arrange-                  graphic card (128 MB). This platform provides real-time
ments with respect to shape and position. When ring shaped                   updates (about 30 − 50 frame/sec) of stereo scene projec-
markers are placed at finger joints, the distance between                     tion. The camera/vision tracker runs on an AMD Athlon XP
markers was too short to distinguish them reliably. In mo-                   2000+ PC with a DFG/BW-1a frame grabber board and two
tion they also might get lost due to occlusion by the fingers                 SONY XC-55 progressive scan cameras, with 30 frame/sec
themselves. As a consequence, our markers are placed on                      in average to track the user’s head and hand simultaneously.
the fingertips (thumb, index, middle and little finger) which                  The rendering frame rate and the scene quality is domi-
are less frequently occluded, compared to other parts of the                 nated by the model complexity and the tessellation update on
hand. A marker is designed like a thimble to fix easily to the                shape deformation. To generate smooth and natural anima-
fingertips. This compact shape can allow comfortable and                      tion while modeling, the scene rendering always has higher
natural motion according to our experience.                                  frequency than the data transmission from the tracking sys-
   For the head tracking interface, the polarized glasses is                 tem.
rimmed with white paper. Figure 5(c) shows the thimble-                         The power as a virtual modeling system can be evaluted
shaped fingertip markers and the white paper-rimmed polar-                    by the performance of building 3D models through the given
ized glasses.                                                                interaction interfaces, hand gestures in our case. It is closely
                                                                             related with and mainly depends on harmony between the
4.1.3. Head tracking                                                         interaction method and interactive model representation.
Eye positions are needed in order to generate accurate stereo                   First, the gesture-based interaction could instantly lead to
views according to the user’s perspective. However, because                  the understanding of each gesture from the users after a short
the user wears stereo glasses, the actual eyes are not visi-                 explanation. They could easily make a set of gestures with
ble and cannot be tracked. So we approximate the eye posi-                   their hand and invoke the reaction from the system. The case
tions by the positions of the glasses on the head. First, we set             of medium level of modeling work could be quickly real-
up the polarized glasses as a second tracking target. In each                ized as well. For more demanding operations such as ’shape

c The Eurographics Association 2005.
198                        H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling




                                        Figure 6: Viewpoint movement by head tracking



design’, we noticed that it needs somewhat more training to          ’workhorse’ for intuitive 3D manipulation in our group. But
produce the intended results right away.                             gestures alone do not make for a complete system. Our inter-
                                                                     action interface builds upon three components that are used
   Figure 7 (a-e) show some results produced by one of the
                                                                     in conjunction:
authors, to represent our model creation in gesture drawings.
It was certainly effective for wide range of shape sketching         a) hand gestures,
and modeling. The time spent to get simle models like ques-          b) gizmos as virtual controller objects, and
tion mark, alphabet characters and flower did not exceed              c) textual menus for mode changes.
1-2 minutes, since they were simply drawn along the path of
hand movement in space. The model in Figure 7 (c) is from            Our experience so far has revealed that the real power of
3D croquis drawing made in a few minutes through similar             ’tangible 3D’ is unleashed only by an intimate fusion of
manner to draw on paper with a pencil. Not only the smooth           all components. Users are enthusiastic about the ’hand-as-
shapes, but also the polygonal forms could be achieved just          a-tool’ metaphor, but the controllability and the precision of
by changing the sharpness flag of edges in the control mesh.          the interaction must still be improved. So our future plans
Figure 7 (f) shows a short process of model generation, start-       are to even better integrate the different methods: Examples
ing from cubical primitives to more refined form.                     include using gestures for mode changes on a per-gizmo ba-
                                                                     sis, written annotations everywhere to clearly flag the current
   Once the source object was prepared, the other defor-             mode states, numeric measurements producing ’tangible val-
mation operators were used for refining or modifying the              ues’ that can be put on a shelf, or used as input values for
model. Figure 7 (a), (d) are the examples of a bit more com-         other operations, and many more.
plex shape creation followed by additional modeling oper-
ations. The tree object in Figure 7 (g) is one of the exam-             Our focus is, and will be, on building a generic infra-
ple models with a complex geometry and a large number                structure for direct, semi-immersive 3D interaction. We be-
of small polygonal facets. The detail control for this kind          lieve our system has distinguished advantages: It is quite af-
of models, such as editing the end of each twig, was quite           fordable, actually, and its performance will improve with in-
difficult without any haptic feedback. In this case the ex-           creasing computer, e.g., recoginition, power. The key com-
tensive deformation operator like a simultaneous control of          ponent for making the step from laboratory to field applica-
multiple vertices was certainly useful. More advanced shape          tion though is the software.
editing were demonstrated in Figure 7 (h-j) shape tilting and
                                                                        Our goal is a software interface that permits to easily
bending. As these experiments indicate, the modeling sys-
                                                                     transfer any given 3D application to such a tangible mod-
tem could easily and quickly reproduce the shapes of the
                                                                     eling space, by wrapping its functionality using gestures,
existing objects by the hand gestures.
                                                                     gizmos, and floating text. This will ultimately give creative
   The results obtained so far are quite promising. In the near      people a new media at hand, to use it in fields such as arts,
future, we will therefore perform systematic usability studies       industrial design, and architecture.
with larger groups of untrained users. We hope that an analy-
sis of the user feed-back will reveal which of the proposed
modeling methods works best for which kinds of application           References
scenario.                                                            [BKL∗ 00] B OWMAN D., K RUIJFF E., L AV IOLA J.,
                                                                       P OUPYREV I., M INE M.: The art and science of 3d inter-
6. Conclusions and Future Work                                         action. Tutorial Notes from the IEEE Virtual Reality 2000
                                                                       Conf., IEEE Computer Society (2000).
Hand gestures have proven a very expressive, convenient,
and robust tool, and they have quickly become a popular              [BO92]     B UTTERWORT J. D. A. H. S., O LANO T.: 3dm:

                                                                                                         c The Eurographics Association 2005.
H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling                     199

  A three dimensional modeler using a head-mounted dis-                         ACM Transactions on Graphics 22, 3 (July 2003),
  play. Computer Graphics. In Proc. of Symposium on In-                         663–668.
  teractive 3D Graphics 25(2) (1992), 135–138.                               [NUK98] N ISHINO H., U TSUMIYA K., KORIDA K.: 3d
[DBW∗ 00] D EISINGER J., B LACH R., W ESCHE G., R.                             object modeling using spatial and pictographic gestures.
  B., S IMON A.: Towards immersive modeling - challenges                       In Proc. of the ACM symposium on Virtual reality software
  and recommendations: A workshop analyzing the needs                          and technology (1998), 51–58.
  of designers. In Proc. of the 6th Eurographics Workshop
                                                                             [PKKG03] PAULY M., K EISER R., KOBBELT L. P.,
  on Virtual Environments (2000).
                                                                               G ROSS M.: Shape modeling with point-sampled geom-
[DDCB01] D EBUNNE G., D ESBRUN M., C ANI M.,                                   etry. ACM Transactions on Graphics 22, 3 (July 2003),
  BARR A.: Dynamic real-time deformations using space                          641–650.
  and time adaptive sampling. In Proc. of SIGGRAPH’01,
                                                                             [PSH97] PAVLOVIC V., S HARMA R., H UANG T.: Visual
  ACM Press (2001), 31–36.
                                                                               interpretation of hand gestures for human-computer inter-
[DTG96] D ESBRUN M., T SINGOS N., G ASCUEL M.:                                 action: A review. IEEE Transaction on Pattern Analysis
  Adaptive sampling of implicit surfaces for interactive                       and Machine Intelligence 19, 7 (1997), 677–695.
  modelling and animation. Computer Graphics Forum
                                                                             [Rye96] RYER A.:           Light measurement handbook.
  15(5) (1996), 319–327.
                                                                               http://www.intl-light.com/handbook/ (1996).
[FS99] F UHRMANN A., S CHMALSTIEG D.: Concept and
                                                                             [SG94] S HAW C., G REEN M.: Two-handed polygonal
  implementation of a collaborative workspace for aug-
                                                                               surface design. In Proc. of ACM Symposium on User In-
  mented reality. GRAPHICS ’99 18(3) (1999).
                                                                               terface Software and Technology (1994), 205–212.
[HF04] H AVEMANN S., F ELLNER D.: Progressive Com-
  bined BReps - Multi-Resolution Meshes for Incremental                      [SIS02] S CHKOLNE S., I SHII H., S CHROEDER P.: Tangi-
  Real-time Shape Manipulation. Tech. rep., Institute of                       ble + virtual = a flexible 3d interface for spatial construc-
  Computer Graphics, TU Braunschweig, 2004.                                    tion applied to dna. CS TR 2002.011, California Institute
                                                                               of Technology (2002).
[HG96] H ALLIDAY S., G REEN M.: A geometric model-
  ing and animation system for virtual reality. Communica-                   [SPP01] S CHKOLNE S., P RUETT M., P. S.: Surface draw-
  tions of the ACM 39(5) (1996), 46–53.                                        ing: Creating organic 3d shapes with the hand and tangible
                                                                               tools. In Proc. of CHI 2001 (2001), 261–268.
[HQ01] H UA J., Q IN H.: Haptic sculpting of volumet-
  ric implicit functions. In Proc. of the 9th Pacific Con-                    [VdPRHP99] VAN DE P OL R., R IBARSKY W., H ODGES
  ference on Computer Graphics and Applications (2001),                        L., P OST F.: Interaction in semi-immersive large display
  254–264.                                                                     environments. In Proc. of EGVE ’99 (1999), 157–168.
[KBS00] KOBBELT L. P., BAREUTHER T., S EIDEL H.-                             [VRCW04] VANDOREN P., R AYMAEKERS C., C ONINX
  P.: Multiresolution shape deformations for meshes with                       K., W EYER T.: Two-handed interaction inspired by tan-
  dynamic vertex connectivity. Computer Graphics Forum                         gible interfaces in 3d modelling. In Proc. of 15th IASTED
  19, 3 (Aug. 2000), 249–260.                                                  Modelling and Simulation Conference (2004), 197–202.
[KF04] K IM H., F ELLNER D.: Interaction with hand ges-
  ture for a back-projection wall. In Proc. of Computer
  Graphics International (2004).
[KHK∗ 96] K IYOKAWA K., H ARUO T., K ATAYAMA Y.,
  I WASA H., YOKOYA N.: Vlego: A simple two-handed
  modeling environment based on toy blocks. In Proc. of
  ACM Simpo. on Virtual Reality Software and Technology
  (VRST’96) (1996), 27–34.
[Kni98] K NIGHT C.: Virtual reality for visualisation.
  Technical Report 13/98, Department of Computer Sci-
  ence, University of Durham (1998).
[KSW00] K AUFMANN H., S CHMALSTIEG D., WAGNER
  M.: Construct3d: A virtual reality application for mathe-
  matics and geometry education. Education and Informa-
  tion Technologies 5(4) (2000), 263–276.
[LKG∗ 03] L LAMAS I., K IM B., G ARGUS J.,
  ROSSIGNAC J., S HAW C. D.:         Twister: A space-
  warp operator for the two-handed editing of 3d shapes.

c The Eurographics Association 2005.

More Related Content

What's hot

Consumer Insights On Innovative 3D Visualization Technologies
Consumer Insights On Innovative 3D Visualization TechnologiesConsumer Insights On Innovative 3D Visualization Technologies
Consumer Insights On Innovative 3D Visualization TechnologiesDr. Alexander van Laack
 
DCT based Image Watermarking
DCT based Image WatermarkingDCT based Image Watermarking
DCT based Image WatermarkingNisarg Shah
 
iMinds insights - 3D Visualization Technologies
iMinds insights - 3D Visualization TechnologiesiMinds insights - 3D Visualization Technologies
iMinds insights - 3D Visualization TechnologiesiMindsinsights
 
Article : Wall grammar for building generation
Article : Wall grammar for building generationArticle : Wall grammar for building generation
Article : Wall grammar for building generationMathieu Larive
 
Developing and comparing an encoding system using vector quantization &
Developing and comparing an encoding system using vector quantization &Developing and comparing an encoding system using vector quantization &
Developing and comparing an encoding system using vector quantization &IAEME Publication
 
Michelangelo: A 3-DOF sketching robot
Michelangelo: A 3-DOF sketching robotMichelangelo: A 3-DOF sketching robot
Michelangelo: A 3-DOF sketching robotUttam Grandhi
 
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...Towards a Smart Control Room for Crisis Response Using Visual Perception of U...
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...Joris IJsselmuiden
 
FOCUS K3D AWG CAD/CAE
FOCUS K3D AWG CAD/CAEFOCUS K3D AWG CAD/CAE
FOCUS K3D AWG CAD/CAEFOCUS K3D
 
Traffic Automation System
Traffic Automation SystemTraffic Automation System
Traffic Automation SystemPrabal Chauhan
 
Computer Graphics 2004
Computer Graphics 2004Computer Graphics 2004
Computer Graphics 2004Sanjay Goel
 
Influence of local segmentation in the context of digital image processing
Influence of local segmentation in the context of digital image processingInfluence of local segmentation in the context of digital image processing
Influence of local segmentation in the context of digital image processingiaemedu
 
Automated License Plate Recognition for Toll Booth Application
Automated License Plate Recognition for Toll Booth ApplicationAutomated License Plate Recognition for Toll Booth Application
Automated License Plate Recognition for Toll Booth ApplicationIJERA Editor
 
Occupy Co-Working, by Julianne Malnick
Occupy Co-Working, by Julianne MalnickOccupy Co-Working, by Julianne Malnick
Occupy Co-Working, by Julianne Malnickjmalnick
 
COMUTER GRAPHICS NOTES
COMUTER GRAPHICS NOTESCOMUTER GRAPHICS NOTES
COMUTER GRAPHICS NOTESho58
 
IRJET - Cardless ATM
IRJET -  	  Cardless ATMIRJET -  	  Cardless ATM
IRJET - Cardless ATMIRJET Journal
 

What's hot (19)

Consumer Insights On Innovative 3D Visualization Technologies
Consumer Insights On Innovative 3D Visualization TechnologiesConsumer Insights On Innovative 3D Visualization Technologies
Consumer Insights On Innovative 3D Visualization Technologies
 
DCT based Image Watermarking
DCT based Image WatermarkingDCT based Image Watermarking
DCT based Image Watermarking
 
project_final
project_finalproject_final
project_final
 
iMinds insights - 3D Visualization Technologies
iMinds insights - 3D Visualization TechnologiesiMinds insights - 3D Visualization Technologies
iMinds insights - 3D Visualization Technologies
 
Article : Wall grammar for building generation
Article : Wall grammar for building generationArticle : Wall grammar for building generation
Article : Wall grammar for building generation
 
Developing and comparing an encoding system using vector quantization &
Developing and comparing an encoding system using vector quantization &Developing and comparing an encoding system using vector quantization &
Developing and comparing an encoding system using vector quantization &
 
Michelangelo: A 3-DOF sketching robot
Michelangelo: A 3-DOF sketching robotMichelangelo: A 3-DOF sketching robot
Michelangelo: A 3-DOF sketching robot
 
G0352039045
G0352039045G0352039045
G0352039045
 
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...Towards a Smart Control Room for Crisis Response Using Visual Perception of U...
Towards a Smart Control Room for Crisis Response Using Visual Perception of U...
 
Print screen
Print screenPrint screen
Print screen
 
FOCUS K3D AWG CAD/CAE
FOCUS K3D AWG CAD/CAEFOCUS K3D AWG CAD/CAE
FOCUS K3D AWG CAD/CAE
 
Traffic Automation System
Traffic Automation SystemTraffic Automation System
Traffic Automation System
 
Computer Graphics 2004
Computer Graphics 2004Computer Graphics 2004
Computer Graphics 2004
 
Influence of local segmentation in the context of digital image processing
Influence of local segmentation in the context of digital image processingInfluence of local segmentation in the context of digital image processing
Influence of local segmentation in the context of digital image processing
 
Automated License Plate Recognition for Toll Booth Application
Automated License Plate Recognition for Toll Booth ApplicationAutomated License Plate Recognition for Toll Booth Application
Automated License Plate Recognition for Toll Booth Application
 
Occupy Co-Working, by Julianne Malnick
Occupy Co-Working, by Julianne MalnickOccupy Co-Working, by Julianne Malnick
Occupy Co-Working, by Julianne Malnick
 
COMUTER GRAPHICS NOTES
COMUTER GRAPHICS NOTESCOMUTER GRAPHICS NOTES
COMUTER GRAPHICS NOTES
 
G1802033543
G1802033543G1802033543
G1802033543
 
IRJET - Cardless ATM
IRJET -  	  Cardless ATMIRJET -  	  Cardless ATM
IRJET - Cardless ATM
 

Viewers also liked

Mid-term Report
Mid-term ReportMid-term Report
Mid-term ReportJongHyoun
 
Ijipm jong hyounkim_ijipm1-070061ip
Ijipm jong hyounkim_ijipm1-070061ipIjipm jong hyounkim_ijipm1-070061ip
Ijipm jong hyounkim_ijipm1-070061ipJongHyoun
 
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...Optimal Performance Consultants Inc
 

Viewers also liked (6)

Mid-term Report
Mid-term ReportMid-term Report
Mid-term Report
 
r_v099
r_v099r_v099
r_v099
 
Ijipm jong hyounkim_ijipm1-070061ip
Ijipm jong hyounkim_ijipm1-070061ipIjipm jong hyounkim_ijipm1-070061ip
Ijipm jong hyounkim_ijipm1-070061ip
 
Etri_V08
Etri_V08Etri_V08
Etri_V08
 
Psr2010
Psr2010Psr2010
Psr2010
 
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...
The Accessibility for Ontarian's with Disabilities Act for Ontario; Status Up...
 

Similar to Tangible 3 D Hand Gesture

International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)CSCJournals
 
Detection of Virtual Passive Pointer
Detection of Virtual Passive PointerDetection of Virtual Passive Pointer
Detection of Virtual Passive PointerCSCJournals
 
3 dmobilenavigationdesign
3 dmobilenavigationdesign3 dmobilenavigationdesign
3 dmobilenavigationdesigntimemagazine90
 
Real time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsReal time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsijujournal
 
Real time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsReal time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsijujournal
 
Virtual Reality in Concept Design
Virtual Reality in Concept Design Virtual Reality in Concept Design
Virtual Reality in Concept Design IDES Editor
 
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSIONMOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSIONIRJET Journal
 
Bt9403, virtual reality
Bt9403, virtual realityBt9403, virtual reality
Bt9403, virtual realitysmumbahelp
 
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devices
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing DevicesFrom Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devices
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devicestoukaigi
 
Review by g siminon latest 2011
Review by g siminon latest 2011Review by g siminon latest 2011
Review by g siminon latest 2011ujjwal9191
 
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...Implementation of Interactive Augmented Reality in 3D Assembly Design Present...
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...AIRCC Publishing Corporation
 
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...ijcsit
 
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...AIRCC Publishing Corporation
 
Hand gesture recognition using support vector machine
Hand gesture recognition using support vector machineHand gesture recognition using support vector machine
Hand gesture recognition using support vector machinetheijes
 
A mobile agent based approach for data management to support 3 d emergency pr...
A mobile agent based approach for data management to support 3 d emergency pr...A mobile agent based approach for data management to support 3 d emergency pr...
A mobile agent based approach for data management to support 3 d emergency pr...Ijrdt Journal
 
Augmented Reality for Fire & Emergency Services
Augmented Reality for Fire & Emergency ServicesAugmented Reality for Fire & Emergency Services
Augmented Reality for Fire & Emergency Servicesidescitation
 

Similar to Tangible 3 D Hand Gesture (20)

International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)
 
Detection of Virtual Passive Pointer
Detection of Virtual Passive PointerDetection of Virtual Passive Pointer
Detection of Virtual Passive Pointer
 
3 dmobilenavigationdesign
3 dmobilenavigationdesign3 dmobilenavigationdesign
3 dmobilenavigationdesign
 
White Paper UMI3D
White Paper UMI3DWhite Paper UMI3D
White Paper UMI3D
 
Real time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsReal time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applications
 
Real time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applicationsReal time hand gesture recognition system for dynamic applications
Real time hand gesture recognition system for dynamic applications
 
Virtual Reality in Concept Design
Virtual Reality in Concept Design Virtual Reality in Concept Design
Virtual Reality in Concept Design
 
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSIONMOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
 
G0342039042
G0342039042G0342039042
G0342039042
 
Bt9403, virtual reality
Bt9403, virtual realityBt9403, virtual reality
Bt9403, virtual reality
 
HGR-thesis
HGR-thesisHGR-thesis
HGR-thesis
 
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devices
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing DevicesFrom Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devices
From Sense to Print: Towards Automatic 3D Printing from 3D Sensing Devices
 
Review by g siminon latest 2011
Review by g siminon latest 2011Review by g siminon latest 2011
Review by g siminon latest 2011
 
1
11
1
 
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...Implementation of Interactive Augmented Reality in 3D Assembly Design Present...
Implementation of Interactive Augmented Reality in 3D Assembly Design Present...
 
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
 
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
IMPLEMENTATION OF INTERACTIVE AUGMENTED REALITY IN 3D ASSEMBLY DESIGN PRESENT...
 
Hand gesture recognition using support vector machine
Hand gesture recognition using support vector machineHand gesture recognition using support vector machine
Hand gesture recognition using support vector machine
 
A mobile agent based approach for data management to support 3 d emergency pr...
A mobile agent based approach for data management to support 3 d emergency pr...A mobile agent based approach for data management to support 3 d emergency pr...
A mobile agent based approach for data management to support 3 d emergency pr...
 
Augmented Reality for Fire & Emergency Services
Augmented Reality for Fire & Emergency ServicesAugmented Reality for Fire & Emergency Services
Augmented Reality for Fire & Emergency Services
 

More from JongHyoun

게임, 인간, 문화 그리고 사회관계 김종현교수
게임, 인간, 문화 그리고 사회관계 김종현교수게임, 인간, 문화 그리고 사회관계 김종현교수
게임, 인간, 문화 그리고 사회관계 김종현교수JongHyoun
 
Game Planning
Game PlanningGame Planning
Game PlanningJongHyoun
 
등장인물분석 가상캐스팅
등장인물분석 가상캐스팅등장인물분석 가상캐스팅
등장인물분석 가상캐스팅JongHyoun
 
TIME Project
TIME ProjectTIME Project
TIME ProjectJongHyoun
 
Tangible&Rendering
Tangible&RenderingTangible&Rendering
Tangible&RenderingJongHyoun
 
AR, the TODAY
AR, the TODAYAR, the TODAY
AR, the TODAYJongHyoun
 
Indoor Location Tracking
Indoor Location TrackingIndoor Location Tracking
Indoor Location TrackingJongHyoun
 
가상세계와클론
가상세계와클론가상세계와클론
가상세계와클론JongHyoun
 
국내게임엔진 1
국내게임엔진 1국내게임엔진 1
국내게임엔진 1JongHyoun
 
Robot Pet Society with Human-being
Robot Pet Society with Human-beingRobot Pet Society with Human-being
Robot Pet Society with Human-beingJongHyoun
 
감성공간을 위한 스토리 인지기술
감성공간을 위한 스토리 인지기술감성공간을 위한 스토리 인지기술
감성공간을 위한 스토리 인지기술JongHyoun
 
Tangible AR Interface
Tangible AR InterfaceTangible AR Interface
Tangible AR InterfaceJongHyoun
 

More from JongHyoun (20)

WG
WGWG
WG
 
Proposal
ProposalProposal
Proposal
 
게임, 인간, 문화 그리고 사회관계 김종현교수
게임, 인간, 문화 그리고 사회관계 김종현교수게임, 인간, 문화 그리고 사회관계 김종현교수
게임, 인간, 문화 그리고 사회관계 김종현교수
 
기획서 2
기획서 2기획서 2
기획서 2
 
2 by Dr.Ahn
2 by Dr.Ahn2 by Dr.Ahn
2 by Dr.Ahn
 
1 by Dr.Ahn
1 by Dr.Ahn1 by Dr.Ahn
1 by Dr.Ahn
 
Game Planning
Game PlanningGame Planning
Game Planning
 
3 by Dr.Ahn
3 by Dr.Ahn3 by Dr.Ahn
3 by Dr.Ahn
 
등장인물분석 가상캐스팅
등장인물분석 가상캐스팅등장인물분석 가상캐스팅
등장인물분석 가상캐스팅
 
TIME Project
TIME ProjectTIME Project
TIME Project
 
Tangible&Rendering
Tangible&RenderingTangible&Rendering
Tangible&Rendering
 
AR, the TODAY
AR, the TODAYAR, the TODAY
AR, the TODAY
 
Indoor Location Tracking
Indoor Location TrackingIndoor Location Tracking
Indoor Location Tracking
 
가상세계와클론
가상세계와클론가상세계와클론
가상세계와클론
 
국내게임엔진 1
국내게임엔진 1국내게임엔진 1
국내게임엔진 1
 
Tangible A
Tangible  ATangible  A
Tangible A
 
Robot Pet Society with Human-being
Robot Pet Society with Human-beingRobot Pet Society with Human-being
Robot Pet Society with Human-being
 
about OWI
about OWIabout OWI
about OWI
 
감성공간을 위한 스토리 인지기술
감성공간을 위한 스토리 인지기술감성공간을 위한 스토리 인지기술
감성공간을 위한 스토리 인지기술
 
Tangible AR Interface
Tangible AR InterfaceTangible AR Interface
Tangible AR Interface
 

Recently uploaded

Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxKatpro Technologies
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 

Recently uploaded (20)

Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 

Tangible 3 D Hand Gesture

  • 1. IPT & EGVE Workshop (2005) R. Blach, E. Kjems (Editors) Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling Hyosun Kim1,2 , Georgia Albuquerque1 , Sven Havemann1,2 , Dieter W. Fellner1,2 1 Institut für ComputerGraphik, TU Braunschweig, Germany 2 Institut für ComputerGraphik und WissensVisualisierung, TU Graz, Austria Abstract Most of all interaction tasks relevant for a general three-dimensional virtual environment can be supported by 6DOF control and grab/select input. Obviously a very efficient method is direct manipulation with bare hands, like in real environment. This paper shows the possibility to perform non-trivial tasks using only a few well-known hand gestures, so that almost no training is necessary to interact with 3D-softwares. Using this gesture interaction we have built an immersive 3D modeling system with 3D model representation based on a mesh library, which is optimized not only for real-time rendering but also accommodates for changes of both vertex positions and mesh connectivity in real-time. For performing the gesture interaction, the user’s hand is marked with just four fingertip- thimbles made of inexpensive material as simple as white paper. Within our scenario, the recognized hand gestures are used to select, create, manipulate and deform the meshes in a spontaneous and intuitive way. All modeling tasks are performed wirelessly through a camera/vision tracking method for the head and hand interaction. Categories and Subject Descriptors (according to ACM CCS): I.3.6 [Computer Graphics]: Interaction Techniques 1. Introduction Bare hands are the ultimate form of three-dimensional human-computer interface (3D-HCI). The vision guiding the work in this paper is the possibility to perform non-trivial tasks using only a few well-known gestures, so that almost no training is necessary to interact with 3D-software. Today, interfaces of this type still seem to be science-fiction, as they appear for instance in the recent Matrix and Minority Report. In both movies, computers are controlled by pointing with bare fingers at 2D interfaces floating in 3D. This has inspired us to look for what is possible with today’s technology to get as close as possible to such an intuitive an un-instrumented way for operating computers. A less far-fetched and more concrete usage scenario for such kind of technique are 3D shopping terminals, located for instance in shopping malls or in specialized stores. They may be thought of as a generalization of the familiar selling machines with a 2D interface, such as ATMs (for drawing cash money), or terminals for selling railway tickets. Antici- Figure 1: User interacts with a 3D virtual model by using pating the increasing use of 3D technology also in the com- non-instrumented hand gestures mercial market, easy to use 3D interfaces and applications are mandatory. c The Eurographics Association 2005.
  • 2. 192 H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling This leads to two fields of research that are tightly re- tesselated, i.e., converted to triangles, in order to display it. lated: First, to find the right kind of input device technology, But object representation that require costly pre-processing and second, to identify suitable interaction metaphors and before they can be displayed are ruled out when interactive devise general design guidelines for responsive 3D appli- update rates are required. cations. This paper presents our first prototype system that integrates both of these aspects, along with the the design Triangle meshes [KBS00] and also point clouds decisions and experiences we’ve made. [PKKG03] are routinely used for direct modeling of free- form surfaces. They permit arbitrary surface deformations Concerning the interface device, our choice was to use while maintaining the surface connected or even closed a vision-based tracking approach, based on a pair of cam- (“water-tight”). Both representations can also work in a eras with black-light sources. The system has been devel- multi-resolution fashion, a feature which is indispensable for oped and set up by ourselves, as described in a previous pa- adaptive real-time display. They have the drawback, how- per [KF04]. It is capable of tracking the fingertips of one ever, of being low-level representations: A distinction be- hand and also the position of the user’s head. It is used in a tween different parts of a 3D object on the level of indi- passive stereo back-projection environment, so that the user vidual points or triangles is difficult. This requires to main- can move freely in a space of approximately (2m)3 . Conse- tain a segmentation of the surface into different patches. A quently, the user has a stereo view on the scene, which fol- good compromise are therefore subdivision surfaces, which lows the head position to optimally adapt to the user’s point provide free-form manipulation, have the multi-resolution of view. A typical situation during 3D interaction is shown property, and are also compatible with polygonal meshes. in Figure 1. Incidentally there are some modeling systems which con- tain their own modeling primitives, and allow to create more This choice of an interface device was motivated by our complex objects by grouping together simpler ones. The targeted application scenario. Our main premise is to be as drawback is that they are specifically designed for the re- minimalistic as possible on the user requirements (see Figure spective applications [KHK∗ 96, HG96, SIS02]. 5 (c)): We consider a pair of black gloves with bright white marks on the fingertips plus a pair of polarized glasses with a The usefulness of interaction between user and virtual white frame as being quite acceptable, and maybe even mass models is the issue of VR based modelers. Debunne et al. market compatible, both in price and in practical handling. [DDCB01] and Hua et al. [HQ01] use a haptic feedback in- The stationary system setup is not prohibitively expensive terface for surface manipulation. Such devices provide the either: Basically a pair of projectors and a pair of cameras are user with the physical feeling of virtual objects, but they gen- sufficient. Both of these components are currently becoming erally restrict the available workspace to a desktop environ- less pricy through the pressure from the mass market. ment. VLEGO [KHK∗ 96] and THRED [SG94] use a pair Concerning our exemple application, our choice was of 6-DOF magnetic trackers (position + orientation data) the creation and manipulation of three-dimensional objects, with additional buttons to issue commands, such as to grab herein referred to as 3D modeling. While this may be an ob- or release the virtual object. These systems support com- vious choice for a genuine 3D interface, it has nevertheless plex modeling tasks by two-handed manipulation, but their turned out to be quite challenging: Controllability definitely input devices are unfortunately less intuitive than a glove- is an issue when working in free 3D, without a table to lean type input device [NUK98]. Schkolne et al. [SPP01] have on or to guide the user’s movements. Again the reason was introduced the Surface Drawing approach for 3D shape cre- to make no assumptions on the physical constraints of our ation, which uses a glove-type sensor together with a num- system, and to gain experiences with ’free 3D’. ber of tangible tools: The motion of the hand draws strokes in the 3D modeling workspace, and the thumb activates the creation of a surface. Other tangible interfaces, specifically 2. Related Research designed for their respective systems, were also found in [VRCW04, SIS02]. However, to our knowledge, bare hands Interactive 3D modeling is one of the areas where immer- have rarely appeared in the virtual modeling field, in spite sive VR can be expected to be most beneficial. There is of their high functionality and excellent naturalness, as was a trade-off, however, between ease of manipulation and noted by Pavlovic [PSH97]. rendering efficiency. One of the crucial ingredients is the choice of a suitable 3D object representation, which must Concerning display devices, head-mounted-display be suited to the target application. See [Kni98] for a dis- (HMD) can be found more often [KSW00, BO92, FS99] cussion of the implications related to the components of than others, such as a virtual workbench [SPP01, SIS02], or a VR system. Implicit surfaces for instance are an exam- a CAVE [DBW∗ 00]. This may be due to the high usability ple of a model representation with intuitive degrees of free- of the HMDs, which do not require fully equipped display dom [DBW∗ 00, HQ01, DTG96]. But they tend to have slow environments. But besides the high price, the drawback update rates when the model is interactively changed, due to of HMD devices is that only one person can follow the the fact that the graphics hardware requires a surface to be interaction. c The Eurographics Association 2005.
  • 3. H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling 193 (a) direct hand manipulation in real environment (b) hand gestures for 3D interactive modeling in virtual environment Figure 2: The five static hand gestures used in our system: Pause, point, grab, rotate, scaling. The gestures mimic characteristic human movements for object manipulation. They are mostly self-explanatory, except for the scaling gesture, which needs to be learned. 3. Hand Gesture based 3D Modeling users. So to keep things simple for the user, we have limited ourselves to a fixed set of five static gestures which are easy Focusing on interaction methods for virtual applications to memorize, shown in Figure 2. Another advantage is a re- in large physical space we find many instances of using duced processing load for gesture recognition, since in most computer-vision based techniques. The popularity of this ap- cases the identical configuration can be used over a number proach comes from the fact that the connection between the of frames. A typical scenario of using these gestures is the user and a computer system is wireless. The unrestrained following: movement encourages a liberal and intuitive interaction. The fact that the user can freely move on the other hand implies Interaction with virtual objects: Pinch the thumb and an that the gesture recognition must be very robust. The set of another finger after pointing to a target object. The se- meaningful gestures must therefore be carefully chosen so lected objects are manipulated in the same way with the that they can be both issued easily by the user, and robustly hand motions. detected by the system. • Grab(x, y, z): Close the hand while pinching the thumb and index finger, then move it. 3.1. Gesture Scenario • Rotate(x, y, z): Open the hand while pinching the An interaction technique is a method that allows the user to thumb and index finger, and then rotate it. perform some tasks in the system as a part of the user in- • Scale(x, y, z): Open the hand while pinching the thumb terface. Depending on the application, it supports different and middle finger, and then move it. modes of interaction, each of them can be issued by a dif- Interaction with a virtual cursor: Release the pinching ferent gesture. Since our chosen application field is shape states. This is to handle the cursor of application. modeling, the mapping from gestures to interaction modes should mimic the way how hands are used for sculpting and • Point(x, y, z): Close the hand except for the thumb and manipulation in the real world. Furthermore, it should sup- index finger, which points straight at the target object. port all necessary means for the basic modeling functions This orders a cursor to move to the position indicated through gestures. by the index finger. • Pause(x, y, z): Open the whole hand. Hand gestures used for 3D-HCI can be classified into two categories: dynamic gestures and static gestures [PSH97]. VR applications have to support four main categories of Using dynamic gestures generally requires a larger amount interaction tasks: navigation, selection, manipulation, sys- of time for training the interaction system, which should tem control [BKL∗ 00]. Table 1 shows the commands sup- also be personalized for each user to get the best recognition ported by our hand gestures. System control is achieved performance. Yet our system should also support occasional through a graphical 3D menu, which is operated in the same c The Eurographics Association 2005.
  • 4. 194 H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling (a) (b) (c) (d) Figure 3: Hand gesture interaction: (a) object manipulation and deformation by (b), (b) the sequence of gesture interaction – pause, point(object), rotate, translate (z-direction), point(object’s vertex), translate(object’s vertex), and rotate, (c) object creation from a simple primitive by (d), (d) the shape design drawn along the hand movement Selection (Point(x, y, z) or Pause(x, y, z)) 6DOF manipulation and selection, our gesture commands + (Grab(x, y, z) or Rotate(x, y, z)) include scaling function as well, which is useful especially Manipulation Translation Grab(x, y, z) for shape modeling. It is triggered by pinching the thumb or Rotation(x, y, z) and middle finger, while pinching the thumb and index fin- Rotation Rotate(x, y, z) ger goes into rotation. The use of different pinching combi- Uniform scaling Scaling(z) nations shows the possibility of extension of the gesture set if necessary. Table 1: Interaction tasks 3.1.1. Visual feedback To control graphical objects the user ’touches’ them with the way as object selection (see section 3.3), namely by point- interaction interfaces. Without the physical collision, how- ing. Navigation is done mainly via head tracking (see sec- ever, it is not easy to perceive the distance between the in- tion 4.1.3). terface and the targets. To reduce the user’s cognitive load, All gestures are associated with a 3D position, which can our modeler provides a number of different gizmos, proxy obviously be changed by moving the hand. So the rotate ges- objects which are occasionally set into the scene as visual ture can simultaneously be used to translate an object. One feedback clues. Gizmos are not part of the 3D model, but crucial point in designing interaction modes is to take into serve as a visual representative for a particular operation, to account the limited freedom of movement of the different indicate a specific mode, or to give information about the parts of the human body. Especially the hand can only per- status of the interaction. form limited rotations, so it is important to allow absolute The great advantage of gizmos is that they also improve as well as relative motions. Consequently, our system allows the robustness of the interaction tasks: Certain modeling op- for cumulative rotations by repeating rotate-and-pause ges- erations become much more controllable if they are not di- tures. rectly issued by gestures, but indirectly: The gizmo is ma- In addition to the basic 3D interaction functionalities of nipulated by gestures, and the gizmo in turn issues the mod- c The Eurographics Association 2005.
  • 5. H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling 195 eling operation, so that it is executed in a more controlled way. One simple example is the arrow-shaped model which is used as a 3D cursor to represent the user’s hand in 3D vir- tual space. The position of the arrow is determined by the projection of the user’s hand position in real space to the vir- tual 3D scene. So it is always in front of the scene, and it works like the familiar 2D mouse pointer on a 2D desktop. When there is a collision between the arrow cursor and a Figure 4: Face representation change by seting the sharp- target object, another gizmo like a control ball or an object ness flags of the edges (from sharp to smooth) bounding-box appears. For other tasks such as rotating or twisting objects, a tripod gizmo with three coordinate axes is drawn, which moves in correspondence to the user’s hand orientation. mesh greatly improved the controllability of surface manip- ulation: Distinct graphical feedback can be given to the user who is going to manipulate either a vertex, an edge, or a face 3.2. Modeling Operations of the control mesh. The mesh as intermediate structure also Figure 3(a-b) illustrates some deformations of a 3D object facilitates the selection of specific parts of the object (sur- together with the corresponding gestures used in the each face navigation), and has proven less tedious than a direct step. The respective gestures in the hand-motion sequences manipulation of the many triangles from the tesselation via also indicate the required modeling operations for the steps, gestures. as they correspond a direct manipulation tool for creating or Our model data are maintained in the Wavefront .obj modifying 3D shapes. file format, thus the input/output objects can be compatibly shared with other modeling products supporting this popular 3.2.1. Shape Representation format. The basis of the modeling operations is a mesh data struc- ture for interactive 3D design and manipulation, the Com- 3.2.2. Selection bined BRep (CBRep). It supports incremental on-line update The pointing approach [VdPRHP99] is used for target selec- and adaptive rendering through a multi-resolution tessela- tion. Note that only visible targets in the user’s field of view tion [HF04]. can be candidates for selection. But since our system sup- The special feature of Combined BReps is that they bridge ports head tracking (see section 4.1.3), it allows to change the gap between polygonal and free-form models. The mesh between different views in a natural way, and unselectable is an ordinary polygonal mesh by default, but any part of targets in one view may become selectable in another. This the mesh can also be used as the control mesh of a Cat- approach is advantageous because: mull/Clark subdivision surface. So a CBRep mesh can have - In most cases, an occluded object can be selected without both smooth free-form and polygonal parts, with crease actually removing the objects obstructing it. This possi- curves establishing a link between them. Both types of bility depends of course on the nature of the scene, and geometry are integrated in a single data structure by associ- works best with scenes that are not too ’opaque’. ating a boolean sharpness flag with each edge, to distinguish - The target must be searched only in 2D, not in 3D. This between sharp and smooth edges. In regions with only sharp enables the use of fast screen-space selection methods, edges, BRep faces are triangulated, just as with standard which accelerates the selection of very distant objects. polygonal meshes, whereas in smooth regions, the tessela- tion is created using recursive subdivision. By just chang- Smooth parts in our model are represented in Cat- ing the sharpness flag, it is possible to switch between both mull/Clark subdivision surfaces. The individual triangles of representations: Any face that contains at least one smooth the tesselation usually cover only a few pixels, so there are edge is treated as a Catmull/Clark surface (see Figure 4). obvious difficulties to accurately intersect a specific trian- Combined BReps are optimized not only for real-time ren- gle by a ray shot from the tracked head position through the dering, but they can also accommodate for changes of both tracked position of the pointing finger. The solution is to re- the vertex positions and the mesh connectivity in real-time, strict the selection to the vertices, edges, and faces of the by incrementally updating the changed parts of the tessela- control mesh, which are much easier to uniquely identify – tion. both for the user and for the system. The availability of CBReps was the basis of the plan to 3.2.3. Navigation realize a 3D modeling application which made use of our vi- sion based tracking system as an intuitive input device. One The size of the work space in front of the projection screen important lesson learned was that the existence of a control is limited by the viewing angle of the cameras used to track c The Eurographics Association 2005.
  • 6. 196 H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling (a) (b) (c) (d) Figure 5: System environment: (a) back-projection screen, (b) two projectors for passive stereo, (c) user interfaces – white paper-rimmed polarized glasses and fingertip markers (d) a black light the user’s motion. In particular, it is not possible to walk 3.3. Changing the modeling mode completely around a 3D object, so it must be possible for the The modeler needs to support various useful operations in user to navigate within the scene. In interaction mode, we terms of the system productivity. Using many kinds of ges- therefore provide the usual possibilities of translation and tures for each operation however can easily overloaded the zooming supported by head tracking (see section 4.1.3), as user with a large amount of the gesture learning. Providing well as the rotation around a particular target (’orbit’). In a graphical menu can be the suitable solution for changing addition to this, objects can be moved in the scene. Target the operation modes, instead of developing new gestures or selection is done as usual, by pointing. adding second interfaces. This is attributable to the fact that users in general already have rich experiences and skills in 3.2.4. Global vs. Local Deformations using menus and windows. The menu is placed on the boder By moving individual vertices and faces of the mesh, the of the scene, typically considered as the ’dead space’ of the user has fine-grained local control about the shape. But the screen. The user can choose the operation by using the hand shape is not only defined by the position of the vertices, but gestures, like selecting a model object. also by the connectivity of the mesh – as the case with all sur- face types based on subdivision. So to allow for controlled changes of mesh connectivity, we have added the Euler op- 4. System Setup erators to the set of manipulation modes. Specifically, the The hardware of our modeling environment consists of a user can passive stereo back-projection display with two projectors, - insert an edge to split a face in two, one PC to run the vision tracking software, and another PC - insert a new vertex on an edge, to run the modeling application. Both computers are con- - and split a vertex in two, connected by a new edge. nected via standard ethernet. It should equally be possible to smoothly deform larger parts of the mesh, which involves moving a number of 4.1. Camera/Vision tracker mesh vertices. We have therefore adopted the Twister ap- proach from Llamas et al. [LKG∗ 03] to use it with Com- The space in front of the back-projection screen (about 2.5m bined BReps. It works by applying a smooth space warp × 2m × 2m) is the work space for interaction. This volume to all vertices inside a spherical region of influence of ad- is typically of sufficient size for any kind of modeling ap- justable radius. So the user first selects a warping-region in- plication. The cameras are mounted near the upper edge of cluding several vertices on the control mesh of a model, and the screen, covering the interaction area but not limiting the moves it then to a new position, with the hand rotating in or- movement of the users. der to tilt, bend, or twist the shape. Since this can have drastic effects on the global shape, this method is quite sketchy and 4.1.1. Interaction environment setup suitable for simulating the spatial deformation of a rubber- like material. It is less applicable though for obtaining pre- The projection display environment usually involves a dim cise results, and is not suitable for CAD-like construction light condition to achieve a high-contrast screen projection. tasks. On the other hand the screen emits a variable amount of c The Eurographics Association 2005.
  • 7. H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling 197 light, so it is not feasible for the vision-based tracking to as- frame the center of the pair of glasses is computed. The left sume a static lighting condition. Variable lighting can deteri- and right eyes are assumed to be 3 cm to the left and to the orate the quality of the images captured by a camera, which right of the center of the glasses. We use linear polarizers in turn affects the image processing, and thus the results for for passive stereo, where the light for one eye is polarized tracking features in the images. To solve these problems we perpendicular to the light for the other eye. It means that the use a ’black-light’ [Rye96] source and white paper markers users must keep their head level or they will lose the stereo on the fingertips: The black light stimulates enough fluores- effect as the polarization of their stereo glasses no longer cent emission from the white paper markers to distinguish matches that of the projector filters. So only the position of the detection features even on the low-contrast camera im- the head is needed. These positions establish the two view- age. Since the black light is barely visible, it hardly influ- points for the stereo rendering. Figure 6 shows the typical ences the projection quality, and only the markers are high- scenes generated from the tracking of the head. lighted [KF04]. Using white paper as a marker material has obvious advantages: it is readily available, of low cost, and 4.2. Network Connection of light weight. Since the black light illuminates a white T- shirt in the same way it illuminates the white paper, this ap- The whole system can be roughly divided into two parts: proach however builds on the assumption that users do not the modeling system and the tracking system. They are con- wear white clothes or reflective accessories. nected via our standard ethernet network. The tracking sys- tem executes the computer-vision algorithms to monitor the 4.1.2. Interface design gestures of the user’s hand, and the position of the head. This information is sent to the second component, the modeling In order to let a user’s field of view cover the whole screen, system, for viewpoint maneuvering and object interaction. the user generally keeps the adequate distance to the screen. The transmission of the data from the tracking system to the Symmetrically, the cameras and the user’s hand are at the modeling system is done via TCP/IP packets containing the same distance, which may reduce the details of the tracking description of user motion. To maintain a nearly real-time features to a limited image resolution. So, if a hand is marked communication the send-receive procedure works in its own with too many markers, segmenting each marker will cause thread on both machines, so that it is independent from the problems. Therefore we limit the number of markers to only rendering of the scene. four. While a user performs an interaction, the postures and positions of the hand will be continuously changed and up- dated. Though the unrestricted motions are desirable for a 5. Experimental Result natural interaction, it often can cause some difficulties such We have developed the modeling system on an AMD Athlon that the markers turn away from the camera view. To com- XP 2200+ PC with 512 MB of RAM and a GForce4 TI4600 pensate this we experimented with different marker arrange- graphic card (128 MB). This platform provides real-time ments with respect to shape and position. When ring shaped updates (about 30 − 50 frame/sec) of stereo scene projec- markers are placed at finger joints, the distance between tion. The camera/vision tracker runs on an AMD Athlon XP markers was too short to distinguish them reliably. In mo- 2000+ PC with a DFG/BW-1a frame grabber board and two tion they also might get lost due to occlusion by the fingers SONY XC-55 progressive scan cameras, with 30 frame/sec themselves. As a consequence, our markers are placed on in average to track the user’s head and hand simultaneously. the fingertips (thumb, index, middle and little finger) which The rendering frame rate and the scene quality is domi- are less frequently occluded, compared to other parts of the nated by the model complexity and the tessellation update on hand. A marker is designed like a thimble to fix easily to the shape deformation. To generate smooth and natural anima- fingertips. This compact shape can allow comfortable and tion while modeling, the scene rendering always has higher natural motion according to our experience. frequency than the data transmission from the tracking sys- For the head tracking interface, the polarized glasses is tem. rimmed with white paper. Figure 5(c) shows the thimble- The power as a virtual modeling system can be evaluted shaped fingertip markers and the white paper-rimmed polar- by the performance of building 3D models through the given ized glasses. interaction interfaces, hand gestures in our case. It is closely related with and mainly depends on harmony between the 4.1.3. Head tracking interaction method and interactive model representation. Eye positions are needed in order to generate accurate stereo First, the gesture-based interaction could instantly lead to views according to the user’s perspective. However, because the understanding of each gesture from the users after a short the user wears stereo glasses, the actual eyes are not visi- explanation. They could easily make a set of gestures with ble and cannot be tracked. So we approximate the eye posi- their hand and invoke the reaction from the system. The case tions by the positions of the glasses on the head. First, we set of medium level of modeling work could be quickly real- up the polarized glasses as a second tracking target. In each ized as well. For more demanding operations such as ’shape c The Eurographics Association 2005.
  • 8. 198 H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling Figure 6: Viewpoint movement by head tracking design’, we noticed that it needs somewhat more training to ’workhorse’ for intuitive 3D manipulation in our group. But produce the intended results right away. gestures alone do not make for a complete system. Our inter- action interface builds upon three components that are used Figure 7 (a-e) show some results produced by one of the in conjunction: authors, to represent our model creation in gesture drawings. It was certainly effective for wide range of shape sketching a) hand gestures, and modeling. The time spent to get simle models like ques- b) gizmos as virtual controller objects, and tion mark, alphabet characters and flower did not exceed c) textual menus for mode changes. 1-2 minutes, since they were simply drawn along the path of hand movement in space. The model in Figure 7 (c) is from Our experience so far has revealed that the real power of 3D croquis drawing made in a few minutes through similar ’tangible 3D’ is unleashed only by an intimate fusion of manner to draw on paper with a pencil. Not only the smooth all components. Users are enthusiastic about the ’hand-as- shapes, but also the polygonal forms could be achieved just a-tool’ metaphor, but the controllability and the precision of by changing the sharpness flag of edges in the control mesh. the interaction must still be improved. So our future plans Figure 7 (f) shows a short process of model generation, start- are to even better integrate the different methods: Examples ing from cubical primitives to more refined form. include using gestures for mode changes on a per-gizmo ba- sis, written annotations everywhere to clearly flag the current Once the source object was prepared, the other defor- mode states, numeric measurements producing ’tangible val- mation operators were used for refining or modifying the ues’ that can be put on a shelf, or used as input values for model. Figure 7 (a), (d) are the examples of a bit more com- other operations, and many more. plex shape creation followed by additional modeling oper- ations. The tree object in Figure 7 (g) is one of the exam- Our focus is, and will be, on building a generic infra- ple models with a complex geometry and a large number structure for direct, semi-immersive 3D interaction. We be- of small polygonal facets. The detail control for this kind lieve our system has distinguished advantages: It is quite af- of models, such as editing the end of each twig, was quite fordable, actually, and its performance will improve with in- difficult without any haptic feedback. In this case the ex- creasing computer, e.g., recoginition, power. The key com- tensive deformation operator like a simultaneous control of ponent for making the step from laboratory to field applica- multiple vertices was certainly useful. More advanced shape tion though is the software. editing were demonstrated in Figure 7 (h-j) shape tilting and Our goal is a software interface that permits to easily bending. As these experiments indicate, the modeling sys- transfer any given 3D application to such a tangible mod- tem could easily and quickly reproduce the shapes of the eling space, by wrapping its functionality using gestures, existing objects by the hand gestures. gizmos, and floating text. This will ultimately give creative The results obtained so far are quite promising. In the near people a new media at hand, to use it in fields such as arts, future, we will therefore perform systematic usability studies industrial design, and architecture. with larger groups of untrained users. We hope that an analy- sis of the user feed-back will reveal which of the proposed modeling methods works best for which kinds of application References scenario. [BKL∗ 00] B OWMAN D., K RUIJFF E., L AV IOLA J., P OUPYREV I., M INE M.: The art and science of 3d inter- 6. Conclusions and Future Work action. Tutorial Notes from the IEEE Virtual Reality 2000 Conf., IEEE Computer Society (2000). Hand gestures have proven a very expressive, convenient, and robust tool, and they have quickly become a popular [BO92] B UTTERWORT J. D. A. H. S., O LANO T.: 3dm: c The Eurographics Association 2005.
  • 9. H. Kim et al. / Tangible 3D: Hand Gesture Interaction for Immersive 3D Modeling 199 A three dimensional modeler using a head-mounted dis- ACM Transactions on Graphics 22, 3 (July 2003), play. Computer Graphics. In Proc. of Symposium on In- 663–668. teractive 3D Graphics 25(2) (1992), 135–138. [NUK98] N ISHINO H., U TSUMIYA K., KORIDA K.: 3d [DBW∗ 00] D EISINGER J., B LACH R., W ESCHE G., R. object modeling using spatial and pictographic gestures. B., S IMON A.: Towards immersive modeling - challenges In Proc. of the ACM symposium on Virtual reality software and recommendations: A workshop analyzing the needs and technology (1998), 51–58. of designers. In Proc. of the 6th Eurographics Workshop [PKKG03] PAULY M., K EISER R., KOBBELT L. P., on Virtual Environments (2000). G ROSS M.: Shape modeling with point-sampled geom- [DDCB01] D EBUNNE G., D ESBRUN M., C ANI M., etry. ACM Transactions on Graphics 22, 3 (July 2003), BARR A.: Dynamic real-time deformations using space 641–650. and time adaptive sampling. In Proc. of SIGGRAPH’01, [PSH97] PAVLOVIC V., S HARMA R., H UANG T.: Visual ACM Press (2001), 31–36. interpretation of hand gestures for human-computer inter- [DTG96] D ESBRUN M., T SINGOS N., G ASCUEL M.: action: A review. IEEE Transaction on Pattern Analysis Adaptive sampling of implicit surfaces for interactive and Machine Intelligence 19, 7 (1997), 677–695. modelling and animation. Computer Graphics Forum [Rye96] RYER A.: Light measurement handbook. 15(5) (1996), 319–327. http://www.intl-light.com/handbook/ (1996). [FS99] F UHRMANN A., S CHMALSTIEG D.: Concept and [SG94] S HAW C., G REEN M.: Two-handed polygonal implementation of a collaborative workspace for aug- surface design. In Proc. of ACM Symposium on User In- mented reality. GRAPHICS ’99 18(3) (1999). terface Software and Technology (1994), 205–212. [HF04] H AVEMANN S., F ELLNER D.: Progressive Com- bined BReps - Multi-Resolution Meshes for Incremental [SIS02] S CHKOLNE S., I SHII H., S CHROEDER P.: Tangi- Real-time Shape Manipulation. Tech. rep., Institute of ble + virtual = a flexible 3d interface for spatial construc- Computer Graphics, TU Braunschweig, 2004. tion applied to dna. CS TR 2002.011, California Institute of Technology (2002). [HG96] H ALLIDAY S., G REEN M.: A geometric model- ing and animation system for virtual reality. Communica- [SPP01] S CHKOLNE S., P RUETT M., P. S.: Surface draw- tions of the ACM 39(5) (1996), 46–53. ing: Creating organic 3d shapes with the hand and tangible tools. In Proc. of CHI 2001 (2001), 261–268. [HQ01] H UA J., Q IN H.: Haptic sculpting of volumet- ric implicit functions. In Proc. of the 9th Pacific Con- [VdPRHP99] VAN DE P OL R., R IBARSKY W., H ODGES ference on Computer Graphics and Applications (2001), L., P OST F.: Interaction in semi-immersive large display 254–264. environments. In Proc. of EGVE ’99 (1999), 157–168. [KBS00] KOBBELT L. P., BAREUTHER T., S EIDEL H.- [VRCW04] VANDOREN P., R AYMAEKERS C., C ONINX P.: Multiresolution shape deformations for meshes with K., W EYER T.: Two-handed interaction inspired by tan- dynamic vertex connectivity. Computer Graphics Forum gible interfaces in 3d modelling. In Proc. of 15th IASTED 19, 3 (Aug. 2000), 249–260. Modelling and Simulation Conference (2004), 197–202. [KF04] K IM H., F ELLNER D.: Interaction with hand ges- ture for a back-projection wall. In Proc. of Computer Graphics International (2004). [KHK∗ 96] K IYOKAWA K., H ARUO T., K ATAYAMA Y., I WASA H., YOKOYA N.: Vlego: A simple two-handed modeling environment based on toy blocks. In Proc. of ACM Simpo. on Virtual Reality Software and Technology (VRST’96) (1996), 27–34. [Kni98] K NIGHT C.: Virtual reality for visualisation. Technical Report 13/98, Department of Computer Sci- ence, University of Durham (1998). [KSW00] K AUFMANN H., S CHMALSTIEG D., WAGNER M.: Construct3d: A virtual reality application for mathe- matics and geometry education. Education and Informa- tion Technologies 5(4) (2000), 263–276. [LKG∗ 03] L LAMAS I., K IM B., G ARGUS J., ROSSIGNAC J., S HAW C. D.: Twister: A space- warp operator for the two-handed editing of 3d shapes. c The Eurographics Association 2005.