Demo Hour

XXIV.1 January - February 2017
Page: 8
Digital Citation


Authors:
Ippei Suzuki, Shuntarou Yoshimitsu, Keisuke Kawahara, Nobutaka Ito, Atushi Shinoda, Akira Ishii, Takatoshi Yoshida, Yoichi Ochiai, Wataru Yamada, Manabe Hiroyuki, Inrak Choi, Elliot Hawkes, David Christensen, Christopher Ploch, Sean Follmer, Oliver Glauser, Benedek Vartok, Wan-Chun Ma, Daniele Panozzo, Alec Jacobson, Otmar Hilliges, Olga Sorkine-Hornung

   

back to top  1. Gushed Diffusers: Fast-moving, Floating, and Lightweight Midair Display

Conventional aerial imaging systems are slow because they require a large, heavy setup. We use aerosol distribution from off-the-shelf spray as a fog screen that resists the wind and has high portability. As application examples, we present wearable applications and aerial imaging on objects with high-speed movements (e.g., a drone, a radio-controlled model car). Our study will contribute to the exploration of new application areas.

Suzuki, I., Yoshimitsu, S., Kawahara, K., Ito, N., Shinoda, A., Ishii, A., Yoshida, T., and Ochiai, Y. Gushed diffusers: Fast-moving, floating, and lightweight midair display. Adjunct Proc. of the 29th Annual ACM Symposium on User Interface Software & Technology. ACM, New York, 2016. DOI: http://dx.doi.org/10.1145/2984751.2985706

Suzuki, I., Yoshimitsu, S., Kawahara, K., Ito, N., Shinoda, A., Ishii, A., Yoshida, T., and Ochiai, Y. Gushed light field: Design method for aerosol-based fog display. Proc. of SIGGRAPH Asia 2016 Emerging Technologies. ACM, New York, 2016. DOI: http://dx.doi.org/10.1145/2988240.2988244

http://digitalnature.slis.tsukuba.ac.jp/2016/08/glf/

https://youtu.be/EPyN0O3RljU

Ippei Suzuki, University of Tsukuba,
1heisuzuki@gmail.com

Shuntarou Yoshimitsu, Waseda University

Keisuke Kawahara, University of Tsukuba

Nobutaka Ito, The University of Tokyo

Atushi Shinoda, University of Tsukuba

Akira Ishii, University of Tsukuba

Takatoshi Yoshida, The University of Tokyo

Yoichi Ochiai, University of Tsukuba

ins01.gif Floating screen with projection (showing a morpho butterfly) under the drone (DJI Phantom 2; DJI Co., Ltd.).
ins02.gif The minimum weight of the whole system is around 600g (when we use balsa wood for the frames).
     

back to top  2. Expanding the Field-of-View of Head-Mounted Displays with Peripheral Blurred Images

Field-of-view (FoV) is one of the key parameters of head-mounted displays (HMDs), because a wider FoV gives higher presence and immersion in a virtual environment. We propose a method that expands the FoV of HMDs by using two kinds of lenses with different levels of magnification. A central convex lens is surrounded by a Fresnel lens with high magnification that fills the peripheral vision with a blurred image. This method doesn’t need complicated optics, and is advantageous in terms of device cost and weight because only two additional Fresnel lenses are necessary.

Yamada, W. and Manabe, H. Expanding the field-of-view of head-mounted displays with peripheral blurred images. Adjunct Proc. of the 29th Annual ACM Symposium on User Interface Software & Technology. ACM, New York, 2016.

Wataru Yamada, NTT DOCOMO,
wataruyamada@acm.org

Manabe Hiroyuki, NTT DOCOMO,
manabehiroyuki@acm.org

ins03.gif Top: entire prototype. Bottom: (a) the prototype with convex lenses and Fresnel lenses (b) the prototype with only convex lenses.
     

back to top  3. Wolverine: A Wearable Haptic Interface for Grasping in Virtual Reality

The Wolverine is a mobile, wearable haptic device designed for simulating the grasping of rigid objects in virtual environments. We focused on creating a low-cost, lightweight device that renders a force directly between the thumb and three fingers to simulate objects held in pad-opposition-type grasps. Leveraging low-power brake-based locking sliders, the system can withstand over 100N of force between each finger and the thumb, and only consumes 2μWh for each braking interaction. Integrated sensors are used both for feedback control and user input: Time-of-flight sensors provide the distance of each finger and IMUs provide orientations.

Choi, I., Hawkes, E.W., Christensen, D.L., Ploch, C.J., and Follmer, S. Wolverine: A wearable haptic interface for grasping in virtual reality. Proc. of IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2016.

Choi, I. and Follmer, S. Wolverine: A wearable haptic interface for grasping in VR. Proc. of the 29th Annual ACM Symposium on User Interface Software & Technology. ACM, 2016.

http://shape.stanford.edu

https://youtu.be/kgE7JwlUuX4

Inrak Choi, Stanford University,
irchoi@stanford.edu

Elliot W. Hawkes, Stanford University,
ewhawkes@stanford.edu

David L. Christensen, Stanford University,
davidc10@stanford.edu

Christopher J. Ploch, Stanford University,
cploch@stanford.edu

Sean Follmer, Stanford University,
sfollmer@stanford.edu

ins04.gif Wolverine, a new wearable haptic user interface for grasping in virtual reality, holding a cylinder-shaped virtual object.
     

back to top  4. Rig Animation with a Tangible and Modular Input Device

This demo presents a novel approach to digital character animation, combining the benefits of modular and tangible input devices and sophisticated rig animation algorithms. We overcome limitations inherent in all previous tangible devices using a symbiotic software and hardware approach. It allows users to directly control complex rigs with only 5 to 10 physical modules. These compact input device configurations—optimized for a specific rig and a set of sample poses—are automatically generated by an algorithm. This avoids oversimplification of the pose space and excessively bulky devices.

Glauser, O., Ma, W.-C., Panozzo, D., Jacobson, A., Hilliges, O., and Sorkine-Hornung, O. Rig animation with a tangible and modular input device. ACM Trans. Graph. 35, 4 (2016); http://dl.acm.org/citation.cfm?id=2925909

http://igl.ethz.ch/projects/rig-animation-input-device/

https://www.youtube.com/watch?v=QW-goz8Cx8U

Oliver Glauser, ETH Zurich,
oliver.glauser@inf.ethz.ch

Benedek Vartok, ETH Zurich

Wan-Chun Ma, Activision, Inc.

Daniele Panozzo, New York University

Alec Jacobson, Columbia University and University of Toronto

Otmar Hilliges, ETH Zurich

Olga Sorkine-Hornung, ETH Zurich

ins05.gif Example posing session of a bunny character, with a device consisting of four modular parts.
     

back to top 

©2017 ACM  1072-5520/17/01  $15.00

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.

   

The Digital Library is published by the Association for Computing Machinery. Copyright © 2017 ACM, Inc.

Post Comment


No Comments Found