Date & Time : September 10 02:15 pm - 03:45 pm Location : HS1 Chair : Dieter Schmalstieg, TU Graz Papers :
AR-IVI – Implementation of In-Vehicle Augmented Reality
Authors: Qing Rao, Tobias Tropper, Christian Grünler, Markus Hammori, Samarjit Chakraborty
Abstract : In the last three years, a number of automotive Augmented Reality (AR)
concepts and demonstrators have been presented, all looking for an
interpretation of what AR in a car may look like. In October 2013,
Mercedes-Benz exhibited to a public audience the AR In-Vehicle Infotainment
(AR-IVI) system aimed at defining an overall in-vehicle electric/electronic
(E/E) architecture for augmented reality rather than showing specific use
cases. In this paper, we explain the requirements and design decisions that
lead to the system-design, and we share the challenges and experiences in
developing the AR-IVI system in the prototype vehicle. Based on our
experiences, we give an outlook on future software and E/E architectural
challenges of in-vehicle augmented reality.
Thermal Touch: Thermography-Enabled Everywhere Touch Interfaces for Mobile Augmented Reality Applications
Author: Daniel Kurz
Abstract : We present an approach that makes any real object a true touch interface for
mobile Augmented Reality applications. Using infrared thermography, we detect
residual heat resulting from a warm fingertip touching the colder surface of
an object. This approach can clearly distinguish if a surface has actually
been touched, or if a finger only approached it without any physical contact,
and hence significantly less heat transfer. Once a touch has been detected in
the thermal image, we determine the corresponding 3D position on the touched
object based on visual object tracking using a visible light camera. Finally
the 3D position of the touch is used by human machine interfaces for
Augmented Reality providing natural means to interact with real and virtual
objects. The emergence of wearable computers and head-mounted displays
desires for alternatives to a touch screen, which is the primary user
interface in handheld Augmented Reality applications. Voice control and
touchpads provide a useful alternative to interact with wearables for certain
tasks, but particularly common interaction tasks in Augmented Reality require
to accurately select or define 3D points on real surfaces. We propose to
enable this kind of interaction by simply touching the respective surface
with a fingertip. Based on tests with a variety of different materials and
different users, we show that our method enables intuitive interaction for
mobile Augmented Reality with most common objects.
AR-Mentor: Augmented Reality Based Mentoring System
Authors: Zhiwei Zhu, Vlad Branzoi, Michael Wolverton, Louise Yarnall, Girish Acharya, Supun Samarasekera, Rakesh Kumar, Glen Murray , Nicholas Vitovitch
Abstract : AR-Mentor is a wearable real time Augmented Reality (AR) mentoring system
that is configured to assist in maintenance and repair tasks of complex
machinery, such as vehicles, appliances, and industrial machinery. The system
combines a wearable Optical-See-Through (OST) display device with high
precision 6-Degree-Of-Freedom (DOF) pose tracking and a virtual personal
assistant (VPA) with natural language, verbal conversational interaction,
providing guidance to the user in the form of visual, audio and locational
cues. The system is designed to be heads-up and hands-free allowing the user
to freely move about the maintenance or training environment and receive
globally aligned and context aware visual and audio instructions (animations,
symbolic icons, text, multimedia content, speech). The user can interact with
the system, ask questions and get clarifications and specific guidance for
the task at hand. A pilot application with AR-Mentor was successfully
developed to instruct a novice to perform an advanced 33-step maintenance
task on a training vehicle. The initial live training tests demonstrate that
AR-Mentor is able to help and serve as an assistant to an instructor, freeing
him or her to cover more students and to focus on higher-order teaching.
Towards Augmented Reality User Interfaces in 3D Media Production
Authors: Max Krichenbauer, Goshiro Yamamoto, Takafumi Taketomi, Christian Sandor, Hirokazu Kato
Abstract : The idea of using Augmented Reality (AR) user interfaces (UIs) to create 3D
media content, such as 3D models for movies and games has been repeatedly
suggested over the last decade. Even though the concept is intuitively
compelling and recent technological advances have made such an application
increasingly feasible, very little progress has been made towards an actual
real-world application of AR in professional media production. To this day,
no immersive 3D UI has been commonly used by professionals for 3D computer
graphics (CG) content creation. In this paper, we are first to publish a
requirements analysis for our target application in the professional domain.
Based on a survey that we conducted with media professionals, the analysis of
professional 3D CG software, and professional training tutorials, we identify
these requirements and put them into the context of AR UIs. From these
findings, we derive several interaction design principles that aim to address
the challenges of real-world application of AR to the production pipeline. We
implemented these in our own prototype system while receiving feedback from
media professionals. The insights gained in the survey, requirements
analysis, and user interface design are relevant for research and development
aimed at creating production methods for 3D media production.