01
april 2012
AUGMENTED REALITY, ART AND TECHNOLOGY
INTRODUCING ADDED WORLDS Yolande Kolstee
THE TECHNOLOGY BEHIND AUGMENTED REALITY Pieter Jonker
RE-INTRODUCING MOSQUITOS Maarten Lamers
HOW DID WE DO IT Wim van Eck
2
AR[t] Magazine about Augmented Reality, art and technology
APRIL 2012
3
COLOPHON
ISSN NUMBER 2213-2481
CONTACT The Augmented Reality Lab (AR Lab) Royal Academy of Art, The Hague (Koninklijke Academie van Beeldende Kunsten)
Prinsessegracht 4 2514 AN The Hague The Netherlands +31 (0)70 3154795 www.arlab.nl info@arlab.nl
EDITORIAL TEAM Yolande Kolstee, Hanna Schraffenberger, Esmé Vahrmeijer (graphic design) and Jouke Verlinden.
CONTRIBUTORS Wim van Eck, Jeroen van Erp, Pieter Jonker, Maarten Lamers, Stephan Lukosch, Ferenc Molnár (photography) and Robert Prevel.
COVER ‘George’, an augmented reality headset designed by Niels Mulder during his Post Graduate Course Industrial Design (KABK), 2008
www.arlab.nl
4
TABLE OF CONTENTS
32 WELCOME to AR[t]
36 07
Yolande Kolstee
08
INTERVIEW WITH HELEN PAPAGIANNIS
12
INTRODUCING ADDED WORLDS
ARTIST IN RESIDENCE PORTRAIT: MARINA DE HAAS Hanna Schraffenberger
A MAGICAL LEVERAGE — IN SEARCH OF THE KILLER APPLICATION.
Pieter Jonker
RE-INTRODUCING MOSQUITOS Maarten Lamers
LIEVEN VAN VELTHOVEN — THE RACING STAR
20 28 30
HOW DID WE DO IT Wim van Eck
PIXELS WANT TO BE FREED! INTRODUCING AUGMENTED REALITY ENABLING HARDWARE TECHNOLOGIES
THE POSITIONING OF VIRTUAL OBJECTS
70
Robert Prevel
MEDIATED REALITY FOR CRIME SCENE INVESTIGATION
72
Stephan Lukosch
DIE WALKÜRE Wim van Eck, AR Lab Student Project
Hanna Schraffenberger
66
Jeroen van Erp
Hanna Schraffenberger
THE TECHNOLOGY BEHIND AR
60
76
36 42
Jouke Verlinden 5
6
WELCOME...
to the first issue of AR[t], the magazine about Augmented Reality, art and technology! Starting with this issue, AR[t] is an aspiring
their designs to clients. Designers of games and
magazine series for the emerging AR commu-
theme parks want to create immersive experi-
nity inside and outside the Netherlands. The
ences that integrate both the physical and the
magzine is run by a small and dedicated team
virtual world. Marketing specialists are working
of researchers, artists and lecturers of the AR
with new interactive forms of communication.
Lab (based at the Royal Academy of Arts, The
For all of them, AR can serve as a powerful tool
Hague), Delft University of Technology (TU
to realize their visions.
Delft), Leiden University and SME. In AR[t], we
Media artists and designers who want to acquire
share our interest in Augmented Reality (AR),
an interesting position within the domain of new
discuss its applications in the arts and provide
media have to gain knowledge about and experi-
insight into the underlying technology.
ence with AR. This magazine series is intended to
At the AR Lab, we aim to understand, develop,
provide both theoretical knowledge as well as a
refine and improve the amalgamation of the
guide towards first practical experiences with AR.
physical world with the virtual. We do this
Our special focus lies on the diversity of contri-
through a project-based approach and with the
butions. Consequently, everybody who wants
help of research funding from RAAK-Pro. In the
to know more about AR should be able to find
magazine series, we invite writers from the in-
something of interest in this magzine, be they
dustry, interview artists working with Augment-
art and design students, students from techni-
ed Reality and discuss the latest technological
cal backgrounds as well as engineers, develop-
developments.
ers, inventors, philosophers or readers who just happened to hear about AR and got curious.
It is our belief that AR and its associated
We hope you enjoy the first issue and invite you
technologies are important to the field of new
to check out the website www.arlab.nl to learn
media: media artists experiment with the inter-
more about Augmented Reality in the arts and
section of the physical and the virtual and probe
the work of the AR Lab.
the limits of our sensory perception in order to create new experiences. Managers of cultural heritage are seeking after new possibilities for worldwide access to their collections. Designers, developers, architects and urban planners are looking for new ways to better communicate
www.arlab.nl
7
INTRODUCING ADDED WORLDS: AUGMENTED REALITY IS HERE! By Yolande Kolstee
Augmented Reality is a relatively recent computer based technology that differs from the earlier known concept of Virtual Reality. Virtual Reality is a computer based reality where the actual, outer world is not directly part of, whereas Augmented Reality can be characterized by a combination of the real and the virtual. Augmented Reality is part of the broader concept of Mixed Reality: environments that consist of the real and the virtual. To make these differences and relations more clear, industrial engineer Paul Milgram and Fumio Kishino introduced the Mixed Reality Continuum diagram in 1994, in which the real world is placed on the one end and the virtual world is placed on the other end. MIXED REALITY (MR) Real Environment
Augmented Reality (AR)
Augmented Virtuality (AV)
Virtuality continuum by Paul Milgram and Fumio Kishino (1994)
8
Virtual Environment
A SHORT OVERVIEW OF AR
Paul C. Lauterbur and Peter Mansfield, who won
We define Augmented Reality as integrating 3-D
ing magnetic resonance imaging (MRI). Although
virtual objects or scenes into a 3-D environment
their original goals were different, in the field of
in real time (cf. A zuma, 1997).
Augmented Reality one might use the 3D virtual
the prize in 2003 for their discoveries concern-
models that are produced by such systems. However, they have to be processed prior to use in
WHERE 3D VIRTUAL OBJECTS OR SCENES COME FROM
AR because they might be too heavy. A 3D laser scanner is a device that analyses a real-world object or environment to collect data on its shape and its appearance (i.e. colour). The collected
What is shown in the virtual world, is created
data can then be used to construct digital, three
first. There are three ways of creating virtual
dimensional models. These scanners are some-
objects:
times called 3D digitizers. The difference is that the above medical scanners are looking inside to
1. By hand: using 3D computer graphics
create a 3D model while the laser scanners are
Designers create 3D drawings of objects, game
creating a virtual image from the reflection of
developers create 3D drawings of (human) figures,
the outside of an object.
(urban) architects create 3D drawings of buildings and cities. This 3D modeling by (product)
3. Photo and/or film images
designers, architects, and visual artists is done
It is possible to use a (moving) 2D image like a
by using specific software. Numerous software
picture as a skin on a virtual 3D model. In this
programs are developed. While some software
way the 2D model gives a three-dimensional
packages can be downloaded for free, others
impression.
are pretty expensive. Well known examples are Maya, Cinema 4D, Studio Max, Blender, Sketch up, Rhinoceros, Solidworks, Revit, Zbrush, AutoCad, Autodesk. By now at least 170 different software programs are available.
INTEGRATING 3-D VIRTUAL OBJECTS IN THE REAL WORLD IN REAL TIME
2. B y computer controlled imaging equipment/3D scanners.
There are different ways of integrating the vir-
We can distinguish different types of three-
tual objects or scenes into the real world. For all
dimensional scanners – the ones used in the
three we need a display possibility. This might
bio-medical world and the ones used for other
be a screen or monitor, small screens in AR
purposes – although there is some overlapping.
glasses, or an object on which the 3D images are
Inspecting a piece of medieval art or inspecting a
projected. We distinguish three types of (visual)
living human being is different but somehow also
Augmented Reality:
alike. In recent years we see a vigorous expansion of the use of image-producing bio-medical
Display type I : Screen based
equipment. We owe these developments to the
AR on a monitor, for example on a flatscreen or
of engineer sir Godfrey Hounsfield and physi-
on a smart phone (using e.g. LAYAR). With this
cist Allan Cormack, among others, who were
technology we see the real world and added at
jointly awarded the Nobel Prize in 1979 for their
the same time on a computer screen, monitor,
pioneering work on X-ray computed tomography
smartphone or tablet computer, the virtual
(CT). Another couple of Nobel Prize winners are
object. In that way, we can, for example,
9
Artist: KAROLINA SOBECKA | http://www.gravitytrap.com
add information to a book, by looking at the book and the screen at the same time.
Display type III: Projection based Augmented Reality With projection based AR we project virtual 3D
Display type II: AR glasses (off-screen)
scenes or objects on a surface of a building of an
A far more sophisticated but not yet consumer
exactly the dimensions of the object we project
friendly method uses AR glasses or a head
AR info onto. The projection is seen on the object
mounted display (HMD), also called a head-up
or building with remarkable precision. This can
display. With this device the extra information is
generate very sophisticated or wild projections
object (or a person). To do this, we need to know
mixed with one’s own perception of the world.
on buildings. The Augmented Matter in Context
The virtual images appear in the air, in the real
group, led by Jouke Verlinden at the Faculty of
world, around you, and are not projected on a
Industrial Design Engineering, TU-Delft, uses
screen. In type II there are two types of mixing
projection-based AR for manipulating the appear-
the real world with the virtual world:
ance of products.
Video see-through: a camera captures the real world. The virtual images are mixed with the captures (video images) of the real world and this mix creates an Augmented Reality. Optical see-through: the real world is perceived
CONNECTING ART AND TECHNOLOGY
directly with one’s own eyes in real time. Via small translucent mirrors in goggles, virtual
The 2011 IEEE International Symposium on Mixed
images are displayed on top of the perceived
and Augmented Reality (ISMAR) was held in
Reality.
Basel, Switzerland. In the track Arts, Media, and
10
Humanities, 40 articles were offered discussing
society.” (p.73., cited in Papagiannis, 2011, p.61)
the connection of ‘hard’ physics and ‘soft’ art.
As Helen Papagiannis concludes, it is then up to
There are several ways in which art and Aug-
the artist “to act as a pioneer, pushing forward
mented Reality technology can be connected:
a new aesthetic that exploits the unique materi-
we can, for example, make art with Augmented
als of the novel technology” (2011, p.61). Like
Reality technology, create Augmented Reality
Helen, we believe this holds also for the emerging
artworks or use Augmented Reality technology
field of AR technologies and we hope, artists will
to show and explain existing art (such as a
set out to create exciting new Augmented Real-
monument like the Greek Pantheon or paintings
ity art and thereby contribute to the interplay
from the grottos of Lascaux). Most of the contri-
between art and technology. An interview with
butions of the conference concerned Augmented
Helen Papagiannis can be found on page 12 of this
Reality as a tool to present, explain or augment
magazine. A portrait of the artist Marina de Haas,
existing art. However, some visual artists use AR
who did a residency at the AR Lab, can be found
as a medium to create art.
on page 60.
The role of the artist in working with the emerging technology of Augmented Reality has been
REFERENCES
discussed by Helen Papagiannis in her ISMAR
■ Milgram P. and Kishino, F., “A Taxonomy of
paper The Role of the Artist in Evolving AR as a
Mixed Reality Visual Displays,” IEICE Trans.
New Medium (2011). In her paper, Helen Papagi-
Information Systems, vol. E77-D, no. 12, 1994,
annis reviews how the use of technology as a creative medium has been discussed in recent years.
pp. 1321-1329. ■ A zuma, Ronald T., “A Survey of Augmented
She points out, that in 1988 John Pearson wrote
Reality”. In Presence: Teleoperators and
about how the computer offers artists “new
Virtual Environments 6, 4 (August 1997),
means for expressing their ideas” (p.73., cited in Papagiannis, 2011, p.61). According to Pearson,
pp. 355-385. ■ Papagiannis, H., “The Role of the Art-
“Technology has always been, the handmaiden of
ist in Evolving AR as a New Medium”, 2011
the visual arts, as is obvious, a technical means is
IEEE International Symposium on Mixed and
always necessary for the visual communication of
Augmented Reality(ISMAR) – Arts, Media, and
ideas, of expression or the development of works
Humanities (ISMAR-AMH), Basel, Switserland,
of art—tools and materials are required.” (p. 73)
pp. 61-65.
However, he points out that new technologies
■ Pearson, J., “The computer: Liberator or
“were not developed by the artistic community
Jailer of The creative Spirit.” Leonardo,
for artistic purposes, but by science and industry
Supplemental Issue, Electronic Art, 1 (1988),
to serve the pragmatic or utilitarian needs of
pp. 73-80.
11
BIOGRAPHY HELEN PAPAGIANNIS
Helen Papagiannis is a designer, artist, and PhD researcher specializing in Augmented Reality (AR) in Toronto, Canada. Helen has been working with AR since 2005, exploring the creative possibilities for AR with a focus on content development and storytelling. She is a Senior Research Associate at the Augmented Reality Lab at York University, in the Department of Film, Faculty of Fine Arts. Helen has presented her interactive artwork and research at global juried
conferences and events including TEDx (Technology, Entertainment, Design), ISMAR (International Society for Mixed and Augmented Reality) and ISEA (International Symposium for Electronic Art). Prior to her Augmented life, Helen was a member of the internationally renowned Bruce Mau Design studio where she was project lead on “Massive Change: The Future of Global Design." Read more about Helen’s work on her blog and follow her on Twitter: @ARstories.
www.augmentedstories.com
12
INTERVIEW WITH HELEN PAPAGIANNIS BY HANNA SCHRAFFENBERGER
What is Augmented Reality?
I wholeheartedly agree that AR can create a magical experience. In my TEDx 2010 talk, “How
Augmented Reality (AR) is a real-time layering of
Does Wonderment Guide the Creative Process”
virtual digital elements including text, images,
(http://youtu.be/ScLgtkVTHDc), I discuss how
video and 3D animations on top of our existing
AR enables a sense of wonder, allowing us to see
reality, made visible through AR enabled devices
our environments anew. I often feel like a magi-
such as smart phones or tablets equipped with
cian when presenting demos of my AR work live;
a camera. I often compare AR to cinema when
astonishment fills the eyes of the beholder ques-
it was first new, for we are at a similar moment
tioning, “How did you do that?” So what happens
in AR’s evolution where there are currently no
when the magic trick is revealed, as you ask,
conventions or set aesthetics; this is a time ripe
when the illusion loses its novelty and becomes
with possibilities for AR’s creative advancement.
habitual? In Virtual Art: Illusion to Immersion
Like cinema when it first emerged, AR has com-
(2004), new media art-historian Oliver Grau
menced with a focus on the technology with
discusses how audiences are first overwhelmed
little consideration to content. AR content needs
by new and unaccustomed visual experiences,
to catch up with AR technology. As a community
but later, once “habituation chips away at the
of designers, artists, researchers and commer-
illusion”, the new medium no longer possesses
cial industry, we need to advance content in AR
“the power to captivate” (p. 152). Grau writes
and not stop with the technology, but look at
that at this stage the medium becomes “stale
what unique stories and utility AR can present.
and the audience is hardened to its attempts at illusion”; however, he notes, that it is at this
So far, AR technologies are still new to many people and often AR works cause a magical experience. Do you think AR will lose its magic once people get used to the technology and have developed an understanding of how AR works? How have you worked with this ‘magical element’ in your work ‘The Amazing Cinemagician’?
stage that “the observers are receptive to content and media competence” (p. 152). When the initial wonder and novelty of the technology wear off, will it be then that AR is explored as a possible media format for various content and receive a wider public reception as a mass medium? Or is there an element of wonder that need exist in the technology for it to be effective and flourish?
13
Picture: PIPPIN LEE
“Pick a card. Place it here. Prepare to be amazed and entertained.”
14
I believe AR is currently entering the stage of
hidden within each physical playing card. Part
content development and storytelling, however,
of the magic and illusion of this project was to
I don’t feel AR has lost its “power to captivate”
disguise the RFID tag as a normal object, out
or “become stale”, and that as artists, design-
of the viewer’s sight. Each of these tags cor-
ers, researchers and storytellers, we continue to
responds to a short film clip by Méliès, which is
maintain wonderment in AR and allow it to guide
projected onto the FogScreen once a selected
and inspire story and content. Let’s not forget
card is placed atop the RFID tag reader. The
the enchantment and magic of the medium. I
RFID card reader is hidden within an antique
often reference the work of French filmmaker
wooden podium (adding to the aura of the magic
and magician George Méliès (1861-1938) as a
performance and historical time period).
great inspiration and recently named him the Patron Saint of AR in an article for The Creators
The following instructions were provided to the
Project (http://www.thecreatorsproject.com/
participant: “Pick a card. Place it here. Prepare
blog/celebrating-georges-méliès-patron-saint-
to be amazed and entertained.” Once the
of-augmented-reality) on what would have been
participant placed a selected card atop the des-
Méliès’ 150th birthday. Méliès was first a stage
ignated area on the podium (atop the concealed
magician before being introduced to cinema at
RFID reader), an image of the corresponding
a preview of the Lumiere brothers’ invention,
card was revealed on the FogScreen, which was
where he is said to have exclaimed, “That’s
then followed by one of Méliès’ films. The deci-
for me, what a great trick”. Méliès became
sion was made to provide visual feedback of the
famous for the “trick-film”, which employed a
participant’s selected card to add to the magic
stop-motion and substitution technique. Méliès
of the experience and to generate a sense of
applied the newfound medium of cinema to
wonder, similar to the witnessing and question-
extend magic into novel, seemingly impossible
ing of a magic trick, with participants asking,
visualities on the screen.
“How did you know that was my card? How did you do that?” This curiosity inspired further
I consider AR, too, to be very much about creat-
exploration of each of the cards (and in turn,
ing impossible visualities. We can think of AR as
Méliès’ films) to determine if each of the par
a real-time stop-substitution, which layers con-
ticipant’s cards could be properly identified.
tent dynamically atop the physical environment and creates virtual actualities with shapeshifting objects, magically appearing and disappearing— as Méliès first did in cinema. In tribute to Méliès, my Mixed Reality exhibit, The Amazing Cinemagician integrates Radio Frequency Identification (RFID) technology with the FogScreen, a translucent projection screen consisting of a thin curtain of dry fog. The
You are an artist and researcher. Your scientific work as well as your artistic work explores how AR can be used as a creative medium. What’s the difference between your work as an artist/ designer and your work as a researcher?
Amazing Cinemagician speaks to technology as magic, linking the emerging technology of the
Excellent question! I believe that artists and
FogScreenwith the pre-cinematic magic lantern
designers are researchers. They propose novel
and phantasmagoria spectacles of the Victorian
paths for innovation introducing detours into the
era. The installation is based on a card-trick,
usual processes. In my most recent TEDx 2011
using physical playing cards as an interface
talk in Dubai, “Augmented Reality and the Power
tointeract with the FogScreen. RFID tags are
of Imagination” (http://youtu.be/7QrB4cYxjmk),
15
I discuss how as a designer/artist/PhD researcher I am both a practitioner and a researcher, a maker and a believer. As a practitioner, I do, create, design; as a researcher I dream, aspire, hope. I am a make-believer working with a technology that is about make-believe, about imagining possibilities atop actualities. Now, more than ever, we need more creative adventurers and make-believers to help AR continue to evolve and become a wondrous new medium, unlike anything we’ve ever seen before! I spoke to the importance and power of imagination and makebelieve, and how they pertain to AR at this critical junction in the medium’s evolution. When we make-believe and when we imagine, we are in two places simultaneously; make-believe is about projecting or layering our imagination on top of a current situation or circumstance. In many ways, this is what AR is too: layering imagined worlds on top of our existing reality.
You’ve had quite a success with your AR pop-up book ‘Who’s Afraid of Bugs?’ In your blog you talk about your inspiration for the story behind the book: it was inspired by AR psychotherapy studies for the treatment of phobias such as arachnophobia. Can you tell us more? Who’s Afraid of Bugs? was the world’s first Augmented Reality (AR) Pop-up designed for iPad2 and iPhone 4. The book combines hand-crafted paper-engineering and AR on mobile devices to create a tactile and hands-on storybook that explores the fear of bugs through narrative and play. Integrating image tracking in the design, as opposed to black and white glyphs commonly
16
Picture: HELEN PAPAGIANNIS
seen in AR, the book can hence be enjoyed alone as a regular pop-up book, or supplemented with Augmented digital content when viewed through a mobile device equipped with a camera.The book is a playful exploration of fears using AR in a meaningful and fun way. Rhyming text takes
the reader through the storybook where various
Hallucinatory Augmented Reality (AR), 2007,
‘creepy crawlies’ (spider, ant, and butterfly) are
was an experiment which investigated the
awaiting to be discovered, appearing virtually
possibility of images which were not glyphs/AR
as 3D models you can interact with. A tarantula
trackables to generate AR imagery. The projects
attacks when you touch it, an ant hyperlinks to
evolved out of accidents, incidents in earlier
educational content with images and diagrams,
experimentsin which the AR software was mis
and a butterfly appears flapping its wings atop
taking non-marker imagery for AR glyphs and
a flower in a meadow. Hands are integrated
attempted to generate AR imagery. This confu-
throughout the book design, whether its placing
sion, by the software, resulted in unexpected
one’s hand down to have the tarantula crawl
and random flickering AR imagery. I decided to
over you virtually, the hand holding the magnify-
explore the creative and artistic possibilities
ing lens that sees the ant, or the hands that pop-
of this effect further and conduct experiments
up holding the flower upon which the butterfly
with non-traditional marker-based tracking.
appears. It’s a method to involve the reader in
The process entailed a study of what types of
the narrative, but also comments on the unique
non-marker images might generate such ‘hallu
tactility AR presents, bridging the digital with
cinations’ and a search for imagery that would
the physical. Further, the story for the AR
evoke or call upon multiple AR imagery/videos
Pop-up Book was inspired by AR psychotherapy
from a single image/non-marker.
studies for the treatment of phobias such as arachnophobia. AR provides a safe, controlled
Upon multiple image searches, one image
environment to conduct exposure therapy
emerged which proved to be quite extraordi-
within a patient’s physical surroundings, creat-
nary. A cathedral stained glass window was
ing a more believable scenario with heightened
able to evoke four different AR videos, the only
presence (defined as the sense of really being in
instance, from among many other images, in
an imagined or perceived place or scenario) and
which multiple AR imagery appeared. Upon close
provides greater immediacy than in Virtual Real-
examination of the image, focusing in and out
ity (VR). A video of the book may be watched at
with a web camera, a face began to emerge in
http://vimeo.com/25608606.
the black and white pattern. A fantastical image of a man was encountered. Interestingly, it
In your work, technology serves as an inspiration. For example, rather than starting with a story which is then adapted to a certain technology, you start out with AR technology, investigate its strengths and weaknesses and so the story evolves. However, this does not limit you to only use the strength of a medium. On the contrary, weaknesses such as accidents and glitches have for example influenced your work ‘Hallucinatory AR’. Can you tell us a bit more about this work?
was when the image was blurred into this face using the web camera that the AR hallucinatory imagery worked best, rapidly multiplying and appearing more prominently. Although numerous attempts were made with similar images, no other such instances occurred; this image appeared to be unique. The challenge now rested in the choice of what types of imagery to curate into this hallucinatory viewing: what imagery would be best suited to this phantasmagoric and dream-like form? My criteria for imagery/videos were like-form and shape, in an attempt to create a collage-like set of visuals. As the sequence or duration of the imagery in Hallucinatory AR could not be predetermined, the goal was to identify imagery
17
that possessed similarities, through which the
example of the technology failing. To the artist,
possibility for visual synchronicities existed.
however, there is poetry in these glitches, with
Themes of intrusions and chance encounters are
new possibilities of expression and new visual
at play in Hallucinatory AR, inspired in part by
forms emerging.
Surrealist artist Max Ernst. In What is the Mechanism of Collage? (1936), Ernst writes:
On the topic of glitches and accidents, I’d like to
One rainy day in 1919, finding myself on a village
return to Méliès. Méliès became famous for the
on the Rhine, I was struck by the obsession
stop trick, or double exposure special effect,
which held under my gaze the pages of an illus
a technique which evolved from an accident:
trated catalogue showing objects designed for
Méliès’ camera jammed while filming the streets
anthropologic, microscopic, psychologic, miner-
of Paris; upon playing back the film, he observed
alogic, and paleontologic demonstration. There
an omnibus transforming into a hearse. Rather
I found brought together elements of figuration
than discounting this as a technical failure, or
so remote that the sheer absurdity of that col
glitch, he utilized it as a technique in his films.
lection provoked a sudden intensification of
Hallucinatory AR also evolved from an accident,
thevisionary faculties in me and brought forth
which was embraced and applied in attempt
an illusive succession of contradictory images,
to evolve a potentially new visual mode in the
double, triple, and multiple images, piling up
medium of AR. Méliès introduced new formal
on each other with the persistence and rapidity
styles, conventions and techniques that were
which are particular to love memories and vi-
specific to the medium of film; novel styles and
sions of half-sleep (p. 427).
new conventions will also emerge from AR artists and creative adventurers who fully embrace
Of particular interest to my work in exploring
the medium.
and experimenting with Hallucinatory AR was Ernst’s description of an “illusive succession of contradictory images” that were “brought forth” (as though independent of the artist), rapidly multiplying and “piling up” in a state of “halfsleep”. Similarities can be drawn to the process of the seemingly disparate AR images jarringly coming in and out of view, layered atop one another. One wonders if these visual accidents are what the future of AR might hold: of unwelcome glitches in software systems as Bruce Sterling describes on Beyond the Beyond in 2009; or perhaps we might come to delight in the visual poetry of these Augmented hallucinations that are “As beautiful as the chance encounter of a sewing machine and an umbrella on an operating table.”
1
To a computer scientist, these ‘glitches’, as applied in Hallucinatory AR, could potentially be viewed or interpreted as a disaster, as an
18
[1] Comte de Lautreamont’s often quoted allegory, famous for inspiring both Max Ernst and Andrew Breton, qtd. in: Williams, Robert. “Art Theory: An Historical Introduction.” Malden, MA: Blackwell Publishing, 2004: 197
“As beautiful as the chance encounter of a sewing machine and an umbrella on an operating table.” Comte de Lautréamont
Picture: PIPPIN LEE 19
THE TECHNOLOGY BEHIND AUGMENTED REALITY Augmented Reality (AR) is a field that is primarily
the data can also be added to the vision of the
concerned with realistically adding computer-
user by means of a head-mounted display (HMD)
generated images to the image one perceives
or Head-Up Display. This is a second, less known
from the real world.
form of Augmented Reality. It is already known to fighter pilots, among others. We distinguish
AR comes in several flavors. Best known is the
two types of HMDs, namely: Optical See Through
practice of using flatscreens or projectors,
(OST) headsets and Video See Through (VST)
butÂÂnowadays AR can be experienced even on
headsets. OST headsets use semi-transparent
smartphones and tablet PCs. The crux is that 3D
mirrors or prisms, through which one can keep
digital data from another source is added to the
seeing the real world. At the same time, virtual
ordinary physical world, which is for example
objects can be added to this view using small
seen through a camera. We can create this ad-
displays that are placed on top of the prisms.
ditional data ourselves, e.g. using 3D drawing
VSTs are in essence Virtual Reality goggles, so
programs such as 3D Studio Max, but we can
the displays are placed directly in front of your
also add CT and MRI data or even live TV images
eyes. In order to see the real world, there are
to the real world. Likewise, animated three
two cameras attached on the other side of the
dimensional objects (avatars), which then can be
little displays. You can then see the Augmented
displayed in the real world, can be made using a
Reality by mixing the video signal coming from
visualization program like Cinema 4D. Instead of
the camera with the video signal containing the
displaying information on conventional monitors,
virtual objects.
20
21
UNDERLYING TECHNOLOGY
is computing power and energy consumption. Companies such as Microsoft, Google, Sony,
Screens and Glasses
Zeiss,... will enter the consumer market soon with AR technology.
Unlike screen-based AR, HMDs provide depth perception as both eyes receive an image. When objects are projected on a 2D screen,
Tracking Technology
one can convey an experience of depth by
A current obstacle for major applications which
letting the objects move. Recent 3D screens
soon will be resolved is the tracking technology.
allow you to view stationary objects in depth.
The problem with AR is embedding the virtual
3D televisions that work with glasses quickly
objects in the real world. You can compare
alternate the right and left image - in sync with
this with color printing: the colors, e.g., cyan,
this, the glasses use active shutters which let
magenta, yellow and black have to be printed
the image in turn reach the left or the right
properly aligned to each other. What you
eye. This happens so fast that it looks like you
often see in prints which are not cut yet, are
view both, the left and right image simultane-
so called fiducial markers on the edge of the
ously. 3D television displays that work without
printing plates that serve as a reference for
glasses make use of little lenses which are
the alignment of the colors. These are also
placed directly on the screen. Those refract
necessary in AR. Often, you see that mark-
the left and right image, so that each eye can
ers are used onto which a 3D virtual object is
only see the corresponding image. See for
projected. Moving and rotating the marker, lets
example www.dimenco.eu/display-technology.
you move and rotate the virtual object. Such
This is essentially the same method as used
a marker is comparable to the fiducial marker
on the well known 3D postcards on which a
in color printing. With the help of computer
beautiful lady winks when the card is slightly
vision technology, the camera of the headset
turned. 3D film makes use of two projectors
can identify the marker and based on it’s size,
that show the left and right images simultane-
shape and position, conclude the relative posi-
ously, however, each of them is polarized in a
tion of the camera. If you move your head rela-
different way. The left and right lenses of the
tive to the marker (with the virtual object), the
glasses have matching polarizations and only
computer knows how the image on the display
let through the light of to the corresponding
must be transformed so that the virtual object
projector. The important point with screens
remains stationary. And conversely, if your
is that you are always bound to the physical
head is stationary and you rotate the marker,
location of the display while headset based
it knows how the virtual object should rotate
techniques allow you to roam freely. This is
so that it remains on top of the marker.
called immersive visualization — you are immersed in a virtual world. You can walk around
AR smartphone applications such as Layar use
in the 3D world and move around and enter
the build in GPS and compass for the tracking.
virtual 3D objects.
This has an accuracy of meters and measures
Video-See-Through AR will become popular
angles of 5-10 degrees. Camera-based tracking,
within a very short time and ultimately be-
however, is accurate to the centimetre and can
come an extension of the smartphone. This is
measure angles of several degrees. Nowadays,
because both display technology and camera
using markers for the tracking is already out
technology have made great strides with the
of date and we use so called “natural feature
advent of smartphones. What currently still
tracking” also called “keypoint tracking”.
might stand in the way of smartphone models
Here, the computer searches for conspicuous
22
(salient) key points in the left and right camera image. If, for example, you twist your head, this shift is determined on the basis of those key points with more than 30 frames per second. This way, a 3D map of these keypoints can be built and the computer knows the relationship (distance and angle) between the keypoints and the stereo camera. This method is more robust than marker based tracking because you have many keypoints — widely spread in the scene — and not just the four corners of the marker close together in the scene. If someone walks in front of the camera and blocks some of the keypoints, there will still be enough keypoints left and the tracking is not lost. Moreover, you do not have to stick markers all over the world.
Collaboration with the Royal Academy of Arts (KABK) in The Hague in the AR Lab (Royal Academy, TU Delft, Leiden University, various SMEs) in the realization of applications. The TU Delft has done research on AR since 1999. Since 2006, the university works with the art academy in The Hague. The idea is that AR is a new technology with its own merits. Artists are very good at finding out what is possible with the new technology. Here are some pictures of realized projects. liseerde projecten
Fig 1. The current technology that replaces the markers with natural feature tracking or so called keypoint tracking. Instead of the four corners of the marker, the computer itself determines which points in the left and right images can be used as anchor points for calculating the 3D pose of the camera in 3D space. From top: 1: you can use all points in the left and right images to slowly build a complete 3D map. Such a map can, for example, be used to relive your past experience because you can again walk in the now virtual space. 2: the 3D keypoint space and the trace of the camera position within it. 3: keypoints (the color indicates the suitability) 4: you can place virtual objects (eyes) on an existing surface
23
Fig 2. Virtual furniture exhibition at the Salone di Mobile in Milan (2008); students of the Royal Academy of Art, The Hague show their furnitures by means of AR headsets. This saves transportation costs.
Fig 3. Virtual sculpture exhibition in KrĂśller-MĂźller (2009). From left: 1) visitors on adventure with laptops on walkers, 2) inside with a optical see-through headset, 3) large pivotable screen on a field of grass, 4) virtual image.
24
Fig 4. Exhibition in Museum Boijmans van Beuningen (2008-2009). From left: 1) Sgraffitto in 3D; 2) the 3D print version may be picked up by the spectator, 3) animated shards, the table covered in ancient pottery can be seen via the headset, 4) scanning antique pottery with the CT scanner delivers a 3D digital image.
Fig 5. The TUD, partially in collaboration with the Royal Academy (with the oldest industrial design course in the Netherlands), has designed a number of headsets.This design of headsets is an ongoing activity. From left: 1) first optical see-through headset with Sony headset and self-made inertia tracker (2000), 2) on a construction helmet (2006), 3) SmartCam and tracker taped on a Cyber Mind Visette headset (2007); 4) headset design with engines by Niels Mulder, a student at Royal Academy of Art, The Hague (2007), based on Cybermind technology, 5) low cost prototype based on the Carl Zeiss Cinemizer headset, 6) future AR Vizor?, 7) future AR lens? 25
There are many applications that can be realized using AR; they will find their way in the coming decades: 1. H ead-Up Displays have already been used for many years in the Air Force for fighter pilots; this can be extended to other vehicles and civil applications. 2. The billboards during the broadcast of a football game are essentially also AR; more can be done by also ivolving the game itself an allowing interaction of teh user, such as off-side line projection. 3. In the professional sphere, you can, for example, visualize where pipes under the street lie or should lie. Ditto for designing ships, houses, planes, trucks and cars. What’s outlined in a CAD drawing could be drawn in the real world, allowing you to see in 3D if and where there is a mismatch. 4. You can easily find books you are looking for in the library. 5. You can find out where restaurants are in a city... 6. Y ou can pimp theater / musical / opera / pop concerts with (immersive) AR decor. 7. You can arrange virtual furniture or curtains from the IKEA catalog and see how they look in your home. 8. Maintenance of complex devices will become easier, e.g. you can virtually see where the paper in the copier is jammed. 9. If you enter a restaurant or the hardware store, a virtual avatar can show you the place to find that special bolt or table.
26
SHOWING THE SERRA ROOM IN MUSEUM BOIJMANS VAN BEUNINGEN DURING THE EXHIBITION SGRAFFITO IN 3D
Picture: JOACHIM ROTTEVEEL
27
28
RE-INTRODUCING MOSQUITOS MAARTEN LAMERS AROUND 2004, MY YOUNGER BROTHER VALENTIJN INTRODUCED ME TO THE FASCINATING WORLD OF AUGMENTED REALITY. HE WAS A MOBILE PHONE SALESMAN AT THE TIME, AND SIEMENS HAD JUST LAUNCHED THEIR FIRST “SMARTPHONE”, THE BULKY SIEMENS SX1. THIS PHONE WAS QUITE MARVELOUS, WE THOUGHT – IT RAN THE SYMBIAN OPERATING SYSTEM, HAD A BUILT-IN CAMERA, AND CAME WITH… THREE GAMES.
One of these games was Mozzies, a.k.a Virtual
that “stuff” is. In Mozzies it was pesky little
Mosquito Hunt, which apparently won some
mosquitos -- nowadays it is anything from
2003 Best Mobile Game Award and my brother
restaurant information to crime scene data.
was eager to show it to me in the store where
But nothing really changed, right?
he worked at that time. I was immediately hooked… Mozzies lets you kill virtual mos-
Right! Technology became more advanced,
quitos that fly around superimposed over the
so we no longer need to hold the phone in
live camera feed. By physically moving the
our hand, but get to wear it strapped to our
phone you could chase after the mosquitos
skull in the form of goggles. But the idea is
when they attempted to fly off the phone’s
unchanged; you look at fake stuff in the real
display. Those are all the ingredients for
world and physically move around to deal
Augmented Reality in my personal opinion:
with it. You still don’t get the tactile sensa-
something that interacts with my perception
tion of swatting a mosquito or collecting
and manipulation of the world around me, at
“virtually heavy” information. You still don’t
that location, at that time. And Mozzies did
even hear the mosquito flying around you…
exactly that. Now almost eight years later, not much
It’s time to focus on those matters also, in my
has changed. Whenever people around me
opinion. Let’s take up the challenge and make
speak of AR, because they got tired of saying
AR more than visual, exploring interaction
“Augmented Reality”, they still refer to bulky
models for other senses. Let’s enjoy the full
equipment (even bulkier than the Siemens
experience of seeing, hearing, and particu-
SX1!) that projects stuff over a live camera
larly swatting mosquitos, but without the
feed and lets you interact with whatever
itchy bites.
29
LIEVEN VAN VELTHOVEN — THE RACING STAR
“IT AIN’T FUN IF IT AIN’T REAL TIME” BY HANNA SCHRAFFENBERGER 30
WHEN I ENTER LIEVEN VAN VELTHOVEN’S ROOM, THE PEOPLE FROM THE EFTELING HAVE JUST LEFT. THEY ARE INTERESTED IN HIS ‘VIRTUAL GROWTH’ INSTALLATION. AND THEY ARE NOT THE ONLY ONES INTERESTED IN LIEVEN’S WORK. IN THE LAST YEAR, HE HAS WON THE JURY AWARD FOR BEST NEW MEDIA PRODUCTION 2011 OF THE INTERNATIONAL CINEKID YOUTH MEDIA FESTIVAL AS WELL AS THE DUTCH GAME AWARD 2011 FOR THE BEST STUDENT GAME. THE WINNING MIXED REALITYGAME ‘ROOM RACERS’ HAS BEEN SHOWN AT THE DISCOVERY FESTIVAL,MEDIAMATIC, THE STRP FESTIVAL AND THE ZKM IN KARLSRUHE. HIS VIRTUAL GROWTH INSTALLATION HAS EMBELLISHED THE STREETS OF AMSTERDAM AT NIGHT. NOW, HE IS GOING TO SHOW ROOM RACERS TO ME, IN HIS LIVING ROOM — WHERE IT ALL STARTED. The room is packed with stuff and on first sight it seems rather chaotic, with a lot of random things laying on the floor. There are a few plants, which probably don’t get enough light, because Lieven likes the dark (that’s when his projections look best). It is only when he turns
Lieven tells me. He hands me a controller and soon we are racing the little projected cars around the chocolate spread, marbles, a remote control and a flash light. Trying not to crash the car into a belt, I tell him what I remember about when I first met him a few years ago at a Media Technology course at Leiden University. Back then, he was programming a virtual bird, which would fly from one room to another, preferring the room in which it was quiet. Loud and sudden sounds would scare the bird away into another room. The course for which he developed it was called sound space interaction, and his installation was solely based on sound. I ask him whether the virtual bird was his first contact with Augmented Reality. Lieven laughs.
“It’s interesting that you call it AR, as it only uses sound!” Indeed, most of Lieven’s work is based on interactive projections and plays with visual augmentations of our real environment. But like the bird, all of them are interactive and work in real-time. Looking back, the bird was not his first AR work.
“My first encounter with AR was during our first Media Technology course — a visit to the Ars Electroncia festival in 2007 — where I saw Pablo Valbuena’s Augmented Sculpture. It was amazing. I was asking myself, can I do something like this but interactive instead?”
on the beamer, that I realize that his room is actually not chaotic at all. The shoe, magnifying
Armed with a bachelor in technical computer
class, video games, tape and stapler which cover
science from TU Delft and the new found possi
the floor are all part of the game.
bility to bring in his own curiosity and ideas at the Media Technology Master program at Leiden
“You create your own race game tracks by placing real stuff on the floor”
University, he set out to build his own inter active projection based works.
31
ROOM RACERS Up to four players race their virtual cars around real objects which are lying on the floor. Players can drop in or out of the game at any time. Everything you can find can be placed on the floor to change the route. Room Racers makes use of projection-based mixed reality. The structure of the floor is analysed in real-time using a modified camera and self-written software. Virtual cars are projected onto the real environment and interact with the detected objects that are lying on the floor. The game has won the Jury Award for Best New Media Production 2011 of the international Cinekid Youth Media Festival, and the Dutch Game Award 2011 for Best Student Game. Room Racers shas been shown at several international media festivals. You can play Room Racers at the 'Car Culture' exposition at the Lentos Kunstmuseum in Linz, Austria until 4th of July 2012. Picture: LIEVEN VAN VELTHOVEN, ROOM RACERS AT ZKM | CENTER FOR ARTS AND MEDIA IN KARLSRUHE, GERMANY ON JUNE 19TH, 2011
32
33
“The first time, I experimented with the combination of the real and the virtual myself was in a piece called shadow creatures which I made with Lisa Dalhuijsen during our first semester in 2007.” More interactive projections followed in the next semester and in 2008, the idea for Room Racers was born. A first prototype was build in a week: a projected car bumping into real world things. After that followed months and months of optimizations. Everything is done by Lieven himself, mostly at night in front of the computer.
“My projects are never really finished, they are always work in progress, but if something works fine in my room, it’s time to take it out in the world.”
His success does surprise him and he especially did not expect the attention it gets in an art context.
“I knew it was fun. That became clear when I had friends over and we played with it all night. But I did not expect the awards. And I did not expect it to be relevant in the art scene. I do not think it’s art, it’s just a game. I don’t consider myself an artist. I am a developer and I like to do interactive projections. Room Racers is my least arty project, nevertheless it got a lot of response in the art context.” A piece which he actually considers more of an artwork is Virtual Growth: a mobile installation which projects autonomous growing structures
After having friends over and playing with the
onto any environment you place it in, be it
cars until six o’clock in the morning, Lieven
buildings, people or nature.
knows it’s time to steer the cars out of his room and show them to the outside world.
“I wanted to present Room Racers but I didn’t know anyone, and no one knew me. There was no network I was part of.” Uninhibited by this, Lieven took the initiative and asked the Discovery Festival if they were interested in his work. Luckily, they were — and showed two of his interactive games at the Discovery Festival 2010. After the festival requests started coming and the cars kept rolling. When I ask him about this continuing success he is divided:
“It’s fun, but it takes a lot of time — I have not been able to program as much as I used to.” 34
“For me AR has to take place in the real world. I don’t like screens. I want to get away from them. I have always been interested in other ways of interacting with computers, without mice, without screens. There is a lot of screen based AR, but for me AR is really about projecting into the real world. Put it in the real world, identify real world objects, do it in real-time, thats my philosophy. It ain’t fun if it ain’t real-time. One day, I want to go through a city with a van and do projections on buildings, trees, people and whatever else I pass.” For now, he is bound to a bike but that does not stop him. Virtual Growth works fast and
stable, even on a bike. That has been witnessed
at university. While talking, he smokes his
in Amsterdam, where the audiovisual bicycle
cigarette and takes the ashtray from the floor.
project ‘Volle Band’ put beamers on bikes and
With the road no longer blocked by it, the cars
invented Lieven to augmented the city with his
take a different route now. Lieven might take a
mobile installation. People who experienced
different route soon as well. I ask him, if he will
Virtual Growth on his journeys around Amster-
still be working from his living room, realizing
dam, at festivals and parties, are enthusiastic
his own ideas, once he has graduated.
about his (‘smashing!’) entertainment-art. As the virtual structure grows, the audience members not only start to interact with the piece but also with each other.
“They put themselves in front of the projector, have it projecting onto themselves and pass on the projection to other people by touching them. I don’t explain anything. I believe in simple ideas, not complicated concepts. The piece has to speak for itself. If people try it, immediately get it, enjoy it and tell other people about it, it works!”
“It’s actually funny. It all started to fill my portfolio in order to get a cool job. I wanted to have some things to show besides a diploma. That’s why I started realizing my ideas. It got out of control and soon I was realizing one idea after the other. And maybe, I’ll just continue doing it. But also, there are quite some companies and jobs I’d enjoy working for. First I have to graduate anyway.” If I have learned anything about Lieven and his work, I am sure his graduation project will be placed in the real world and work in in real-
Virtual Growth works, that becomes clear from
time. More than that, it will be fun. It ain’t
the many happy smiling faces the projection
Lieven, if it ain’t’ fun.
grows upon. And that’s also what counts for Lieven.
“At first it was hard, I didn’t get paid for doing these projects. But when people see them and are enthusiastic, that makes me happy. If I see people enjoying my work, and playing with it, that’s what really counts.”
Name:
Lieven van Velthoven
Born:
1984
Study: Media Technology MSc, Leiden University Background: Computer Science, TU Delft Selected AR Works: R oom Racers, Virtual Growth Watch: http://www.youtube.com/ user/lievenvv
I wonder where he gets the energy to work that much alongside being a student. He tells me, what drives him, is that he enjoys it. He likes to spend the evenings with the programming language C#. But the fact that he enjoys working on his ideas, does not only keep him motivated but also has caused him to postpone a few courses
35
HOW DID WE DO IT: ADDING VIRTUAL SCULPTURES AT THE KRÖLLER-MÜLLER MUSEUM By Wim van Eck ALWAYS WANTED TO CREATE YOUR OWN AUGMENTED REALITY PRO JECTS BUT NEVER KNEW HOW? DON’T WORRY, AR[T] IS GOING TO HELP YOU! HOWEVER, THERE ARE MANY HURDLES TO TAKE WHEN REALIZING AN AUGMENTED REALITY PROJECT. IDEALLY YOU SHOULD BE A SKILLFUL 3D ANIMATOR TO CREATE YOUR OWN VIRTUAL OBJECTS, AND A GREAT PROGRAMMER TO MAKE THE PROJECT TECHNICALLY WORK. PROVIDING YOU DON’T JUST WANT TO MAKE A FANCY TECH-DEMO, YOU ALSO NEED TO COME UP WITH A GREAT CONCEPT!
My name is Wim van Eck and I work at the AR
Blender (www.blender.org). These are all great
Lab, based at the Royal Academy of Art. One of
programs, however at the AR Lab we mostly use
my tasks is to help art-students realize their Aug-
Cinema 4d (image 1) since it is very user friendly
mented Reality projects. These students have
and because of that easier to learn. It is a shame
great concepts, but often lack experience in 3d
that the free Blender still has a steep learning
animation and programming. Logically I should
curve since it is otherwise an excellent program.
tell them to follow animation and programming
You can download a demo of Cinema 4d at
courses, but since the average deadline for their
http://www.maxon.net/downloads/demo-ver-
projects is counted in weeks instead of months
sion.html, these are some good tutorial sites to
or years there is seldom time for that... In the
get you started:
coming issues of AR[t] I will explain how the AR
http://www.cineversity.com
Lab helps students to realize their projects and
http://www.c4dcafe.com
how we try to overcome technical boundaries,
http://greyscalegorilla.com
showing actual projects we worked on by example. Since this is the first issue of our magazine I will give a short overview of recommendable programs for Augmented Reality development. We will start with 3d animation programs, which we need to create our 3d models. There are many 3d animation packages, the more well known ones include 3ds Max, Maya, Cinema 4d, Softimage, Lightwave, Modo and the open source
36
Image 1
Image 2
Image 3 | Picture by Klaas A. Mulder
Image 4
quid.com), for example, offers good quality but
Sweet summer nights at the Kröller-Müller Museum.
often at a high price, while free sites such as
As mentioned before in the introduction we
Artist-3d (http://artist-3d.com) have a more var-
will show the workflow of AR Lab projects with
ied quality. When a 3d model is not constructed
these ‘How did we do it’ articles. In 2009 the AR
properly it might give problems when you import
Lab was invited by the Kröller-Müller Museum to
it or visualize it. In coming issues of AR[t] we
present during the ‘Sweet Summer Nights’, an
will talk more about optimizing 3d models for
evening full of cultural activities in the famous
Augmented Reality usage. To actually add these
sculpture garden of the museum. We were asked
3d models to the real world you need Aug-
to develop an Augmented Reality installation
mented Reality software. Again there are many
aimed at the whole family and found a diverse
options, with new software being added continu-
group of students to work on the project. Now
ously. Probably the easiest to use software is
the most important part of the project started,
BuildAR (http://www.buildar.co.nz) which is
brainstorming!
In case you don’t want to create your own 3d models you can also download them from various websites. Turbosquid (http://www.turbos-
available for Windows and OSX. It is easy to import 3d models, video and sound and there is
Our location in the sculpture garden was in-
a demo available. There are excellent tutorials
between two sculptures, ‘Man and woman’, a
on their site to get you started. In case you want
stone sculpture of a couple by Eugène Dodeigne
to develop for iOS or Android the free Junaio
(image 2) and ‘Igloo di pietra’, a dome shaped
(http://www.junaio.com) is a good option. Their
sculpture by Mario Merz (image 3). We decided
online GLUE application is easy to use, though
to read more about these works, and learned
their preferred .m2d format for 3d models is
that Dodeigne had originally intended to create
not the most common. In my opinion the most
two couples instead of one, placed together in a
powerful Augmented Reality software right now
wild natural environment. We decided to virtu-
is Vuforia (https://developer.qualcomm.com/
ally add the second couple and also add a more
develop/mobile-technologies/Augmented-reality)
wild environment, just as Dodeigne initially had
in combination with the excellent game-engine
in mind. To be able to see these additions we
Unity (www.unity3d.com). This combination
placed a screen which can rotate 360 degrees
offers high-quality visuals with easy to script
between the two sculptures (image 4).
interaction on iOS and Android devices
37
Image 5
A webcam was placed on top of the screen,
actually build what the camera will see. This will
and a laptop running ARToolkit (http://www.
already save us quite some work. We can also
hitl.washington.edu/artoolkit) was mounted
see the screen is positioned quite far away from
on the back of the screen. A large marker was
the sculpture, and when an object is viewed
placed near the sculpture as a reference point
from a distance it will optically lose its depth.
for ARToolkit.
When you are one meter away from an object and take one step aside you will see the side of
Now it was time to create the 3d models of the
the object, but if the same object is a hundred
extra couple and environment. The students
meter away you will hardly see a change in per-
working on this part of the project didn’t have
spective when changing your position (see image
much experience with 3d animation, and there
6). From that distance people will hardly see the
wasn’t much time to teach them, so manually
difference between an actual 3d model and a
modeling the sculptures would be a difficult task.
plain 2d image. This means we could actually use
Soon options such as 3d scanning the sculpture
photographs or drawings instead of a complex 3d
were opted, but it still needs quite some skill
model, making the whole process easier again.
to actually prepare a 3d scan for Augmented
We decided to follow this route.
Reality usage. We will talk more about that in a coming issue of this magazine. But when we look carefully at our setup (image 5) we can draw some interesting conclusions. Our screen is immobile, we will always see our added 3d model from the same angle. So since we will never be able to see the back of the 3d model there is no need to actually model this part. This is a common practice while making 3d models, you can compare it with set construction for Hollywood movies where they also only Image 6 38
= Image 7
Image 8
Image 9
Image 10
Image 11 39
Original photograph by Klaas A. Mulder Image 12
To be able to place the photograph of the
of an image are visible, white is opaque, black
sculpture in our 3d scene we have to assign
is transparent. Detailed tutorials about alpha
it Âto a placeholder, a single polygon, image 7
channels are easily found on the internet.
shows how this could look.
As you can see this looks much better (image 9). We followed the same procedure for the second
This actually looks quite awful, we see the
statue and the grass (image 10), using many
statue but also all the white around it from the
separate polygons to create enough randomness
image. To solve this we need to make usage of
for the grass. As long as you see these models
something called an alpha channel, an option
from the right angle they look quite realistic
you can find in every 3d animation package
(image 11). In this case this 2.5d approach prob-
(image 8 shows where it is located in the mate-
ably gives even better results than a ‘normal’ 3d
rial editor of Cinema 4d). An alpha channel is
model, and it is much easier to create. Another
a grayscale image which declares which parts
advantage is that the 2.5d approach is very easy
40
The Lab collaborated in this project with students from different departments of the KABK: Ferenc Molnar, Mit Koevoets, Jing Foon Yu, Marcel Kerkmans and Alrik Stelling. The AR Lab team consisted of: Yolande Kolstee, Wim van Eck, Melissa Coleman en Pawel Pokutycki, supported by Martin Sjardijn and Joachim Rotteveel.
to compute since it uses few polygons, so you
We can conclude that it is good practice to
don’t need a very powerful computer to run it
analyze your scene before you start making your
or you can have many models on screen at the
3d models. You don’t always need to model all
same time. Image 12 shows the final setup.
the detail, and using photographs or drawings
For the iglo sculpture by Mario Merz we used
can be a very good alternative. The next issue
a similar approach. A graphic design student
of AR[t] will feature a new ‘How did we do it’, in
imagined what could be living inside the iglo,
case you have any questions you can contact me
and started drawing a variety of plants and
at w.vaneck@kabk.nl
creatures. Using the same 2.5d approach as described before we used these drawings and placed them around the iglo, and an animation was shown of a plant growing out of the iglo (image 12).
41
PIXELS WANT TO BE FREED! INTRODUCING AUGMENTED REALITY ENABLING HARDWARE TECHNOLOGIES
BY JOUKE VERLINDEN
42
1. Introduction
From the early head-up display in the movie “Robocop” to the present, Augmented Reality (AR) has evolved to a manageable ICT environment that must be considered by product designers of the 21st century. Instead of focusing on a variety of applications and software solutions, this article will discuss the essential hardware of Augmented Reality (AR): display techniques and tracking techniques. We argue that these two fields differentiate AR from regular human-user interfaces and tuning these is essential in realizing an AR experience. As often, there is a vast body of knowledge behind each of the principles discussed below, hence a large variety of literature references is given. Furthermore, the first author of this article found it important to elude his own preferences and experiences throughout this discussion. We hope that this material strikes a chord and makes you consider employing AR in your designs. After all, why should digital information always be confined to a dull, rectangular screen? 43
2. Display Technologies To categorise AR display technologies, two
3. Projector-based systems: one or more
important characteristics should be identified: imaging generation principle and physical
projectors cast digital imagery directly on the physical environment.
layout.
As Raskar and Bimber (2004, p.72) argued, an
Generic AR technology surveys describe a
important consideration in deploying an Aug-
large variety of display technologies that sup-
mented system is the physical layout of the
port imaging generation (Azuma, 1997; Azuma
image generation. For each imaging genera-
et al., 2001); these principles can be catego-
tion principle mentioned above, the imaging
rised into:
display can be arranged between user and physical object in three distinct ways:
1. Video-mixing. A camera is mounted somewhere on the product; computer graphics
a) head-attached, which presents digital
are combined with captured video frames
images directly in front of the viewer’s
in real time. The result is displayed on an
eyes, establishing a personal information
oblique surface, for example, an immer-
display.
sive Head-Mounted Display (HMD).
b) hand-held, carried by a user and does not
2. See-through: Augmentation by this
cover the whole field of view
principle typically employs half-silvered
c) spatial, which is fixed to the environment.
mirrors to superimpose computer graphics onto the user’s view, as found in head-up
The resulting imaging and arrangement combi-
displays of modern fighter jets.
nations are summarised in Table 1.
1.Video-mixing A. Head-attached
2. See-through
3. Projection-based
Head-mounted display (HMD)
B. Hand-held
Handheld devices
C. Spatial
Embedded display
See-through boards
Spatial projection-based
Table 1. Image generation principles for Augmented Reality
When the AR image generation and layout principles are combined, the following collection of display technologies are identified: HMD, Handheld devices, embedded screens, see-through boards and spatial projection-based AR. These are briefly discussed in the following sections.
44
2.1 Head-mounted display Head-attached systems refer to HMD solutions, which can employ either of the three image generation technologies. Even the first headmounted displays developed by virtue of the Virtual Reality already considered a see-through system with half-silvered mirrors to merge virtual line drawings with the physical environment (Sutherland, 1967). Since then, the variety of head-attached imaging systems has been expanded and encompasses all three principles for AR: video-mixing, see-through and direct projection on the physical world (Azuma et al., 2001). A benefit of this approach is its handsfree nature. Secondly, it offers personalised content, enabling each user to have a private view of the scene with customised and sensitive data that das not have to be shared. For most applications, HMDs have been considered inadequate, both in the case of see-through and video-mixing imaging. According to Klinker et al. (2002), HMDs introduce a large barrier between the user and the object and their resolution is insufficient for IAP — typically 800 × 600 pixels for the complete field of view (rendering the user “legally blind”by American standards). Similar reasoning was found in Bochenek et al. (2001), in which both the objective and subjective assessment of HMDs were less than those of hand-held or spatial imaging devices. However, new developments (specifically high-resolution OLED displays) show promising new devices, specifically for the professional market (Carl Zeiss) and enterntainment (Sony), see figure right.
Figure 1. RECENT HEAD MOUNTED DISPLAYS (ABOVE: KABK THE HAGUE AND UNDER: CARL ZEISS).
2.2 Handheld display Hand-held video-mixing solutions are based on
Augmented Reality technologies are emerging.
smartphones, PDAs or other mobile devices
By employing built-in cameras on smartphones
equipped with a screen and camera. With the
or PDAs, video mixing is enabled while concur-
advent of powerful mobile electronics, handheld
rent use is being supported by communication
45
GPS Antenna
Camera + IMU
Joystick Handles
UMPC
Figure 2. THE VESP´R DEVICE FOR UNDERGROUND INFRASTRUCTURE VISUALIZATION (SCHALL ET AL., 2008).
through wireless networks (Schmalstieg and
tems are found in each modern smartphone,
Wagner, 2008). The resulting device acts as a
and apps such as Layar (www.layar.com) and
hand-held window of a mixed reality. An exam-
Junaio (www.junaio.com) offer such functions
ple of such a solution is shown in Figure 2, which
for free to the user — allowing different layers of
is a combination of an Ultra Mobile Personal
content to the user (often social-media based).
Computer (UMPC), a Global Positioning System
‘such systems are found in each modern smartphone’
The advantage of using a video-mixing approach is that the lag times in processing are less influential than with the see-through or projector-based systems — the live video feed is also delayed and, thus, establishes a consistent combined image. This hand-held solution works well for occasional, mobile use. Long-term use can cause strain in the arms. The challenges in employing this principle are the limited screen coverage/resolution (typically with a 4-in diameter and a resolution of 320 × 240 pixels). Furthermore, memory, processing power and graphics processing is limited to ren-
(GPS) antenna for global position tracking, a
dering relatively simple 3D scenes, although these
camera for local position and orientation sensing
capabilities are rapidly improving by the upcom-
along with video mixing. As of today, such sys-
ing dual-core and quad-core mobile CPUs.
46
2.3 Embedded display Another AR display option is to include a number
with such screens. To our knowledge, no such
of small LCD screens in the observed object in
systems have been developed or commercialised
order to display the virtual elements directly on
so far. Although it does not support changing
the physical object. Although arguably an aug-
light effects, the Luminex material approximates
mentation solution, embedded screens do add
this by using an LED/fibreglass based fabric (see
digital information on product surfaces.
Figure 4). A Dutch company recently presented
This practice is found in the later stages of pro-
a fully interactive light-emitting fabric based on
totyping mobile phones and similar information
integrated RGB LEDs labelled ‘lumalive’. These
appliances. Such screens typically have a similar
initiatives can manifest as new ways to support
resolution as that of PDAs and mobile phones,
prototyping scenarios that require a high local
which is QVGA: 320 × 240 pixels. Such devices
resolution and complete unobstructedness. How-
are connected to a workstation by a specialised
ever, the fit to the underlying geometry remains
cable, which can be omitted if autonomously
a challenge, as well as embedding the associated
components are used, such as a smartphone.
control electronics/wiring. An elegant solution
Regular embedded screens can only be used on
to the second challenge was given by (Saakes et
planar surfaces and their size is limited while
al 2010) entitled “the slow display: by temporar-
their weight impedes larger use. With the ad-
ily changing the color of photochromatic paint
vent of novel, flexible e-Paper and Organic Light-
properties by UV laser projection. This effect
Emitting Diode (OLED) technologies, it might
lasts for a couple of minutes and demonstrates
be possible to cover a part of a physical model
how fashion and AR could meet.
Figure 3. IMPRESSION OF THE LUMINEX MATERIAL
47
2.4 See-through board See-through boards vary in size between desk-
compelling display system for exhibits and trade
top and hand-held versions. The Augmented
fairs. However, see-through boards obstruct user
engineering system (Bimber et al., 2001) and
interaction with the physical object. Multiple
the AR extension of the haptic sculpting project
viewers cannot share the same device, although
(Bordegoni and Covarrubias, 2007) are examples
a limited solution is offered by the virtual
of the use of see-through technologies, which
showcase by establishing a faceted and curved
typically employ a half-silvered mirror to mix
mirroring surface (Bimber, 2002).
virtual models with a physical object (Figure 4). Similar to the Pepper’s ghost phenomenon, standard stereoscopic Virtual Reality (VR) workbench systems such as the Barco Baron are used to project the virtual information. In addition to the need to wear shutter glasses to view stereoscopic graphics, head tracking is required to align the virtual image between the object and the viewer. An advantage of this approach is that digital images are not occluded by the users’ hand or environment and that graphics can be displayed outside the physical object (i.e., to display the environment or annotations and tools). Furthermore, the user does not have to wear heavy equipment and the resolution of the
Figure 4. THE AUGMENTED ENGINEERING SEE-THROUGH DISPLAY (BIMBER ET AL., 2001).
projection can be extremely high — enabling a
2.5 Spatial projection-based displays This technique is also known as Shader Lamps
mentary to constructing a perspective image
by (Raskar et al., 2001) and was extended in
of a virtual object by a pinhole camera. If the
(Raskar&Bimber, 2004) to a variety of imaging so-
physical object is of the same geometry as the
lutions, including projections on irregular surface
virtual object, a straightforward 3D perspective
textures and combinations of projections with
transformation (described by a 4 × 4 matrix)
(static) holograms. In the field of advertising and
issufficient to predistort the digital image. To
performance arts, this technique recently gained
obtain this transformation, it suffices to indicate
popularity labelled as Projection Mapping: to
6 corresponding points in the physical world
project on buildings, cars or other large objects,
and virtual world: an algorithm entitled Linear
replacing traditional screens as display means, cf.
Camera Calibration can then be applied (see
Figure 5. In such cases, theatre projector systems
Appendix). If the physical and virtual shapes dif-
are used that are prohibitively expensive (>30.000
fer, the projection is viewpoint-dependent and
euros). The principle of spatial projection-based
the head position needs to be tracked. Impor-
technologies is shown in Figure 6. Casting an im-
tant projector characteristics involve weight
age to a physical object is considered comple-
and size versus the power (in lumens) of the
48
Figure 5. TWO PROJECTIONS ON A CHURCH CHAPEL IN UTRECHT (HOEBEN, 2010).
49
projector. There are initiatives to employ LED lasers for direct holographic projection, which also decreases power consumption compared to traditional video projectors and ensures that the projection is always in focus without requiring optics (Eisenberg, 2004). Both fixed and handheld spatial projection-based systems have been demonstrated. At present, hand-held projectors measure 10 × 5 × 2 cm and weigh 150 g, including the processing unit and battery. However, the light output is little (15–45 lumens). The advantage of spatial projection-based tech nologies is that they support the perception of all visual and tactile/haptic depth cues without the need for shutter glasses or HMDs. Furthermore, the display can be shared by multiple co-located users. It requires less expensive equipment, which are often already available at design studios. Challenges to projector-based AR approaches include optics and occlusion. First, only a limited field of view and focus depth can be achieved. To reduce these problems, multiple video projectors can be used. An alternative solution is to employ a portable projector, as Figure 6. PROJECTION-BASED DISPLAY PRINCIPLE
proposed in the iLamps and the I/O Pad concepts
(ADAPTED FROM (RASKAR AND LOW, 2001)), ON THE
(Raskar et al., 2003) (Verlinden et al., 2008).
RIGHT THE DYNAMIC SHADER LAMPS DEMONSTRATION
Other issues include occlusion and shadows,
(BANDYOPADHYAY ET AL., 2001)).
which are cast on the surface by the user or other parts of the system. Projection on nonconvex geometries depends on the granularity and orientation of the projector. The perceived quality is sensitive to projection errors (also known as registration errors), especially projection overshoot (Verlinden et al., 2003b). A solution for this problem is either to include an offset (dilatation) of the physical model or introduce pixel masking in the rendering pipeline. As projectors are now being embedded in consumer cameras and smartphones, we are expecting this type of augmentation in the years to come.
50
3. Input Technologies
Logitech 3D Tracker, Microscribe and Minolta VI900). All these should be considered for object tracking in Augmented prototyping scenarios.
In order to merge the digital and physical, posi-
There are significant differences in the tracker/
tion and orientation tracking of the physical
marker size, action radius and accuracy. As
components is required. Here, we will discuss
the physical model might consist of a number
two different types of input technologies: track-
of parts or a global shape and some additional
ing and event sensing. Furthermore, we will
components (e.g., buttons), the number of items
briefly discuss other input modalities.
to be tracked is also of importance. For simple tracking scenarios, either magnetic or passive optical technologies are often used.
3.1 Position tracking
In some experiments we found out that a projector could not be equipped with a standard Flock of Birds 3D magnetic tracker due to interfer-
Welch and Foxlin (2002) presented a compre-
ence. Other tracking techniques should be used
hensive overview of the tracking principles
for this paradigm. For example, the ARToolkit
that are currently available. In the ideal case,
employs complex patterns and a regular web-
the measurement should be as unobtrusive and
camera to determine the position, orientation
invisible as possible while still offering accurate
and identification of the marker. This is done by
and rapid data. They concluded that there is
measuring the size, 2D position and perspective
currently no ideal solution (‘silver bullet’) for
distortion of a known rectangular marker, cf.
position tracking in general, but some respect-
Figure 7 (Kato and Billinghurst, 1999).
able alternativesare available. Table 2 summarises the most important characteristics of
Passive markers enable a relatively untethered
these tracking methods for Augmented Reality
system, as no wiring is necessary. The optical
purposes. The data have been gathered from
markers are obtrusive when markers are visible
commercially available equipment (the As-
to the user while handling the object. Although
cension Flock of Birds, ARToolkit, Optotrack,
computationally intensive, marker-less optical
Tracking type
Size of tracker (mm)
Typical number of trackers
Action radius/ accuracy
DOF
Issues
Magnetic
16x16x16
2
1.5 m (1 mm)
6
Ferro-magnetic interference
Optical passive
80x80x0.01
>10
3m (1 mm)
6
line of sight
Optical active
10x10x5
>10
3m (0.5 mm)
3
line of sight, wired connections
Ultrasound
20x20x10
1
1m (3 mm)
6
line of sight
Mechanical linkage
defined by working envelope
1
0.7 m (0.1 mm)
5
limited degrees of freedom, inertia
Laser scanning
none
infinite
2m ( 0.2mm)
6
line of sight, frequency, object recognition
Table 2. SUMMARY OF TRACKING TECHNOLOGIES. 51
Figure 7. WORKFLOW OF THE ARTOOLKIT OPTICAL TRACKING ALGORITHM,
http://www.hitl.washington.edu/artoolkit/documentation/userarwork.html
tracking has been proposed (Prince et al.,2002).
bodies. This method has a number of challenges
The employment of Laser-Based tracking sys-
when used as a real-time tracking means, includ-
tems is demonstrated by the illuminating Clay
ing the recognition of objects and their posture.
system by Piper et al. (2002): a slab of Plasti-
However, with the emergence of depth cameras
cine acts as an interactive surface — the user
for gaming such as the Kinect (Microsoft), similar
influences a 3D simulation by sculpting the clay,
systems are now being devised with a very small
while the simulation results are projected on the
technological threshold.
surface. A laser-based Minolta Vivid 3D scanner is employed to continuously scan the clay
In particular cases, a global measuring system is
surface. In the article, this principle was applied
combined with a different local tracking principle
to geodesic analysis, yet it can be adapted to
to increase the level of detail, for example, to
design applications, e.g., the sculpting of car
track the position and arrangement of buttons on
Figure 8. ILLUMINATING CLAY SYSTEM WITH A PROJECTOR/LASER SCANNER (PIPER ET AL., 2002).
52
the object’s surface. Such local positioning sys-
sliders, rotation knobs and sensors to measure
tems might have less advanced technical require-
force, touch and light. More elaborate compo-
ments; for example, the sampling frequency can
nents like a mini joystick, Infrared (IR) motion
be decreased to only once a minute. One local
sensor, air pressure and temperature sensor are
tracking system is based on magnetic resonance,
commercially available. Similar initiatives are
as used in digital drawing tablets. The Sensetable
iStuff (Ballagas et al., 2003), which also hosts a
demonstrates this by equipping an altered com-
number of wireless connections to sensors. Some
mercial digital drawing tablet with custom-made
systems embed switches with short-range wire-
wireless interaction devices (Patten et al., 2001).
less connections, for example, the Switcheroo
The Senseboard (Jacob et al., 2002) has similar
and Calder systems (Avrahami and Hudson, 2002;
functions and an intricate grid of RFID receivers
Lee et al., 2004) (cf. Figure 9). This allows a
to determine the (2D) location of an RFID tag on
greater freedom in modifying the location of the
a board. In practice, these systems rely on a rigid
interactive components while prototyping. The
tracking table, but it is possible to extend this to
Switcheroo system uses custom-made RFID tags.
a flexible sensing grid. A different technology was
A receiver antenna has to be located nearby
proposed by Hudson (2004) to use LED pixels as
(within a 10-cm distance), so the movement en-
light emitters and sensors. By operating one pixel
velope is rather small, while the physical model
as a sensor whilst its neighbours are illuminated,
is wired to a workstation. The Calder toolkit
it is possible to detect light reflected from a
(Lee et al., 2004) uses a capacitive coupling
fingertip close to the surface. This principle could
technique that has a smaller range (6 cm with
be applied to embedded displays, as mentioned
small antennae), but is able to receive and trans-
in Section 2.3.
mit for long periods on a small 12 mm coin cell. Other active wireless technologies would draw more power, leading to a system that would
3.2 Event sensing
only fit a few hours. Although the costs for this system have not been specified, only standard electronics components are required to build
Apart from location and orientation tracking,
such a receiver.
Augmented prototyping applications require interaction with parts of the physical object, for example, to mimic the interaction with the
Hand tracking
artefact. This interaction differs per AR scenario, so a variety of events should be sensed
Instead of attaching sensors to the physi-
to cater to these applications.
cal environment, fingertip and hand tracking technologies can also be used to generate user
Physical sensors
events. Embedded skins represent a type of interactive surface technology that allows the accurate measurement of touch on the object’s
The employment of traditional sensors labelled
surface (Paradiso et al., 2000). For example, the
‘physical widgets’ (phidgets) has been studied
Smartskin by Reikimoto (2002) consists of a flex-
extensively in the Computer-Human Interface
ible grid of antennae. The proximity or touch of
(CHI) community. Greenberg and Fitchett (2001)
human fingers changes the capacity locally in the
introduced a simple electronics hardware and
grid and establishes a multi-finger tracking cloth,
software library to interface PCs with sensors
which can be wrapped around an object. Such a
(and actuators) that can be used to discern
solution could be combined with embedded dis-
user interaction. The sensors include switches,
plays, as discussed in Section 2.3. Direct electric
53
Figure 9. MOCKUP EQUIPPED WITH WIRELESS SWITCHES THAT CAN BE RELOCATED TO EXPLORE USABILITY (LEE ET AL., 2004).
contact can also be used to track user interac-
tip and hand tracking as well. A simple example
tion; the Paper Buttons concept (Pedersen et
is the light widgets system (Fails and Olsen,
al.,2000) embeds electronics on the objects and
2002) that traces skin colour and determines
equips the finger with a two-wire plug that sup-
finger/hand position by 2D blobs. The OpenNI
plies power and allows bidirectional communica-
library enables hand and body tracking of depth
tion with the embedded components when they
range cameras such as the Kinect (OpenNi.org).
are touched. Magic Touch (Pedersen, 2001) uses
A more elaborate example is the virtual drawing
a similar wireless system; the user wears an RFID
tablet by Ukita and Kidode (2004); fingertips
reader on his or her finger and can interact by
are recognised on a rectangular sheet by a
touching the components, which have hidden
head-mounted infrared camera. Traditional VR
RFID tags. This method has been adapted to
gloves can also be used for this type of tracking
Augmented Reality for design by Kanai et al.
(Schäfer et al., 1997).
(2007). Optical tracking can be used for finger
54
3.3 Other input modalities
tions and constraints in terms of the field of view and resolution and lend themselves to a kind of isolation. For all display technologies,
Speech and gesture recognition require consid-
the current challenges include an untethered
eration in AR as well. In particular, pen-based
interface, the enhancement of graphics capabili-
interaction would be a natural extension to the
ties, visual coverage of the display and improve-
expressiveness of today’s designer skills. Oviatt
ment of resolution. LED-based laser projection
et al. (2000) offer an comprehensive overview of
and OLEDs are expected to play an important
the so-called Recognition-Based User Interfaces
role in the next generation of IAP devices
(RUIs), including the issues and Human Factors
because this technology can be employed by
aspects of these modalities. Furthermore,
see-through or projection-based displays.
speech-based interaction can also be useful to
To interactively merge the digital and physical
activate operations while the hands are used for
parts of Augmented prototypes, position and
selection.
orientation tracking of the physical components is needed, as well as additional user input means.For global position tracking, a variety of
4. Conclusions and Further reading
principles exist. Optical tracking and scanning suffer from the issues concerning line of sight and occlusion. Magnetic, mechanical linkage and ultrasound-based position trackers are obtrusive
This article introduces two important hardware
and only a limited number of trackers can be
systems for AR: displays and input technologies.
used concurrently.
To superimpose virtual images onto physical models, head mounted-displays (HMDs), see-
The resulting palette of solutions is summarized
through boards, projection-based techniques
in Table 3 as a morphological chart. In devising a
and embedded displays have been employed.
solution for your AR system, you can use this as
An important observation is that HMDs, though
a checklist or inspiration of display and input.
best known by the public, have serious limita-
Display Imaging principle
Video Mixing
Projectorbased
See-through
Display arrangment
Headattached
Handheld/ wearable
Spatial
Position tracking Input technologies
Event sensing
Optical Magnetic
Passive markers Physical sensors
Wired connection
Wireless
Active markers
3D laser scanning
Ultrasound
Mechanical
Virtual Surface tracking
3D tracking
Table 3. Morphological chart of AR enabling technologies.
55
Further reading
eters of the projector, respectively. Then a point P in 3D-space is transformed to:
For those interested in research in this area,
p=[I·E] ·P
(1)
the following publication means offer a range of detailed solutions:
where p is a point in the projector’s coordinate
■ International Symposium on Mixed and
system. If we decompose rotation and transla-
Augmented Reality (ISMAR) – ACM-sponsored
tion components in this matrix transformation
annual convention on AR, covering both spe-
we obtain:
cific applications as emerging technologies.
p=[R t] ·P
(2)
accesible through http://dl.acm.org ■ Augmented Reality Times — a daily update
In which R is a 3x3 matrix corresponding to the
on demos and trends in commercial and aca-
rotational components of the transformation and
demic AR systems: http://artimes.rouli.net
t the 3x1 translation vector. Then we split the
■ Procams workshop — annual workshop on
rotation columns into row vectors R1, R2, and R3
projector-camera systems, coinciding with
of formula 3. Applying the perspective division
the IEEE conference on Image Recognition
results in the following two formulae:
and Robot Vision. The resulting proceedings are freely accessible at http://www.procams.
(3)
(4)
org ■ Raskar, R. and Bimber, O. (2004) Spatial Augmented Reality, A.K. Peters, ISBN: 1568812302 – personal copy can be downloaded for free at http://140.78.90.140/medien/ar/SpatialAR/download.php ■ BuildAR – download simple webcam-based application that uses markers, http://www.
in which the 2D point pi is split into (ui,vi).
buildar.co.nz/buildar-free-version
Given n measured point-point correspondences (pi; Pi); (i = 1::n), we obtain 2n equations:
Appendix: Linear Camera
R1·Pi – ui·R3·Pi + t x - ui·tz = 0
(5)
Calibration
R 2·Pi – vi·R3·Pi + ty - ui·tz = 0
(6)
This procedure has been published in (Raskar
We can rewrite these 2n equations as a matrix
and Bimber, 2004) to some degree, but is slightly
multiplication with a vector of 12 unknown
adapted to be more accessible for those with
variables, comprising the original transformation
less knowledge of the field of image processing.
components R and t of formula 3. Due to mea-
C source code that implements this mathemati-
surement errors, a solution is usually non-singu-
cal procedure can be found in appendix A1 of
lar; we wish to estimate this transformation with
(Faugeras, 1993). It basically uses point corres
a minimal estimation deviation. In the algorithm
pondences between original x,y,z coordinates
presented at (Bimber & Raskar, 2004), the mini-
and their projected u,v, counterparts to resolve
max theorem is used to extract these based on
internal and external camera parameters.
determining the singular values. In a straightfor-
In general cases, 6 point correspondences are
ward matter, internal and external transforma-
sufficient (Faugeras 1993, Proposition 3.11).
tions I and E of formula 1 can be extracted from
Let I and E be the internal and external param-
the resulting transformation.
56
References ■ Avrahami, D. and Hudson, S.E. (2002) ‘Forming interactivity: a tool for rapid pro
empirical testing’, Int. J. Technology Management, Vol. 21, Nos. 3–4, pp.340–352.
totyping of physical interactive products’, Proceedings of DIS ‘02, pp.141–146.
■ Bordegoni, M. and Covarrubias, M. (2007) ‘Augmented visualization system for a haptic
■ Azuma, R. (1997)
interface’, HCI International 2007 Poster.
‘A survey of augmented reality’, Presence: Teleoperators and Virtual Environments, Vol. 6, No. 4, pp.355–385.
■ Eisenberg, A. (2004) ‘For your viewing pleasure, a projector in your pocket’, New York Times, 4 November.
■ Azuma, R., Baillot, Y., Behringer, R., Feiner, S., Julier, S. and MacIntyre, B. (2001)
■ Faugeras, O. (1993)
‘Recent advances in augmented reality’, IEEE
‘Three-Dimensional Computer Vision:
Computer Graphics and Applications, Vol. 21,
a Geometric Viewpoint’, MIT press.
No. 6, pp.34–47. ■ Fails, J.A. and Olsen, D.R. (2002) ■ Ballagas, R., Ringel, M., Stone, M. and Borchers, J. (2003)
‘LightWidgets: interacting in everyday spaces’, Proceedings of IUI ‘02, pp.63–69.
‘iStuff: a physical user interface toolkit for ubiquitous computing environments’, Proceedings of CHI 2003, pp.537–544.
■ Greenberg, S. and Fitchett, C. (2001) ‘Phidgets: easy development of physical interfaces through physical widgets’, Proceed-
■ Bandyopadhyay, D., Raskar, R. and Fuchs,
ings of UIST ‘01, pp.209–218.
H. (2001) ‘Dynamic shader lamps: painting on movable
■ Hoeben, A. (2010)
objects’, International Symposium on Aug-
“Using a projected Trompe L’Oeil to highlight
mented Reality (ISMAR), pp.207–216.
a church interior from the outside”, EVA 2010
■ Bimber, O. (2002) ‘Interactive rendering for projection-based
■ Hudson, S. (2004)
augmented reality displays’, PhD disserta-
‘Using light emitting diode arrays as touch-
tion, Darmstadt University of Technology.
sensitive input and output devices’, Proceedings of the ACM Symposium on User Interface
■ Bimber, O., Stork, A. and Branco, P. (2001)
Software and Technology, pp.287–290.
‘Projection-based augmented engineering’, Proceedings of International Conference on
■ Jacob, R.J., Ishii, H., Pangaro, G. and Pat-
Human-Computer Interaction (HCI’2001),
ten, J. (2002)
Vol. 1, pp.787–791.
‘A tangible interface for organizing information using a grid’, Proceedings of CHI ‘02,
■ Bochenek, G.M., Ragusa, J.M. and Malone,
pp.339–346.
L.C. (2001) ‘Integrating virtual 3-D display systems into product design reviews: some insights from
57
■ Kanai, S., Horiuchi, S., Shiroma, Y., Yo-
■ Pederson, T. (2001)
koyama, A. and Kikuta, Y. (2007)
‘Magic touch: a simple object location track-
‘An integrated environment for testing and
ing system enabling the development of
assessing the usability of information appli-
physical-virtual artefacts in office environ-
ances using digital and physical mock-ups’,
ments’, Personal Ubiquitous Comput., Janu-
Lecture Notes in Computer Science, Vol.
ary, Vol. 5, No. 1, pp.54–57.
4563, pp.478–487. ■ Piper, B., Ratti, C. and Ishii, H. (2002) ■ Kato, H. and Billinghurst, M. (1999)
‘Illuminating clay: a 3-D tangible interface
‘Marker tracking and HMD calibration for a
for landscape analysis’, Proceedings of CHI
video-based augmented reality conferenc-
‘02, pp.355–362.
ing system’, Proceedings of International Workshop on Augmented Reality (IWAR 99), pp.85–94.
■ Prince, S.J., Xu, K. and Cheok, A.D. (2002) ‘Augmented reality camera tracking with homographies’, IEEE Comput. Graph. Appl.,
■ Klinker, G., Dutoit, A.H., Bauer, M., Bayer,
November, Vol. 22, No. 6, pp.39–45.
J., Novak, V. and Matzke, D. (2002) ‘Fata Morgana – a presentation system for
■ Raskar, R., Welch, G., Low, K-L. and
product design’, Proceedings of ISMAR ‘02,
Bandyopadhyay, D. (2001)
pp.76–85.
‘Shader lamps: animating real objects with image based illumination’, Proceedings
■ Oviatt, S.L., Cohen, P.R., Wu, L., Vergo, J., Duncan, L., Suhm, B., Bers, J., et al.
of Eurographics Workshop on Rendering, pp.89–102.
(2000) ‘Designing the user interface for multimodal
■ Raskar, R. and Low, K-L. (2001)
speech and gesture applications: state-of-
‘Interacting with spatially augmented real-
the-art systems and research directions’,
ity’, ACM International Conference on Virtual
Human Computer Interaction, Vol. 15, No. 4,
Reality, Computer Graphics and Visualization
pp.263–322.
in Africa (AFRIGRAPH), pp.101–108.
■ Paradiso, J.A., Hsiao, K., Strickon, J.,
■ Raskar, R., van Baar, J., Beardsley, P.,
Lifton, J. and Adler, A. (2000)
Willwacher, T., Rao, S. and Forlines, C.
‘Sensor systems for interactive surfaces’,
(2003)
IBM Systems Journal, Vol. 39, Nos. 3–4,
‘iLamps: geometrically aware and self-con-
pp.892–914.
figuring projectors’, SIGGRAPH, pp.809–818.
■ Patten, J., Ishii, H., Hines, J. and Pangaro,
■ Raskar, R. and Bimber, O. (2004)
G. (2001)
Spatial Augmented Reality, A.K. Peters,
‘Sensetable: a wireless object tracking plat-
ISBN: 1568812302.
form for tangible user interfaces’, Proceedings of CHI ‘01, pp.253–260.
■ Reikimoto, J. (2002) ‘SmartSkin: an infrastructure for freehand
■ Pedersen, E.R., Sokoler, T. and Nelson, L. (2000) ‘PaperButtons: expanding a tangible user interface’, Proceedings of DIS ’00, pp.216–223.
58
manipulation on interactive surfaces’, Proceedings of CHI ‘02, pp.113–120.
■ Saakes, D.P., Chui, K., Hutchison, T.,
■ Verlinden, J., Horvath, I. (2008)
Buczyk, B.M., Koizumi, N., Inami, M.
”Enabling interactive augmented prototyp-
and Raskar, R. (2010)
ing by portable hardware and a plugin-based
’ Slow Display’. In SIGGRAPH 2010 emerg-
software architecture” Journal of Mechani-
ing technologies: Proceedings of the 37th
cal Engineering, Slovenia, Vol 54(6), pp.
annual conference on Computer graphics and
458-470.
interactive techniques, July 2010. ■ Welch, G. and Foxlin, E. (2002) ■ Schäfer, K., Brauer, V. and Bruns, W.
‘Motion tracking: no silver bullet, but a re-
(1997)
spectable arsenal’, IEEE Computer Graphics
‘A new approach to human-computer
and Applications, Vol. 22, No. 6, pp.24–38.
interaction – synchronous modelling in real and virtual spaces’, Proceedings of DIS ‘97, pp.335–344.
.
■ Schall, G., Mendez, E., Kruijff, E., Veas, E., Sebastian, J., Reitinger, B. and Schmalstieg, D. (2008) ‘Handheld augmented reality for underground infrastructure visualization’, Journal of Personal and Ubiquitous Computing, Springer, DOI 10.1007/s00779-008-0204-5. ■ Schmalstieg, D. and Wagner, D. (2008) ‘Mobile phones as a platform for augmented reality’, Proceedings of the IEEE VR 2008 Workshop on Software Engineering and Architectures for Realtime Interactive Systems, pp.43–44. ■ Sutherland, I.E. (1968) ‘A head-mounted three-dimensional display’, Proceedings of AFIPS, Part I, Vol. 33, pp.757–764. ■ Ukita, N. and Kidode, M. (2004) ‘Wearable virtual tablet: fingertip drawing on a portable plane-object using an activeinfrared camera’, Proceedings of IUI 2004, pp.169–175. ■ Verlinden, J.C., de Smit, A., Horváth, I., Epema, E. and de Jong, M. (2003) ‘Time compression characteristics of the augmented prototyping pipeline’, Proceedings of Euro-uRapid’03, p.A/1.
59
LIKE RIDING A BIKE. LIKE PARKING A CAR. PORTRAIT OF THE ARTIST IN RESIDENCE
MARINA DE HAAS BY HANNA SCHRAFFENBERGER
60
"Hi Marina. Nice to meet you! I have heard a lot about you."
support of the lab, she has realized the AR artworks Out of the blue and Drops of white in the course of her study. In 2008 she graduated with
I usually avoid this kind of phrases. Judging from
an AR installation that shows her 3d animated
my experience, telling people that you have
portfolio. Then, having worked with AR for three
heard a lot about them makes them feel uncom-
years, she decided to take a break from technol-
fortable. But this time I say it. After all, it’s no
ogy and returned to photography, drawing and
secret that Marina and the AR Lab in The Hague
painting. Now, after yet another three years,
share a history which dates back much longer
she is back in the mixed reality world. Convinced
than her current residency at the AR Lab. At the
by her concepts for future works, the AR Lab
lab, she is known as one of the first students
has invited her as an Artist in Residence. That is
who overcame the initial resistance of the fine
what I have heard about her, and made me want
arts program and started working with AR. With
to meet her for an artist-portrait. Knowing quite
61
a lot about her past, I am interested in what she is currently working on, in the context of her residency. When she starts talking, it becomes
ogy, the audience will then see a dyingdove or dying crane fly with a missing foot.”
clear that she has never really stopped thinking about AR. There’s a handwritten notebook
Marina tells me her current piece is about imper-
full of concepts and sketches for future works.
manence and mortality, but also about the fact
Right now, she is working on animations of two
that death can be the beginning of something
animals. Once she is done animating, she'll use
new. Likewise, the piece is not only about death
AR technology to place the animals — an insect
but also intended as an introduction and begin-
and a dove — in the hands of the audience.
ning for a forthcoming work. The AR Lab makes
“I usually start out with my own photographs and a certain space I want to augment.” "They will hold a little funeral monument in the shape of a tile in their hands. Using AR technol62
this beginning possible through financial support but also provides technical assistance and serves as a place for mutual inspiration and exchange. Despite her long break from the digital arts, the young artist feels confident about working with AR again:
“It’s a bit like biking, once you’ve learned it, you never unlearn it. It’s the same with me and AR, of course I had to practice a bit, but I still have the feel for it. I think working with AR is just a part of me.”
After having paused for three years, Marina is positively surprised about how AR technology has emerged in the meantime:
“AR is out there, it’s alive, it’s growing and finally, it can be markerless. I don’t like the use of markers. They are not part of my art and people see them, when they don’t wear AR glasses. I am also glad that so many people know AR from their mobile phones or at least have heard about it before. Essentially, I don’t want the audience to wonder about the technology, I want them to look at the pictures and animations I create. The more people are used to the technology the more they will focus on the content. I am really happy and excited how AR has evolved in the last years!”
“When I was a child I found a book with code and so I programmed some games. That was fun, I just understood it. It’s the same with creating AR works now. My way of thinking perfectly matches with how AR works. It feels completely natural to me.” Nevertheless, working with technology also has its downside:
“The most annoying thing about working with AR is that you are always facing technical limitations and there is so much that can go wrong. No matter how well you do it, there is always the risk that something won’t work. I hope for technology to get more stable in the future.” When realizing her artistic augmentations,
I ask, how working with brush and paint differs
Marina sticks to an established workflow:
from working with AR, but there seems to be surprisingly little difference.
“The main difference is that with AR I am working with a pen-tablet, a computer and a screen. I control the software, but if I work with a brush I have the same kind of control over it. In the past, I used to think that there was a difference, but now I think of the computer as just another medium to work with. There is no real difference between working with a brush and working with a computer. My love for technology is similar to my love for paint.”
“I usually start out with my own photographs and a certain space I want to augment. Preferably I measure the dimensions of the space, and then I work with that
Marina discovered her love for technology at a young age:
63
room in my head. I have it in my inner vision and I think in pictures. There is a photo register in my head which I can access. It’s a bit like parking a car. I can park a car in a very small space extremely well. I can feel the car around me and I can feel the space I want to put it in. It’s the same with the art I create. Once I have clear idea of the artwork I want to create, I use Cinema4D software to make 3d models. Then I use BuildAR to place my 3d models it the real space. If everything goes well, things happen that you could not have imagined.” A result of this process is, for example, the AR installation Out of the blue which was shown at Today’s Art festival in The Hague in 2007:
“The idea behind ‘Out of the blue’ came from a photograph I took in an elevator. I took the picture so that the lights in the elevator looked like white ellipses on a black background. I took this basic elliptical shape as a basis for working in a very big space. I was very curious if I could use such a simple shape and still convince the audience that it really existed in the space. And it worked — people tried to touch it with their hands and were very surprised when that wasn’t possible.” The fact that people believe in the existence of her virtual objects is also important for Marina’s personal understanding of AR:
64
“For me, Augmented Reality means using digital images to create something which is not real. However, by giving meaning to it, it becomes real and people realize that it might as well exist.” I wonder whether there is a specific place or space she’d like to augment in the future and Marina has quite some places in mind. They have one thing in common: they are all known museums that show modern art.
“I would love to create works for the big museums such as the TATE Modern or MoMa. In the Netherlands, I’d love to augment spaces at the Stedelijk Museum in Amsterdam or Boijmans museum in Rotterdam. That’s my world. Going to a museum means a lot to me. Of course, one can place AR artworks everywhere, also in public spaces. But it is important to me that people who experience my work have actively chosen to go somewhere to see art. I don’t want them to just see it by accident at a bus stop or in a park.” Rather than placing her virtual models in a specific physical space, her current work follows a different approach. This time, Marina will place the animated dying animals in the hands of the audiences. The artist has some ideas about how to design this physical contact with the digital animals.
“In order for my piece to work, the viewer needs to feel like he is holding something in his hand. Ideally, he will feel the weight of
the animal. The funeral monuments will therefor have a certain weight.”
Coming from a fine arts background, Marina has a tip for art students who want to to follow in her footsteps and are curious about working with AR:
It is still open where and when we will be able to experience the piece:
“My residency lasts 10 weeks. But of course that’s not enough time to finish. In the past, a piece was finished when the time to work on it was up. Now, a piece is finished when it feels complete. It’s something I decide myself, I want to have control over it. I don’t want any more restrictions. I avoid deadlines.”
“I know it can be difficult to combine technology with art, but it is worth the effort. Open yourself up to for art in all its possibilities, including AR. AR is a chance to take a step in a direction of which you have no idea where you’ll find yourself. You have to be open for it and look beyond the technology. AR is special — I couldn’t live without it any more...”
65
BIOGRAPHY JEROEN VAN ERP Under Jeroen’s joint leadership, Fabrique has grown through the years into a multifaceted design bureau. It currently employs more than 100 artists, engineers and storytellers working for a wide range of customers: from supermarket chain Albert Heijn to the Rijksmuseum. Fabrique develops visions, helps its clients think about strategies, branding and innovation and realises designs. Preferably cutting straight through the design disciplines, so that the traditional borders between graphic design, industrial design, spatial design and interactive media are sometimes barely re cognisable. In the bureau’s vision, this cross
JEROEN VAN ERP GRADUATED FROM THE FACULTY OF INDUSTRIAL DESIGN AT THE TECHNICAL UNIVERSITY OF DELFT IN 1988. IN 1992, HE WAS ONE OF THE FOUNDERS OF FABRIQUE IN DELFT, WHICH POSITIONED ITSELF AS A MULTIDISCIPLINARY DESIGN BUREAU. HE ESTABLISHED THE INTERACTIVE MEDIA DEPARTMENT IN 1994, FOCUSING PRIMARILY ON DEVELOPING WEBSITES FOR THE WORLD WIDE WEB - BRAND NEW AT THAT TIME.
media approach will be the only way to create apparently simple solutions for complex and relevant issues. The bureau also opened a studio in Amsterdam in 2008. Jeroen is currently CCO (Chief Creative Officer) and a partner, and in this role he is responsible for the creative policy of the company. He has also been closely involved in various projects as art director and de signer. He is a guest lecturer for various courses and is a board member at NAGO (the Netherlands Graphic Design Archive) and the Design & Emotion Society.
www.fabrique.nl 66
A MAGICAL LEVERAGE IN SEARCH OF THE KILLER APPLICATION
BY JEROEN VAN ERP The moment I was confronted with the technology of Augmented Reality, back in 2006 at the Royal Academy of Arts in The Hague, I was thrilled. Despite the heavy helmet, the clumsy equipment, the shaky images and the lack of a well-defined purpose, it immediately had a profound impact on me. From the start, it was clear that this technology had a lot of potential, although at first it was hard to grasp why. Almost six years later, the fog that initially surrounded this new technology has gradually faded away. To start with, the technology itself is developing rapidly, as is the equipment.
But more importantly: companies and cul-
tural institutions are starting to understand how they can benefit from this technology. At the moment there are a variety of applications available (mainly mobile applications for tablets or smart phones) that create added value for the user or consumer. This is great, because it not only allows the audience to gain experience in the field of this still-developing technology, but also the industry. But to make Augmented Reality a real success, the next step will be of vital importance.
67
INNOVATING OR INNOVATING?
a social goal. A business goal is often derived
Let’s have a look at different forms of innovat-
which is expected to generate an economic
ing in figure 1. On the left we see innovations
benefit for the company. A marketing special-
with a bottom-up approach, and on the right a
ist would state that there is already a market.
top-down approach to innovating. A bottom-
This approach means that you have to inno-
up approach means that we have a promising
vate with an intended goal in mind. A business
new technique, concept or idea although the
goal-driven innovation can be a product inno-
exact goal or matching business model aren’t
vation (either physical products, services or a
clear yet. In general, bottom-up developments
combination of the two) or a brand innovation
are technological or art-based, and are there-
(storytelling, positioning), but always with an
fore what I would call autonomous: the means
intended economical or social benefit in mind.
are clear, but the exact goal has still to be
As there is an expected benefit, people are
defined. The usual strategy to take it further
willing to invest.
from a benefit for the user or the consumer,
is to set up a start-up company in order to develop the technique and hopefully to create
It’s interesting to note the difference on the
a market.
vertical axis between radical innovations and
This is not always that simple. Innovating from
incremental changes (Robert Verganti – Design
a top-down approach means that the innova-
Drive Innovation). Incremental changes are
tion is steered on the basis of a more or less
improvements of existing concepts or prod-
clearly defined goal. In contrast with bottom-
ucts. This is happening a lot, for instance in
up innovations, the goal is well-defined and
the automotive industry. In general, a radical
the designer or developer has to choose the
innovation changes the experience of the
right means, and design a solution that fits
product in a fundamental way, and as a result
the goal. This can be a business goal, but also
of this often changes an entire business.
Figure 1.
68
This is something Apple has achieved several
me that the experience of AR wasn’t suitable
times, but it has also been achieved by Tom-
at all for this form of publishing. AR doesn’t do
Tom, and by Philips and Douwe Egberts with
well on a projection screen. It does well in the
their Senseo coffee machine.
user’s head, where time, place, reality and imagination can play an intriguing game with our senses. It is unlikely that the technique of
HOW ABOUT AR?
Augmented Reality will lead to mass consump-
What about the position of Augmented Real-
a lot of people at the same time’. No, by their
ity? To start with, the Augmented Reality
nature, AR applications are intimate and in-
technique is not a standalone innovation. It’s
tense, and this is one of its biggest assets.
tion as in ‘experiencing the same thing with
not a standalone product but a technique or feature that can be incorporated into products or services with a magical leverage. At its core it is a technique that was developed — and is
FUTURE
still developing — with specialist purposes in
We have come a long way, and the things we
mind. In principle there was no big demand
can do with AR are becoming more amazing by
from ‘the market’. Essentially, it is a bottom-
the day. The big challenge is to make it appli-
up technological development that needs a
cable in relevant solutions. There’s no discus-
concept, product or service.
sion about the value of AR in specialist areas, such as the military industry. Institutions in
You can argue about whether it is an incre-
the field of art and culture have discovered
mental innovation or a radical one. A virtual
the endless possibilities, and now it is the
reality expert will probably tell you that it is
time to make the big leap towards solutions
an improvement (incremental innovation) of
with social or economic value (the green area
the VR technique. But if you look from an ap-
in figure 1). This will give the technique the
plication perspective, there is a radical aspect
chance to develop further in order to flourish
to it. I prefer to keep the truth in the middle.
at the end. From that perspective, it wouldn’t
At this moment in time, AR is in the blue area
surprise me if the first really good, efficient
(figure 1).
and economically profitable application will
It is clear that bottom-up innovation and top-
emerge for educational purposes.
down innovation are different species. But when it comes to economic leverage, it is a
Let’s not forget we are talking about a tech-
challenge to be part of the top-down game.
nology that is still in its infant years. When I
This provides a guarantee for further develop-
look back at the websites we made 15 years
ment, and broad acceptation of the technique
ago, I realize the gigantic steps we have made,
and principles. So the major challenge for AR
and I am aware of the fact that we could
is to make the big step to the right part of fig-
hardly imagine then what the impact of the
ure 1 as indicated by the red arrow. Although
internet would be on society today. Of course,
the principles of Augmented Reality are very
it’s hard to compare the concept of Augment-
promising, it’s clear we aren’t there yet. An
ed Reality with that of the internet, but it is
example: we recently received a request to
a valid comparison, because it gives the same
‘do something’ with Augmented Reality. The
powerless feeling of not being able to predict
idea was to project the result of an AR appli-
its future. But it will probably be bigger than
cation onto a big wall. Suddenly it occurred to
you can imagine.
69
THE POSITIONING OF VIRTUAL OBJECTS ROBERT PREVEL
WHEN USING AUGMENTED REALITY (AR) FOR VISION, VIRTUAL OBJECTS ARE ADDED TO THE REAL WORLD AND DISPLAYED IN SOME WAY TO THE USER; BE THAT VIA A MONITOR, PROJECTOR, OR HEAD-MOUNTED DISPLAY (HMD). OFTEN IT IS DESIRABLE, OR EVEN UNAVOIDABLE, FOR THE VIEWPOINT OF THE USER TO MOVE AROUND THE ENVIRONMENT (THIS IS PARTICULARLY THE CASE IF THE USER IS WEARING A HMD). THIS PRESENTS A PROBLEM, REGARDLESS OF THE TYPE OF DISPLAY USED: HOW CAN THE VIEWPOINT BE DECOUPLED FROM THE AUGMENTED VIRTUAL OBJECTS?
position and orientation (pose) in 3D space, and its scale, should be. However, if the view point changes, then how we view the virtual object should also change. For example, if Iwalk around to face the back of a virtual object, I expect to be able to see the rear of that object. The solution to this problem is to keep track of the user’s viewpoint and, in the event that the viewpoint changes, to update the pose of any virtual content accordingly. There are a number of ways in which this can be achieved, by using, for example: positional sensors (such as inertia trackers), a global positioning system, computer vision techniques, etc. Typically the best results are those systems that take the data from many tracking systems and blend them together. At TU Delft, we have been researching and developing techniques to track position using computer vision techniques. Often it is the case that video cameras are used in AR systems; indeed, in the case where the AR system uses video see-through, the use of cameras is necessary. Using computer vision techniques, we can identify landmarks in the
To recap, virtual objects are blended with the
environment, and, using these landmarks, we
real world view in order to achieve an Aug-
can determine the pose of our camera with
mented world view. From our initial viewpoint
basic geometry. If the camera is not used
we can determine what the virtual object’s
directly as the viewpoint (as is the case in
70
optical see-through systems), then we can still
does not change, should deliver the same AR
keep track of the viewpoint by attaching the
experience each time. Sometimes however,
camera to it. Say, for example, that we have
it is not feasible to prepare an environment
an optical see-through HMD with an attached
with markers. Often it is desirable to use an
video camera. Then, if we calculate the pose
AR application in an unknown or unprepared
of the camera, we can then determine the
environment. In these cases, an alternative
pose of the viewpoint, provided that the
to using markers is to identify the natural
camera’s position relative to the viewpoint
features found in the environment.
remains fixed. The term ‘natural features’ can be used to The problem then, has been reduced to
describe the parts of an image that stand out.
identifying landmarks in the environment.
Examples include: edges, corners, areas of
Historically, this has been achieved by the
high contrast, etc. In order to be able to use
use of fiducial markers, which act as points
the natural features to track the camera posi-
of reference in the image. Fiducial markers
tion in an unknown environment, we need to
provide us with a means of determining the
be able to first identify the natural features,
scale of the visible environment, provided
and then determine their relative positions
that: enough points of reference are visible,
in the environment. Whereas you could place
we know their relative positions, and these
20 markers in an environment and still only
relative positions don’t change. A typical
have 80 identifiable corners, there are often
marker often used in AR applications consists
hundreds of natural features in any one image.
of a card with a black rectangle in the centre,
This makes using natural features a more ro-
a white border, and an additional mark to
bust solution than using markers, as there are
determine which edge of the card is consid-
far more landmarks we can use to navigate,
ered the bottom. As we know that the corners
not all of which need to be visible. One of the
of the black rectangle are all 90 degrees, and
key advantages to using natural features over
we know the distance between corners, we
markers is that: as we already need to identify
can identify the marker and determine the
and keep track of those natural features seen
pose of the camera with regard to the points
from our initial view point, we can use the
of reference (in this case the four corners of
same method to continually update a 3D map
the card).
of features as we change our view point. This allows our working environment to grow,
A large number of simple ‘desktop’ AR applica-
which could not be achieved in a prepared
tions make use of individual markers to track
environment.
camera pose, or conversely, to track the posi-
Although we are able to determine the rela-
tion of the markers relative to our viewpoint.
tive distance between features, the question
Larger applications require multiple markers
remains: how can we determine the absolute
linked together, normally distinguishable by
position of features in an environment without
a unique pattern or barcode in the centre
some known measurement? The short answer
of the marker. Typically the more points of
is that we cannot; either we need to estimate
reference that are visible in a scene, the bet-
the distance or we can introduce a known
ter the results when determining the camera’s
measurement. In a future edition we will
pose. The key advantage to using markers
discuss the use of multiple video cameras and
for tracking the pose of the camera is that
how, given the absolute distance between the
an environment can be carefully prepared
cameras, we can determine the absolute posi-
in advance, and provided the environment
tion of our identified features. 71
MEDIATED REALITY FOR CRIME SCENE INVESTIGATION1 STEPHAN LUKOSCH CRIME SCENE INVESTIGATION IN THE NETHERLANDS IS PRIMARILY THE RESPONSIBILITY OF THE LOCAL POLICE. FOR SEVERE CRIMES, A NATIONAL TEAM SUPPORTED BY THE NETHERLANDS FORENSIC INSTITUTE (NFI) IS CALLED IN. INITIALLY CAPTURING ALL DETAILS OF A CRIME SCENE IS OF PRIME IMPORTANCE (SO THAT EVIDENCE IS NOT ACCIDENTLY DESTROYED). NFI’S DEPARTMENT OF DIGITAL IMAGE ANALYSIS USES THE INFORMATION COLLECTED FOR 3D CRIME SCENE RECONSTRUCTION AND ANALYSIS. Within the CSI The Hague project (http://
for future crime scene investigation and to
www.csithehague.com) several companies and
tackle current issues in crime scene investi-
research institutes cooperate under the guid-
gation. In AugmentedReality, virtual data is
ance of the Netherlands Forensic Institute in
spatially overlaid on top of physical reality. With
order to explore new technologies to improve
this technology the flexibility of virtual reality
crime scene investigation by combining differ-
can be used and is grounded in physical reality
ent technologies to digitize, visualize and in-
(Azuma, 1997). Mediated reality refers to the
vestigate the crime scene. The major motiva-
ability to add to, subtract information from, or
tion for the CSI The Hague project is that one
otherwise manipulate one’s perception of real-
can investigate a crime scene only once. If you
ity through the use of a wearable computer or
do not secure all possible evidence during this
hand-held device (Mann and Barfield, 2003).
investigation, it will not be available for solving the committed crime. The digitalization
In order to reveal the current challenges for
of the crime scene provides opportunities for
supporting spatial analysis in crime scene
testing hypotheses and witness statements,
investigation, structured interviews with five
but can also be used to train future investi-
international experts in the area of 3D crime
gators. For the CSI The Hague project, two
scene reconstruction were conducted. The
groups at the Delft University of Technology,
interviews showed a particular interest for
Systems Engineering2 and BioMechanical En-
current challenges in spatial reconstruction
gineering3, joined their efforts to explore the
and the interaction with the reconstruction
potential of mediated and Augmented Reality
data. The identified challenges are:
This article is based upon (Poelman et al., 2012). http://www.sk.tbm.tudelft.nl 3 http://3me.tudelft.nl/en/about-the-faculty/departments/biomechanical-engineering/research/dbl-delft-biorobotics-lab/people/ 1
2
72
Figure 1. MEDIATED REALITY HEAD MOUNTED DEVICE IN USE DURING THE EXPERIMENT IN THE DUTCH FORENSIC FIELD LAB.
■ Time needed for reconstruction: data cap-
tion: The hands of the CSIs have to be free to
ture, alignment, data clean-up, geometric
physically interact with the crime scene when
modelling and analyses are manual steps.
needed, e.g. to secure evidence, open doors,
■ Expertise required to deploy dedicated soft-
climb, etc. Additional hardware such as data
ware and secure evidence at the crime scene. ■ Complexity: Situations differ significantly. ■ Time freeze: Data capture is often conducted once after a scene has been contaminated.
gloves or physically touching an interface such as a mobile device is not acceptable. ■ Remote connection to and collaboration with experts: Expert crime scene investigators are a scarce resource and are not often available
The interview sessions ended with an open
at location on request. Setting up a remote
discussion on how mediated reality can support
connection to guide a novice investigator
crime scene investigation in the future. Based
through the crime scene and to collaborative-
on these open discussions, the following require-
ly analyze the crime scene has the potential
ments for a mediated reality system that is to
to improve the investigation quality.
support crime scene investigation were identi-
To address the above requirements, a novel
fied:
mediated reality system for collaborative spatial
■ Lightweight head-mounted display (HMD):
analysis on location has been designed, devel-
It became clear that the investigators whom
oped and evaluated together with experts in the
arrive first on the crime scene currently carry
field and the NFI. This system supports collabo-
a digital camera. Weight and ease of use are
ration between crime scene investigators (CSIs)
important design criteria. Experts would like
on location who wear a HMD (see Figure 1) and
those close to a pair of glasses.
expert colleagues at a distance.
■ Contactless augmentation alignment (no markers on the crime scene): The first
The mediated reality system builds a 3D map of
investigator who arrives on a crime scene has
the environment in real-time, allows remote users
to keep the crime scene as untouched as pos-
to virtually join and interact together in shared
sible. Technology that involves preparing the
Augmented space with the wearer of the HMD,
scene is therefore unacceptable.
and uses bare hand gestures to operate the 3D
■ Bare hands gestures for user interface opera-
multi-touch user interface. The resulting medi-
73
ated reality system supports a lightweight headmounted display (HMD), contactless augmentation alignment, and a remote connection to and collaboration with expert crime scene investigators. The video see-through of a modified Carl Zeiss Cinemizer OLED (cf. Figure 2) for displaying content fulfills the requirement for a lightweight HMD, as its total weight is ~180 grams. Two Microsoft HD-5000 webcams are stripped and mounted in front of the Cinemizer providing a full stereoscopic 720p resolution pipeline. Both cameras record at ~30hz in 720p, images are Figure 2. HEAD MOUNTED DISPLAY, MODIFIED CINEMIZER OLED (CARL ZEISS) WITH TWO MICROSOFT HD-5000 WEBCAMS.
projected in our engine, and render 720p stereoscopic images to the Cinemizer. As for all mediated reality systems, robust realtime pose estimation is one of the most crucial parts, as the 3D pose of the camera in the physical world is needed to render virtual objects correctly on required positions. We use a heavily modified version of PTAM (Parallel Tracking and Mapping) (Klein and Murray, 2007), in which a single camera setup is replaced by a stereo camera setup using 3D natural feature matching and estimation based on natural features. Using this algorithm, a sparse metric map (cf. Figure 3) of the environment is created. This sparse metric map can be used for pose estimation in our Augmented Reality system. In addition to the sparse metric map, a dense
Figure 3. SPARSE 3D FEATURE MAP GENERATED BY THE POSE ESTIMATION MODULE.
3D map of the crime scene is created. The dense metric map provides a detailed copy of the crime scene enabling detailed analysis and is created from a continuous stream of disparity maps that are generated while the user moves around the scene. Each new disparity map is registered (combined) using the pose information from the PE module to construct or extend the 3D map of the scene. The point clouds are used for occlusion and collision checks, and for snapping digital objects to physical locations. By using an innovative hand tracking system, the mediated reality system can recognize bare hands gestures for user interface operation. This hand tracking system utilizes the stereo
Figure 4. GRAPHICAL USER INTERFACE OPTIONS MENU.
74
camera rig to detect the hand movements in 3D.
The cameras are part of the HMD and an adap-
activity could possibly help to overcome this is-
tive algorithm has been designed to determine
sue. Whether traditional patterns for computer-
whether to rely on the color, disparity or on both
mediated interaction (Schümmer and Lukosch,
depending on the lighting conditions. This is the
2007) support awareness in mediated reality
core technology to fulfill the requirement of
or rather new forms of awareness need to be
bare hand interfacing. The user interface and
designed, will be the subject of future research.
the virtual scene are general-purpose parts of
Further tasks for future research include the
the mediated reality system. They can be used
design and evaluation of alternative interaction
for CSI, but also for any other mediated reality
possibilities, e.g. by using physical objects that
application. The tool set, however, needs to be
are readily available in the environment, sensor
tailored for the application domain. The current
fusion, image feeds from spectral cameras or
mediated reality system supports the following
previously recorded laser scans, to provide more
tasks for CSIs: recording the scene, placing tags,
situational awareness and the privacy, security
loading 3D models, bullet trajectories and plac-
and validity of captured data. Finally, though
ing restricted area ribbons. Figure 4 shows the
IT is being tested and used for educational
corresponding menu attached to a user’s hand.
purposes within the CSI Lab of the Netherlands Forensic Institute (NFI), only the application and
The mediated reality system has been evalu-
test of the mediated reality system in real set-
ated on a staged crime scene at the NFI’s Lab
tings can show the added value for crime scene
with three observers, one expert and one
investigation.
layman with only limited background in CSI. Within the experiment the layman, facilitated
REFERENCES
by the expert, conducted three spatial tasks,
■ R. Azuma, A Survey of Augmented Reality,
i.e. tagging a specific part of the scene with information tags, using barrier tape and poles
Presence 6, Vol 4, 1997, 355-385
■ J. Burkhardt, F. Détienne, A. Hébert, L. Perron,
to spatially secure the body in the crime scene
S. Safin, P. Leclercq, An approach to assess the
and analyzing a bullet trajectory analysis with
quality of collaboration in technology-mediated
ricochet. The experiment was analyzed along
design situation, European Conference on Cognitive
seven dimensions (Burkhardt et al., 2007): fluid-
Ergonomics: Designing beyond the Product - Under-
ity of collaboration, sustaining mutual under-
standing Activity and User Experience in Ubiquitous
standing, information exchanges for problem solving, argumentation and reaching consensus,
Environments, 2009, 1-9
■ G. Klein, D. Murray, Parallel Tracking and Map-
task and time management, cooperative ori-
ping for Small AR Workspaces, Proc. International
entation, and individual task orientation. The
Symposium on Mixed and Augmented Reality, 2007,
results show that the mediated reality system supports remote spatial interaction with the physical scene as well as collaboration in shared augmented space while tackling current issues in
225-234
■ S. Mann, W. Barfield, Introduction to Mediated Reality, International Journal of Human-Computer Interaction, 2003, 205-208
crime scene investigation. The results also show
■ R. Poelman, O. Akman, S. Lukosch, P. Jonker, As
that there is a need for more support to identify
if Being There: Mediated Reality for Crime Scene
whose turn it is and who wants the next turn,
Investigation, CSCW ‘12: Proceedings of the 2012
etc. Additionally, the results show the need to
ACM conference on Computer Supported Coopera-
represent the expert in the scene to increase
tive Work, ACM New York, NY, USA, 2012, 1267-1276,
the awareness and trust of working in a team and to counterbalance the feeling of being observed. Knowing the expert’s focus and current
http://dx.doi.org/10.1145/2145204.2145394
■ T. Schümmer, S. Lukosch, Patterns for ComputerMediated Interaction, John Wiley & Sons, Ltd. 2007
75
76
ON FRIDAY DECEMBER 16TH 2011 THE SYMPHONY ORCHESTRA OF THE ROYAL CONSERVATOIRE PLAYED DIE WALKÜRE (ACT 1) BY RICHARD WAGNER, AT THE BEAUTIFUL CONCERT HALL ‘DE VEREENIGING’ IN NIJMEGEN. THE AR LAB WAS INVITED BY THE ROYAL CONSERVATOIRE TO PROVIDE VISUALS DURING THIS LIVE PERFORMANCE. TOGETHER WITH STUDENTS FROM DIFFERENT DEPARTMENTS OF THE ROYAL ACADEMY OF ART, WE DESIGNED A SCREEN CONSISTING OF 68 PIECES OF TRANSPARENT CLOTH (400X20 CM), HANGING IN FOUR LAYERS ABOVE THE ORCHESTRA. BY PROJECTING ON THIS CLOTH WE CREATED VISUALS GIVING THE ILLUSION OF DEPTH. WE CHOSE 7 LEITMOTIVS (RECURRING THEME, ASSOCIATED WITH A PARTICULAR PERSON, PLACE, OR IDEA), AND CREATED ANIMATIONS REPRESENTING THESE USING COLOUR, SHAPE AND MOVEMENT. THESE ANIMATIONS WERE PLAYED AT KEY-MOMENTS OF THE PERFORMANCE.
77
CONTRIBUTORS WIM VAN ECK Royal Academy of Art (KABK) w.vaneck@kabk.nl
Wim van Eck is the 3D animation specialist of the AR Lab. His main tasks are developing Augmented Reality projects, supporting and supervising students and creating 3d content. His interests are, among others, real-time 3d animation, game design and creative research.
JEROEN VAN ERP Fabrique jeroen@fabrique.nl
Jeroen van Erp co-founded Fabrique, a multidisciplinary design agency in which the different design disciplines (graphic, industrial, spatial and new media) are closely interwoven. As a designer he was recently involved in the flagship store of Giant Bicycles, the website for the Dutch National Ballet and the automatic passport control at Schiphol airport, among others.
PIETER JONKER Delft University of Technology P.P.Jonker@tudelft.nl
Pieter Jonker is Professor at Delft University of Technology, Faculty Mechanical, Maritime and Materials Engineering (3ME). His main interests and fields of research are: real-time embedded image processing, parallel image processing architectures, robot vision, robot learning and Augmented Reality.
YOLANDE KOLSTEE Royal Academy of Art (KABK) Y.Kolstee@kabk.nl
Yolande Kolstee is head of the AR Lab since 2006. She holds the post of Lector (Dutch for 78
r esearcher in professional universities) in the field of ‘Innovative Visualisation Techniques in higher Art Education’ for the Royal Academy of Art, The Hague.
MAARTEN LAMERS Leiden University lamers@liacs.nl
Maarten Lamers is assistant professor at the Leiden Institute of Advanced Computer Science (LIACS) and board member of the Media Technology MSc program. Specializations include social robotics, bio-hybrid computer games, scientific creativity, and models for perceptualization.
STEPHAN LUKOSCH Delft University of Technology S.g.lukosch@tudelft.nl
Stephan Lukosch is associate professor at the Delft University of Technology. His current research focuses on collaborative design and engineering in traditional as well as emerging interaction spaces such as augmented reality. In this research, he combines recent results from intelligent and context-adaptive collaboration support, collaborative storytelling for know ledge elicitation and decision-making, and design patterns for computer-mediated interaction.
FERENC MOLNÁR Photographer info@baseground.nl
Ferenc Molnár is a multimedia artist based in The Hague since 1991. In 2006 he has returned to the KABK to study photography and that’s where he started to experiment with AR. His focus is on the possibilities and on the impact of this new technology as a communication platform in our visual culture.
ROBERT PREVEL Delft University of Technology r.g.prevel@tudelft.nl
Robert Prevel is working on a PhD focusing on localisation and mapping in Augmented Reality applications at the Delft Biorobotics Lab, Delft University of Technology under the supervision of Prof.dr.ir P.P.Jonker.
HANNA SCHRAFFENBERGER Leiden University hkschraf@liacs.nl Hanna Schraffenberger works as a researcher and PhD student at the Leiden Institute of Advanced Computer Science (LIACS) and at the AR Lab in The Hague. Her research interests include interaction in interactive art and (non-visual) Augmented Reality.
ESMÉ VAHRMEIJER Royal Academy of Art (KABK) e.vahrmeijer@kabk.nl
Context” lab that focuses on blend between bits and atoms for design and creativity. Co-founder and lead of the minor on advanced prototyping programme and editor of the International Journal of Interactive Design, Engineering and Manufacturing.
SPECIAL THANKS We would like to thank Reba Wesdorp, Edwin van der Heide, Tama McGlinn, Ronald Poelman, Karolina Sobecka, Klaas A. Mulder, Joachim Rotteveel and last but not least the Stichting Innovatie Alliantie (SIA) and the RAAK (Regionale Aandacht en Actie voor Kenniscirculatie) initiative of the Dutch Ministry of Education, Culture and Science.
NEXT ISSUE The next issue of AR[t] will be out in October 2012.
Esmé Vahrmeijer is graphic designer and webmaster of the AR Lab. Besides her work at the AR Lab, she is a part time student at the Royal Academy of Art (KABK) and runs her own graphic design studio Ooxo. Her interests are in graphic design, typography, web design, photography and education.
JOUKE VERLINDEN Delft University of Technology j.c.verlinden@tudelft.nl
Jouke Verlinden is assistant professor at the section of computer aided design engineering at the Faculty of Industrial Design Engineering. With a background in virtual reality and interaction design, he leads the “Augmented Matter in 79