You are on page 1of 9

review articles

doi:10.1145/ 2580723.2580730
information.33 Since the 1990s, AR as a
AR systems pose potential security research area has focused on overcom-
ing challenges with display technology,
concerns that should be addressed tracking, and registration to properly
before the systems become widespread. align virtual and real objects, user in-
terfaces and human factors, auxiliary
by Franziska Roesner, Tadayoshi Kohno, and David Molnar sensing devices, and the design of novel
AR applications.1,2,6,22,36,41

Security and
However, it is only recently that early-
generation AR technologies have begun
shipping commercially. For example,
Google recently released a limited num-

Privacy for
ber of its Google Glass heads-up glasses
for AR applications. Many other early-
generation AR applications are enabled
by the ubiquity of smartphones and oth-

Augmented
er mobile devices. Examples include the
Word Lens iPhone application—an ap-
plication that overlays translated text on
the camera’s view of foreign text—and

Reality
Layar, a geolocation-based AR platform
that allows developers to create AR lay-
ers for the world (for example, for game
playing); see Figure 1. The recent advent

Systems
of 1GHz processors, location sensors,
and high resolution, autofocusing cam-
eras in mobile phones has made these
applications possible.
In this article, we take a broad view
of the AR space, considering both di-
rect applications of AR as well the tech-
nologies necessary to support these ap-
plications. Beyond the mobile phone,
devices are becoming available that
enhance sensing, display, and data
sharing, which will enable more com-
Augm ented re ali t y ( AR)technologies promise to
key insights
enhance our perception of and interaction with
the real world. Unlike virtual reality systems, which 
Augmented reality technologies, which
are rapidly evolving and becoming
replace the real world with a simulated one, AR commercially available, will create new
challenges and opportunities for security
systems sense properties of the physical world and and privacy. These challenges can be
characterized along two axes: system
overlay computer-generated visual, audio, and haptic scope and functionality.
signals onto real-world feedback in real time. In this 
Security and privacy challenges with AR
Photo Illustration by Barry Downa rd

technologies include conflicts among


article, we consider the security and privacy concerns applications sharing input and output
associated with AR systems themselves as well as devices, as well as more complex access
control for sensor data. While some
those that arise from the supporting technologies. issues can be addressed by adapting
existing solutions for smartphones,
Researchers have explored the idea of AR since the others will require novel approaches.

1960s, when Ivan Sutherland described a transparent 


AR technologies provide an opportunity
to address existing security and privacy
head-mounted display showing three-dimensional challenges in new ways.

88 com municatio ns o f th e ac m | apr i l 201 4 | vo l . 5 7 | no. 4


apr i l 2 0 1 4 | vo l. 57 | no. 4 | c om m u n ic at ion s of t he acm 89
review articles

Figure 1. Phone-based augmented reality. On the left, a picture of Word Lens, an iPhone cles—such as handling conflicts be-
application that provides seamless “in-picture” translation (source: http://www.flickr.com/ tween multiple applications sharing
photos/neven/5269418871/). Here the app translates the word “craft” from English to an AR system’s output—that are simul-
Spanish and then back again. On the right, a picture of Layar, an “augmented reality taneously intellectually challenging
browser” shipping on Android phones (source: http://site.layar.com/company/blog/
make-your-ownlayar-screen-shot-with-the-dreamcatcher/).
yet surmountable. Other challenges,
such as access control for data, are well
known in other arenas but become
even more important for AR technolo-
gies with their always-on, always-sens-
ing inputs. Given the future impor-
tance of AR technologies, researchers
already tackling these issues in other
domains can find value in refocusing
their attention on AR applications.
In addition to presenting new chal-
lenges, AR systems present opportuni-
ties for new applications that improve
security and privacy. For example,
these technologies can provide person-
Figure 2. Wearable input and output. On the left, a Looxcie body-worn camera worn by al digital views of content on personal
a ranger in Kenya (source: http://looxcie.com/index.php/image-gallery). On the right, displays. Imagine a password manager
a Google Glass prototype in June 2012 (source: http://www.flickr.com/photos/azugal-
that superimposes visual indicators
dia/7457645618).
over the correct keys for a complex
password when a user looks at a key-
board, or an application that alerts the
user when someone is lying.
In this article, we explore new se-
curity and privacy challenges pre-
sented by AR technologies, defensive
directions, and new applications of
AR systems to known security and
privacy issues.

Challenges
The AR applications and technologies
plex AR systems. For example, Loox- may also bring unforeseen computer we consider may have any or all of the
cie—an over-the-ear, always-on video security and privacy risks. Previous re- following characteristics, in addition
camera—includes a feature enabling search in the AR space has rarely con- to the traditional definition of aligning
wearers to share their live video feed sidered these issues. Rather than wait real and virtual objects in real time:
with anyone else in the world. Micro- for these technologies to fully mature ˲˲ A complex set of input devices and
soft’s SDK for Kinect,20 which provides and then retroactively try to develop se- sensors that are always on (for exam-
accurate motion sensing by combining curity and privacy safeguards, we argue ple, camera, GPS, microphone).
an RGB camera, a depth camera, and a that now is the time to consider secu- ˲˲ Multiple output devices (for exam-
multi-array microphone, has enabled rity and privacy issues, while the tech- ple, display, earpiece).
numerous prototype AR applications. nologies are still young and malleable. ˲˲ A platform that can run multiple
In addition to Google Glass, transpar- To guide this process, we ask the fol- applications simultaneously.
ent, wearable displays are now avail- lowing questions: What new security ˲˲ The ability to communicate wire-
able for research purposes from sev- and privacy research challenges arise lessly with other AR systems.
eral companies, such as Vuzix, Lumus, with AR systems and the technologies Here, we present a set of security
and Meta SpaceGlasses. Figure 2 shows that support them? What novel oppor- and privacy challenges that come with
examples of such input and output de- tunities do AR technologies create for these novel technologies and their ap-
vices. (Table 1 offers a summary of AR- improving security and privacy? plications, as summarized in Table 2.
enabling technologies; many of these We find that AR technologies form We organize these challenges along
technologies are shipping today, while an important, new, and fertile play- two axes: system scope and function-
others are still experimental.) ground for computer security and pri- ality. On one axis, we consider AR
These technologies will enable com- vacy research and industry. Of course, systems of increasing scope: single
mercial AR applications and are at the these technologies should leverage applications, multiple applications
cusp of significant innovation, which standard security best practices, such within a single AR platform, and mul-
will bring significant benefits to many as on-device and network encryption. tiple communicating AR systems. The
users. However, these technologies Nevertheless, we find unique obsta- challenges in each category first ap-

90 communicatio ns o f th e ac m | apr i l 201 4 | vo l . 5 7 | no. 4


review articles

pear at that level of system complexity. vices verifiably turned off. systems should be aware of their neces-
For each scope, we further categorize In the near term, removing the system sity in this new context.
challenges as related to input, output, is a simple way to achieve this return to Data access. To provide their intend-
or data access. We encourage future reality. However, future wearable sys- ed functionality, AR applications may
designers of AR technologies to con- tems may be difficult or impossible for require access to a variety of sensor
sider security and privacy challenges users to remove (for example, contact data, including video and audio feeds,
along both axes. lenses23 or implanted devices), and to- GPS data, temperature, accelerometer
Readers familiar with smartphone day’s non-wearable systems may already readings, and more. As in desktop and
security may observe some overlap be difficult for users to evade. For exam- smartphone operating systems, an im-
between those challenges and the set ple, several automotive manufacturers portant challenge for AR systems will
we present here. We note that some have produced windshields that display be to balance the access required for
techniques from smartphone security augmented content over the user’s view functionality with the risk of an appli-
may be applicable to AR technologies; of the road.5 In these cases, the system cation stealing data or misusing that
others will need to be rethought in this should have a trusted path for the user access. For example, a malicious appli-
new context. to return to reality, analogous to Ctrl-Alt- cation may leak the user’s location or
Challenges with single applica- Del on Windows computers. Determin- video feed to its backend servers. The
tions. We first consider threats and ing the best such sequence, or the right existing proof-of-concept PlaceRaider
challenges limited in scope to a single input mode (for example, gestures or attack34 shows that smartphone sen-
AR application. speech), requires research for each AR sors can be used to gather enough in-
Output. Users must place significant system. Another approach may be to re- formation to create three-dimensional
trust in AR applications that overlay serve a trusted region of the display that models of indoor environments.
real-world visual, auditory, or haptic always shows the real world. Unlike most of today’s desktop and
perceptions with virtual feedback. De- Input. AR applications will undoubt- smartphone applications, complex AR
vices providing immersive feedback edly face similar input validation and applications will require rich, always-
can be used by malicious applications sanitization challenges as convention- on sensing. For example, an applica-
to deceive users about the real world. al applications. For example, a trans- tion that automatically detects and
For example, a future malicious ap- lation application that parses text in scans QR codes requires constant ac-
plication might overlay an incorrect the real world may be exploited by ma- cess to video stream data, as does an
speed limit on top of a real speed limit liciously crafted text on a sign. Tradi- application that automatically detects
sign (or place a fake sign where there is tional input validation techniques are when the user is entering a password
none), or intentionally provide an in- likely to apply, but the designers of AR on another device and provides pass-
correct translation for real-world text
in a foreign language. More generally, Table 1. Summary of commercial and emerging AR technologies.
such an application can trick users into
falsely believing that certain objects Commercially Available Today Experimentally Only
are or are not present in the real world. Sensors Body-worn RGB cameras Haptic sensors29
Malicious applications can use sim- (Inputs) GPS (error of 5 meters or more)
ilar techniques to cause sensory over- Accurate motion sensing (for example, Kinect)
load for users. By flashing bright lights Feedback Opaque near-eye display Transparent near-eye display
in the display, playing loud sounds, or (Outputs) Phone display/speaker Embedded displays (contact lenses23)
Invisible Bluetooth earpiece Haptic feedback17
delivering intense haptic feedback, ap-
Services Simple cloud services (photo gallery) Complex cloud services (object recognition)
plications could physically harm users. Marker-based tracking39 Markerless tracking
Such attacks are not unprecedented: Good face detection (not recognition)37 Good face recognition
attackers have targeted epilepsy fo- Expensive or cheap but inaccurate transcription Cheap accurate transcription
rums, posting messages containing Sharing Selective sharing (photos, videos, location) Automatic sharing
flashing animated gifs to trigger head-
aches or seizures.24 Emerging AR plat-
forms must consider and prevent these
types of attacks. Table 2. Security and privacy challenges for AR technologies. We categorize these
challenges by two axes: challenges related to output, input, and data access, as arise
These output attacks are more seri- in single applications, multi-application systems, and multiple interacting systems.
ous in immersive AR applications than
they are in today’s desktop or handheld
computing scenarios both because it is Single Application Multiple Applications Multiple Systems

more difficult for users to distinguish Output Deception attacks Handling conflicts Conflicting views
Overload attacks Clickjacking
virtual from real feedback and because Trusted path to reality
it may be more difficult for users to re- Input Input validation Resolving focus Aggregate input
move or shut down the system. As a last Data Access Access control for sensor data Cross-app sharing Cross-system sharing
resort for output attacks, users must be Bystander privacy
able to easily and reliably return to the
real world, that is, with all output de-

apr i l 2 0 1 4 | vo l. 57 | no. 4 | c om m u n ic at ion s of t he acm 91


review articles

word assistance (as we will discuss). As requests from other devices or the en- lead to security concerns. For example,
a result, these privacy risks are much vironment; news reports suggest that a malicious application might try to
greater than in conventional systems. Apple has considered adding such a ca- obscure content presented by another
AR systems should take approach- pability to the iPhone to prevent video- application (for example, visually or
es that limit these risks. For example, taping of live events, such as concerts.4 aurally covering up a correct transla-
individual applications will likely not Cameras may also alert bystanders tion with an incorrect one).
need access to all sensor data. Perhaps while recording, such as by flashing a Nevertheless, output sharing will
an application only requires access to light36 or by providing access to more be necessary to provide desirable func-
a portion of the screen when the user complex policy information.19 tionality in AR systems. For example, a
is in a certain location, or only needs The CVDazzle project10 pursues a user may wish to simultaneously view
to know about certain objects the sys- different approach—using makeup to content overlaid on their view of real-
tem recognizes (for example, via the confuse face detection algorithms— ity from multiple applications, such as
Kinect’s skeleton recognizer), rather that provides privacy without compli- directions supplied by a maps appli-
than needing access to the entire raw ant cameras. The key limitation is that cation, a social feed summarizing the
camera feed. AR system designers CVDazzle is painstakingly hand-tuned activity of nearby friends, the track cur-
must consider the appropriate granu- for one particular face detection algo- rently playing in a music application,
larity for these permissions, and the rithm. A research question is to find and so on. Thus, the naive solution, in
design of usable permission manage- a general algorithm for synthesizing which only one application controls
ment interfaces will be important. makeup that fools face detection. the display at a time (as in Android to-
Existing manifest or prompt-based Challenges with multiple applica- day, for instance), is insufficient.
solutions as used in smartphones are tions. Though AR applications are Thus, future AR systems must han-
unlikely to scale in a usable way, and often conceived and prototyped in dle conflicts between multiple appli-
the long-term (rather than one-time) isolation, we can expect future AR plat- cations attempting to produce output.
data access needs of AR applications forms, like those built on Google Glass For example, five applications may all
make the application of in-context ac- or the Microsoft Kinect, will support want to annotate the same object (for
cess control solutions like user-driven multiple applications running simul- example, with a translation subtitle),
access control28 not straightforward. taneously, sharing input and output and the system will need to prioritize
Always-on cameras and other sen- devices, and exposing data and APIs to them. Furthermore, it may be impor-
sors will also create a privacy risk for each other (see Figure 3). Researchers tant for users to know which content
bystanders, which Krevelen and Poel- must anticipate these developments was generated by which application—
man identify as a challenge for wide- and ensure an “operating system for for instance, whether an annotated
spread social acceptance of AR.36 By- augmented reality” is designed with product recommendation comes from
standers should be able to opt out of or appropriate considerations for security a friend or an advertiser. AR system
be anonymized (for example, blurred) and privacy. designers must create interfaces that
in the recordings of others; prior work Output. In a multi-application AR make the origins of displayed content
has examined such issues.9,31 AR users system, applications will share output clear to or easily discoverable by users.
may need methods to prove to skepti- devices, including displays, audio out- Traditional attacks based on the
cal bystanders that such safeguards are put, and haptic feedback. Conflicts manipulation of output may require
in place. Legislation or market forces among multiple applications attempt- new approaches or new formulations
may lead to cameras that respond to ing to use these output devices can in the AR context. For example, in to-
day’s systems, applications can mount
Figure 3. Multi-application AR. Emerging and future AR platforms will support multiple clickjacking attacks that trick users
applications running simultaneously, sharing input and output devices, and exposing data
and APIs to each other. In a multi-application AR system, applications like those depicted
into clicking on sensitive user inter-
in this mockup will share output devices, including displays, audio output, and haptic face elements from another applica-
feedback. Conflicts among these applications can result in security concerns. tion (for example, to post something
on the user’s social media profile).
These attacks generally work either by
manipulating the display of the sensi-
tive element—by making it transpar-
ent or partially obscuring it in a clever
way—or by suddenly displaying sensi-
tive elements just before users click
in a predictable place. Future appli-
cations on AR systems may develop
new techniques for tricking users into
interacting with elements, and sys-
tem designers must anticipate these
threats. For example, an AR applica-
tion could attempt to trick a user into
interacting with an object in the physi-

92 comm unicatio ns o f the ac m | apr i l 201 4 | vo l . 5 7 | no. 4


review articles

cal, rather than the virtual, world. and thus have access control implica-
Input. Users will likely not interact tions. A long line of work in desktop
with AR systems using traditional in- and smartphone systems has attempt-
put methods like clicking on a mouse ed to map user actions to application
or even using a touchscreen. Instead,
users may increasingly interact with We argue that privileges (examples include Miller21
and Roesner et al.28); AR systems will
these systems using subtle input to now is the time need to evolve new user gestures to in-
haptic sensors (for example, embed-
ded in gloves), using voice, or with the to consider AR dicate sharing intent. Additionally, AR
systems are unlikely to display applica-
aid of gaze-tracking technologies. With
these input techniques and multiple
security and privacy tions in labeled windows the way that
traditional desktop operating systems
running applications, it will be non- issues, while do, so new interaction paradigms are
trivial for the system to resolve which
application is in focus and should thus
the technologies needed to enable users to identify ap-
plications and indicate which applica-
receive input. are still young tion should receive shared data.
For example, today’s voice interac-
tions happen either following an explic-
and malleable. Challenges with multiple systems.
Moving beyond a single AR system run-
it user action indicating the destina- ning multiple applications, we consid-
tion application (for example, clicking er the interactions between multiple
on the “Siri” button on an iPhone) or AR systems belonging to different us-
on systems in which only one applica- ers. Prior work in AR proposes collab-
tion can ever receive voice input (for orative applications among multiple
example, on the Xbox). When multiple users of an AR system. These applica-
applications are active and might re- tions include multiplayer games,11,32,40
ceive voice or other input at any given telepresence for remote conferenc-
time, there must be either a usable way ing,16 and face-to-face collaboration.26
for users to bring applications into fo- These types of applications pose addi-
cus, or for the system to determine the tional security and privacy challenges.
correct intended destination for input Output. Multiple users may have dif-
commands when focus is ambiguous. fering views of the world presented by
We emphasize that future AR systems their respective AR systems. For exam-
are likely to run multiple applications ple, different users may see different
simultaneously, many of them run- virtual advertisements superimposed
ning and listening for input without on real-world billboards, or different
having any visible output. Improperly users watching a presentation may be
designed focus resolution may make it shown different content based on their
easy for malicious applications to steal access levels (such as, one user may see
user input intended for another appli- top-secret footnotes while others do
cation (for example, to steal a password not). Such conflicting views will require
intended for the login box of another users to manage mental models of who
application). For example, a malicious can perceive which information, lest
application may attempt to register a they accidentally reveal private infor-
similar-sounding verbal keyword as an- mation intended only for themselves.
other, sensitive application, intention- Addressing this concern will require
ally increasing input ambiguity. innovations in interface design for aid-
Data access. As in traditional op- ing users in this task.
erating systems, AR applications will Input. A rise in the complexity of AR
likely wish to expose APIs to each systems and applications will be tightly
other, and users may wish to share coupled with a rise in the number and
virtual objects between applications. complexity of sensor inputs provided
Researchers must explore appropri- by enabling technologies. This abun-
ate access control models for cross- dance of sensor input from many users
application sharing. Certainly lessons will in turn lead to novel collaborative
from traditional access control design sensing applications, which can them-
can be applied in this space, but new selves feed data back into AR applica-
technologies and environments may tions. For example, Google already uses
require new approaches. For example, data collected by users’ smartphones
copy-and-paste and drag-and-drop are to estimate traffic conditions, which is
established user gestures for sharing then reported back to user’s phones.8
data between traditional applications This type of data is necessary to enable

apr i l 2 0 1 4 | vo l. 57 | no. 4 | c om m u n ic at ion s of t he acm 93


review articles

Figure 4. AR pipeline. AR applications (1) gather sensory data, from which they (2) extract to users and to bystanders as in the
objects with high-level semantics. Finally, they (3) render on top of the user’s senses. smartphone context—for example,
an AR system’s camera will always be
on, whereas a smartphone’s camera,
even if turned on by malware, provides
Jane TA.
much less data while the phone is in
the user’s pocket. Thus, we argue it is
“Play” important to consider full-fledged fu-
“Call” ture AR contexts when designing solu-
tions in this space.
Sensing Recognition Rendering
Along these lines, new research into
AR-specific solutions will be needed.
For example, researchers have begun
future AR applications displayed on vide a given user with a real-time 3D considering operating system support
the car’s windshield, for example. model of him or herself. specific to AR.7 AR applications—and
However, this type of aggregate in- The implicit or explicit sharing of the underlying OS—naturally follow
put can be used by malicious users to data across separate AR systems can the pipeline shown in Figure 4, so re-
fool the data collection systems. For enable valuable applications. How- search can be characterized accord-
example, a review site might leverage ever, appropriate access control mod- ingly, and different research models
location tracking to measure a restau- els and interfaces are needed to allow can assume different boundaries be-
rant’s popularity by noting the average users to manage this sharing. Today, tween the application and the OS. In
number of people present during the users already have difficulty forming the first stage, sensing, an application
day. A canny restaurateur may then pay mental models of their privacy settings (or the OS) gathers raw sensory data
people to stand in the restaurant with- on services like Facebook because of such as audio, video, or radio waves;
out buying anything. The restaurant’s the complexity of relationships be- research here includes limiting which
measured popularity rises but has no tween people and data items.18 The sensed information is collected (for
relationship to its quality. vast amount of data collected by AR example, polite cameras9,31) or limit-
AR technologies that constantly systems and the integration of virtual ing its use (for example, retention poli-
collect data will drive the adoption of objects with the real world will make cies). Second, in the recognition stage,
such collaborative sensing applica- this problem only more difficult. machine-learning algorithms extract
tions; thus, these security concerns objects with high-level semantics: as
will increase in importance. As another Defensive Directions an example, the figure shows a Kinect
example, the Community Seismic Net- Here, we outline several defensive di- skeleton, a face, the associated name,
work aggregates accelerometer sensor rections for AR technologies. First, and voice command triggers. Related
data of many individuals to detect and some of the security and privacy chal- research includes changing objects
predict earthquakes; an attacker could lenges associated with AR technolo- to cause false negatives (for example,
manipulate the sensors to “spoof” un- gies are similar to those faced by CVDazzle10) and policies governing ap-
usual seismic activity, for example, by smartphones today, such as the privacy plication access to objects.15 Finally,
encouraging many individuals moni- of sensor data and cross-application the application (or the OS) renders on
tored by the project to jump at once in sharing. In some cases, an appropriate top of the user’s senses, such as vision
the context of an unrelated game. (For defensive direction for AR is to adapt and hearing. Research here includes
example, Improv Everything13 asks us- smartphone solutions. For example, uncovering invariants that must be re-
ers to play provided audio files as a des- permission manifests and the app spected to avoid harming the user and
ignated time and follow the audio in- store review process may be adopted in building a performant “trusted render-
structions.) Trusted sensors30—while the short term. er” that respects these invariants.
important to prevent other attacks—do In the long term, however, there Not all AR defensive directions will
not help in these cases, as real-world are several reasons that approaches in consist of technical solutions. Some
conditions are manipulated. the AR context must differ from smart- challenges may call for social, policy,
Data access. In addition to show- phone solutions. First, an analysis of or legal approaches; for example, po-
ing different content to different us- the resource needs of smartphone ap- tential policies for bystander opt-outs
ers, communicating AR systems will plications28 showed that most require and compliant cameras, as discussed
allow users to share virtual content only one-time or short-term access to earlier. Other issues will similarly ben-
with each other. For example, one most resources, making solutions that efit from non-technical approaches.
user may create a virtual document require in-context user interactions Finally, we call for an AR testbed
within their private AR system and (such as user-driven access control28) for researchers working in this space.
later choose to share its display with feasible. By contrast, AR applications Most experimental AR applications
the systems of other users. Some shar- will require long-term or permanent today rely on the Microsoft Kinect
ing may even be implicit; imagine an access to sensor data at a scale beyond or smartphone platforms like Layar;
AR system that automatically uses the smartphone applications. Further, both involve only single applications
camera feeds of nearby users to pro- AR resource access will not be as clear running at one time, thereby hiding

94 comm unicatio ns o f the ac m | apr i l 201 4 | vo l . 5 7 | no. 4


review articles

challenges that arise as AR systems in- servers.) Precursors of such a system ing the voice command “OK Glass,
crease in complexity. are possible today using smartphones find password.” If the user has previ-
and 2D barcodes that encode URLs to ously stored a password for that web-
Novel Applications data with appropriate access control; site, the application displays the pass-
Though AR technologies create im- augmented heads-up displays will pre- word; otherwise, the user can enroll a
portant security and privacy concerns, vent the need for a manual scan. new password by asking the Chrome
there is also an unexplored opportu- AR systems can also act as an en- extension to generate an enrollment
nity for them to enhance security and hanced password manager for users, QR code and asking the Glass to store
privacy through their application to ex- presenting passwords or password the new password using the “enroll
isting problems. Here, we consider op- hints via the personal display. For ex- password” voice command. We have
portunities for new security and priva- ample, a display could outline the ap- made the code for our prototype avail-
cy enhancing applications enabled by propriate characters the user must able at https://github.com/froeschele/
AR technologies and systems. Our list enter on legacy devices like ATM PIN GlassPass.
is undoubtedly incomplete; we hope to pads. Users could then be assigned By designing the QR code displayed
see rich future work in this area. strong passwords, as they would never by the browser extension to include a
Leveraging personal views. AR sys- need to actually remember them. This secret shared between the browser and
tems that integrate heads-up or other application requires markerless track- the phone, this application could fur-
personal displays can leverage these ing and a system design that properly thermore serve as phishing protection,
personal views to address existing se- protects the stored passwords. as websites would not be able to cre-
curity and privacy concerns—in par- As a concrete example, we have ate and display forged QR codes that
ticular, protecting private data and im- implemented a prototype password would map to legitimate passwords in
proving password management. manager application consisting of a the password manager.
Personal displays present a strong Google Glass application and a brows- Leveraging complex sensor systems.
defense against shoulder surfing, as er (Chrome) extension (see Figure 5). AR systems benefit from the combina-
users may interact with applications The Chrome extension modifies the tion of multiple input and sensing
visible in their own view only. For ex- browser’s user interface to display devices, which can be combined to
ample, someone using a laptop on a QR code representing the website enhance digital and physical security,
an airplane today exposes everything currently displayed to the user (the privacy, and safety.
they view and type to their seat neigh- website in the browser’s address bar). Future systems can leverage AR
bors, and researchers have demon- Users can ask the Google Glass ap- technologies to detect privacy or security
strated that footage from low-cost plication to scan these QR codes and conditions of which the user should be
cameras can be used to reconstruct a consult its password database by us- alerted. For example, rather than rely-
user’s typing on a virtual mobile key-
board.25 A personal heads-up display Figure 5. Prototype AR Password Manager. Our Chrome extension (background) displays
a QR code representing the current website. In response to the voice command “find
combined with a haptic sensor for password,” our Google Glass application (foreground) scans this QR code and displays
discreet input would allow for greatly the stored password for that site privately in the heads-up display.
improved privacy.a
Personal displays further enable
encrypted content in the real world
that can only be decrypted by the AR
systems of the intended recipients.
For example, a company can post en-
crypted notices on a bulletin board
that employees can read through their
company-issued AR systems, but that
visitors to the company’s building can-
not. (Storing only the key, not the en-
crypted content, on a server accessible
by the AR system requires adversaries
to find the physical notice rather than
simply compromise the company’s

a We observe that see-through displays, such


as that used by Google Glass, may not be fully
private from external observers. For example,
images taken of the display using a telephoto
lens may be used to reconstruct the content
of the screen, similar to reconstructing con-
tent from screen reflections.3 Future research
should fully characterize this threat and de-
sign appropriate defenses.

apr i l 2 0 1 4 | vo l. 57 | no. 4 | c om m u n ic at ion s of t he acm 95


review articles

ing on compliant cameras to shield nologies to create novel privacy- and 20. Microsoft. Kinect for Windows, 2012; http://www.
microsoft.com/en-us/kinectforwindows/.
users from unwanted recording, a sys- security-enhancing applications. 21. Miller, M.S. Robust Composition: Towards a Unified
tem could alert users when it detects Approach to Access Control and Concurrency Control. Ph.D.
thesis. Johns Hopkins University, Baltimore, MD, 2006.
camera lenses pointed at them, using Acknowledgments 22. Papagiannakis, G., Singh, G. and Magnenatthalmann,
(for instance) computer vision to detect This work is supported in part by the Na- N. A survey of mobile and wireless technologies for
augmented reality systems. Computer Animation and
the glint of light reflecting off a lens.35 It tional Science Foundation (Grants CNS- Virtual Worlds 19 (2008), 3–22.
could also detect some forms of eaves- 0846065, CNS-0905384, and a Graduate 23. Parviz, B. For your eye only. IEEE Spectrum 46 (2009),
36–41.
dropping, for example, a laser micro- Research Fellowship under Grant DGE- 24. Poulsen, K. Hackers assault epilepsy patients via
computer. Wired (2008); http://www.wired.com/
phone pointed at a window. 0718124) and by a Microsoft Research politics/security/news/2008/03/epilepsy.
Such systems could also detect physi- Fellowship. We thank Luis Ceze, Lydia 25. Raguram, R., White, A.M., Goswami, D. et al. iSpy:
automatic reconstruction of typed input from
cal deception attempts. For example, an Chilton, Alexei Czeskis, Nicki Dell, Ta- compromising reflections. In Proceedings of the 18th
AR system could estimate the size and mara Denning, Karl Koscher, Brandon ACM Conf. Computer and Communications Security.
26. Reitmayr, G. and Schmalstieg, D. Mobile collaborative
shape of an ATM card slot, then issue a Lucia, Alex Moshchuk, Bryan Parno, augmented reality. In Proceedings of the 4th
warning if it appears a card-skimming Karin Strauss Helen Wang, and anony- International Symp. on Augmented Reality (2001).
27. Riva, O., Qin, C., Strauss, K., And Lymberopoulos,
device has been added. Similarly, exist- mous reviewers. D. Progressive authentication: Deciding when to
ing work on computerized interpreta- authenticate on mobile phones. In Proceedings of the
21st USENIX Security Symposium (2012).
tion of facial expressions12 could be ap- References 28. Roesner, F., Kohno, T., Moshchuk, A. et al. User-driven
1. Azuma, R.T. A survey of augmented reality. Presence:
plied to behavior-based lie detection.38 Teleoperators and Virtual Environments 6 (1997),
access control: Rethinking permission granting in
modern operating systems. IEEE Symposium on
One of our colleagues refers to this ap- 355–385. Security and Privacy (2012).
2. Azuma, R., Baillot, Y., Behringer, R., Feiner, S., Julier, 29. Saponas, T.S., Tan, D.S., Morris, D. et al. Enabling
plication as “spidey sense.” S. and Macintyre, B. Recent advances in augmented always-available input with muscle-computer
Beyond storing passwords, AR sys- reality. IEEE Computer Graphics and Applications 21, interfaces. In Proceedings of the 22nd ACM Symposium
6 (2001), 34–47. on User Interface Software and Technology (2009).
tems can be used for implicit authen- 3. Backes, M., Chen, T., Duermuth, M., Lensch, H. 30. Saroiu, S. and Wolman, A. I am a sensor, and I
tication of their users. The plethora of and Welk, M. Tempest in a teapot: Compromising approve this message. In Proceedings of the 11th
reflections revisited. IEEE Symposium on Security and Workshop on Mobile Computing Systems and
sensors attached to people using these Privacy (2009). Applications (2010), ACM.
technologies can be used to authenti- 4. Business Insider. This apple patent will shut 31. Schiff, J., Meingast, M., Mulligan, D.K., Sastry, S. and
down your camera at live concerts; http://www. Goldberg, K.Y. Respectful cameras: Detecting visual
cate them with biometric and behav- businessinsider.com/iphone-concert-patent-2011-6. markers in real-time to address privacy concerns. In
ioral characteristics. Prior work has 5. CNN. Augmented-reality windshields and the future Proceeding of the 2007 Int’l Conference on Intelligent
of driving, 2012; http://virtual.vtt.fi/virtual/proj2/ Robots and Systems.
examined the possibility of such mech- multimedia/alvar.html. 32. Starner, T., Leibe, B., Singletary, B. and Pair, J.
6. Costanza, E., Kunz, A. and Fjeld, M. Mixed reality: A
anisms on mobile phones14,27 AR sys- survey. Human Machine Interaction. Springer-Verlag,
Mindwarping: Towards creating a compelling
collaborative augmented reality game. In ACM
tems would provide far more powerful 2009, 47–68. Intelligent User Interfaces (2000).
7. D’Antoni, L., Dunn, A., Jana, S., et al. Operating
authentication. Similarly, sensor data system support for augmented reality applications.
33. Sutherland, I.E. A head-mounted three-dimensional
display. In Proceedings of the Fall Joint Computer
could be used to help with authoriza- In Proceedings of USENIX Workshop on Hot Topics in Conference, American Federation of Information
Operating Systems (2013).
tion and access control decisions. 8. Google. Crowdsourcing road congestion data; http://
Processing Societies (1968).
34. Templeman, R., Rahman, Z., Crandall, D.J., And
Beyond the sensors attached to an googleblog.blogspot.com/2009/08/bright-side-of- Kapadia, A. Placeraider: Virtual theft in physical spaces
sitting-in-traffic.html. with smartphones. CoRR abs/1209.5982 (2012).
individual (for example, Alice), the sen- 9. Halderman, J.A., Waters, B. and Felten, E.W. Privacy 35. Truong, K., Patel, S., Summet, J. and Abowd, G.
sors of bystanders could also be used management for portable recording devices. In Preventing camera recording by designing a capture-
Proceedings of the 3rd ACM Workshop on Privacy in resistant environment. Proceedings of Ubicomp (2005).
to authenticate her by providing the Electronic Society (2004). 36. Van Krevelen, D. and Poelman, R. A survey of
authentication system with third-party 10. Harvey, A. CVDazzle: Camouflage from Computer augmented reality technologies, applications, and
Vision; http://cvdazzle.com/. limitations. The International Journal of Virtual
visual, audio, and other sensory views 11. Henrysson, A., Billinghurst, M., And Ollila, M. Face to Reality 9 (2010), 1–20.
of Alice. This third-party authentication face collaborative AR on mobile phones. In Proceeding 37. Viola, P. and Jones, M. Robust real-time object
of the 4th IEEE/ACM International Symposium on detection. International Journal of Computer Vision 57
system would distribute trust to sys- Mixed & Augmented Reality (2005). (2004), 137–154, Hingham, MA.
tems and persons with no incentive to 12. Hoque, M.E., McDuff, D. and Picard, R.W. Exploring 38. Vrij, A., Edward, K., Roberts, K. and Bull, R. Detecting
temporal patterns in classifying frustrated and deceit via analysis of verbal and nonverbal behavior.
falsely authenticate Alice. delighted smiles. IEEE Transactions on Affective Journal of Nonverbal Behavior 24 (2000), 239–263.
Computing 3 (2012), 323–334. 39. VTT Technical Research Centre Of Finland. Alvar
13. Improv Everywhere. The Mp3 Experiments, 2012; Software Library, 2009. http://cnn.com/2012/01/13/
Conclusion http://improveverywhere.com/missions/the-mp3- tech/innovation/ces-future-driving/.
experiments/.
AR systems, with their sophisticated 14. Jakobsson, M., Shi, E., Golle, P., and Chow, R. Implicit
40. Wagner, D., Pintaric, T., Ledermann, F. and
Schmalstieg, D. Towards massively multi-user
and pervasive input, output, and pro- authentication for mobile devices. In Proceedings of augmented reality on handheld devices. In
the 4th USENIX Workshop on Hot Topics in Security
cessing capabilities, have the potential (2009), USENIX.
Proceedings of the 3rd International Conference on
Pervasive Computing (2005).
to significantly benefit many users. To 15. Jana, S., Molnar, D., Moshchuk, A. et al. Enabling 41. Zhou, F., Duh, H. B.-L. and Billinghurst, M. Trends in
fine-grained permissions for augmented reality augmented reality tracking, interaction and display: A
complement ongoing innovations in applications with recognizers. Tech. Rep. MSR- review of 10 years of ISMAR. In Proceedings of the
AR technologies, we argue now is also TR-2013-11, Microsoft Research, Feb. 2013. 7th IEEE/ACM International Symposium on Mixed and
16. Kato, H. and Billinghurst, M. Marker tracking and Augmented Reality (2008).
the time to define a roadmap for pro- HMD calibration for a video-based augmented reality
tecting the computer security and pri- conferencing system. In IEEE/ACM Workshop on
Augmented Reality (1999). Franziska Roesner (franzi@cs.washington.edu) is a Ph.D.
vacy of AR systems—before these sys- 17. Laycock, S. and Day, A. A survey of haptic rendering candidate at the University of Washington, Seattle.
tems become widely deployed and their techniques. Comp. Graphics Forum. 26, 1 (2007), 50–65.
18. Madejski, M., Johnson, M. and Bellovin, S.M. The Tadayoshi Kohno (yoshi@cs.washington.edu) is an
architectures become entrenched. To Failure of Online Social Network Privacy Settings. associate professor at the University of Washington, Seattle.
catalyze this roadmap, we consider Tech. Rep. CUCS-010-11, Dept. of Comp. Science,
Columbia University, 2011. David Molnar (dmolnar@microsoft.com) is a researcher
new security and privacy challenges 19. Maganis, G., Jung, J., Kohno, T. et al. Sensor at Microsoft Research, Redmond, WA.
Tricorder: What does that sensor know about me? In
posed by these systems, and we explore Proceedings of the 12th Workshop on Mobile Computing Copyright held by Owner(s)/Author(s). Publication rights
opportunities afforded by these tech- Systems and Applications (2011), ACM. licensed to ACM. $15.00.

96 comm unicatio ns o f th e ac m | apr i l 201 4 | vo l . 5 7 | no. 4

You might also like