You are on page 1of 8

ACCEPTED FROM OPEN CALL

An Edge-Computing Based Architecture for Mobile Augmented Reality


Jinke Ren, Yinghui He, Guan Huang, Guanding Yu, Yunlong Cai, and Zhaoyang Zhang

Abstract pointlessly increasing the cost of mobile devices.


Hence, novel techniques are urgently expected to
In order to mitigate the long processing delay and tackle the above challenges.
high energy consumption of mobile augmented To overcome the computational resource
reality (AR) applications, mobile edge computing shortage of mobile devices, another major innova-
(MEC) has been recently proposed and is envi- tion in the past decade is mobile cloud computing
sioned as a promising means to deliver better (MCC), which allows users to offload computa-
Quality of Experience (QoE) for AR consumers. tion-intensive tasks to a number of powerful cloud
In this article, we first present a comprehensive servers deployed at the remote cloud platform
AR overview, including the indispensable com- for processing [3]. However, with the strict delay
ponents of general AR applications, fashionable requirement of mobile AR applications, MCC suf-
AR devices, and several existing techniques for fers from extra propagation delay due to the long
overcoming the thorny latency and energy con- physical distance between mobile devices and
sumption problems. Then we propose a novel cloud servers [4]. To this end, an improved tech-
hierarchical computation architecture by insert- nique, which is referred to as mobile edge com-
ing an edge layer between the conventional user puting (MEC), has been recently proposed by the
layer and cloud layer. Based on the proposed European Telecommunication Standards Institute
architecture, we further develop an innovative (ETSI) and is also considered by the Third Gen-
operation mechanism to improve the perfor- eration Partnership Project (3GPP) in their future
mance of mobile AR applications. Three key tech- standards. By distributing the conventional cen-
nologies are also discussed to further assist the tralized cloud computing resources to the edge
proposed AR architecture. Simulation results are of mobile networks, MEC offers an adjacent com-
finally provided to verify that our proposals can puting environment for mobile subscribers and
significantly improve latency and energy perfor- provides a variety of benefits, including ultra-low
mance as compared to existing baseline schemes. latency, real-time access, location-aware services,
and so on [5–7].
Introduction On the other hand, by taking advantage of the
As a revolutionary innovation, mobile augmented proximity to mobile users in MEC and the abun-
reality (AR) has become a miraculous technology dant computation capacity in MCC, effective col-
in the past few years. By combining computer-gen- laboration between cloud and edge computing
erated and sensor-extracted elements with real can further improve system performance. Sever-
objects and enabling real-time 3D interaction al hierarchical edge-to-cloud architectures have
between mobile users and physical surroundings been proposed to collaborate the computation
at a user’s finger tips, mobile AR is envisioned as a capacity of edge servers and cloud servers [8, 9].
new paradigm to submerge mobile subscribers in However, they are not specifically designed for
a fabulous mixed-reality world [1, 2]. In view of the AR applications. Motivated by this, in this work
great business opportunities involved in mobile AR, we propose a novel hierarchical computation
many leading companies have continuously scram- architecture to deal with the long processing
bled to design and promote their own mobile AR delay and high energy consumption of mobile AR
applications and products, such as Google Glass, applications. The newly-designed architecture is
Microsoft HoloLens, and Recon Jet. composed of three layers: the user layer, the edge
However, despite the fully immersive user layer, and the cloud layer. By integrating the com-
experience, mobile AR still faces many techni- munication, computation, and control modules in
cal challenges that hinder its widespread com- the edge layer while seamlessly collaborating the
mercial use. Due to hardware limitations, running edge and cloud computing capacities, several crit-
sophisticated AR algorithms at mobile devices ical components of AR applications can be intelli-
generally causes long processing delay and high gently offloaded to both edge and cloud servers
energy consumption, which are the main hurdles for further processing. In accordance with this
for implementation. To handle these issues, many architecture, an advanced operation mechanism
cutting-edge technologies, such as advanced hard- is developed by embedding the MEC functions
ware structure, acceptable approximate comput- into the AR process. Moreover, to support the
ing, and partial video frame updating have been proposed AR framework, we further discuss three
developed to improve mobile AR performance. key technologies, that is, joint communication and
Nevertheless, these methods still cannot catch computation resource allocation, collaborative
up with the rapid growth of user demand while cloud and edge computing, and content-based
Digital Object Identifier:
10.1109/MNET.2018.1800132 The authors are with Zhejiang University; Jinke Ren and Guanding Yu (corresponding author) are also with Xidian University.

162 0890-8044/18/$25.00 © 2018 IEEE IEEE Network • July/August 2019

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
image retrieval from the viewpoints of energy effi- Resolution Image Format
ciency and delay optimization. The main merit of
the proposed MEC-based AR framework is that it
merges the advantages of cloud computing and
Video Source Renderer
edge computing to enormously improve the qual-
ity of experience (QoE) for AR applications.
In what follows, we first provide a compre-
hensive AR overview, including the indispensable Object
components of general AR applications, fashion- Recognizer
able AR devices, and existing technologies for
alleviating the inherent long processing delay and
high energy consumption problems. To improve Tracker Mapper
system performance, our key contributions are
the novel hierarchical computation architecture
and the related AR operation mechanism as well Number of Features
as the three key technologies, which can signifi-
cantly facilitate the implementation of the MEC- FIGURE 1. Main components of general AR applications.
based AR framework. Finally, simulation results
demonstrate that the proposed AR framework can HoloLens, and Osterhout Design Group (ODG)
evidently improve energy and delay performance, R-9 operate independently, whereas Meta 2 and
as compared against some baseline schemes. Our Recon Jet need to rely on external equipment such
design in this article has great potential to provide as personal computers (PCs) or smart phones.
an essential reference for the future development The Google Glass is the first-appeared AR
of mobile AR technology. glass, which has a stylish appearance and provides
users with requested information, such as calen-
Overview of Augmented Reality dar, weather, and message via natural language
In this section, we present an overview of AR, voice commands. By comparison, Microsoft Holo-
including the specific components of general AR Lens enables users to engage with digital content
applications, the compelling mobile AR devices, and interact with holograms of the surroundings
and the incumbent AR technologies, from the to provide subscribers with a mixed-reality expe-
viewpoints of energy efficiency and latency opti- rience. This function is extended in ODG R-9 by
mization. combining with some advanced modules, such
as Bluetooth 5.0, built-in GPS, and six degrees
AR Application Components of freedom tracking. However, these devices all
To combine colorful computer-generated and suffer from high energy consumption and cannot
sensor-extracted data with physical reality, differ- work for a long time. On the other hand, by con-
ent AR applications possess diverse modules for necting powerful computers, Meta 2 allows users
specific processing purposes. However, five com- to intuitively touch, grab, and move the comput-
putation components, that is, the video source, er-generated digital objects as real ones, which
tracker, mapper, object recognizer, and renderer, inevitably leads to its non-portability for outdoor
are indispensable for almost all AR applications, use. In particular, Recon Jet is extensively used in
and they play the critical roles in general AR archi- sport scenes and can exhibit the real-time motion
tectures [10, 11]. information of mobile users.
As illustrated in Fig. 1, the aforementioned five Although the above AR devices can provide a
components collaborate closely to accomplish fully immersive experience for mobile subscribers,
an integrated AR process. First, the video source they generally suffer from long processing delay
fetches the raw videos from mobile cameras and high energy consumption, which are the
and clips these videos into frames with a specific most-critical performance metrics for AR appli-
image format, such as JPEG and PNG. Next, the cations. To overcome these drawbacks, several
video frames are delivered to the tracker to deter- existing technologies have been recently devel-
mine the user’s position with respect to the physi- oped, which will be discussed in the following.
cal surroundings. Given the tracking results, virtual
coordinate of the environment can be established Existing Technologies
by the mapper. Then, the internal objects in video Presently, there are mainly three kinds of technol-
frames are identified by the object recognizer ogies to reduce the processing delay and energy
with robust features. Afterward, the augmented consumption of AR applications, as summarized
information of the identified object can be accu- in the following [1].
rately retrieved from local memory (or cloud data- Advanced Hardware Structure: It is hard for
base) and properly mixed with the original videos traditional batteries to sustainably support the
by the renderer. The results are finally displayed complex operations of AR applications. As a
on the screens of AR devices so that the subscrib- result, advanced hardware structures such as a
ers are able to enjoy a magic interactive experi- multi-core CPU with low frequency and voltage
ence with physical reality at their fingertips. can be utilized to replace the single-core CPU.
Furthermore, dynamic voltage and frequency scal-
Fashionable AR Devices ing (DVFS) technology can be adopted to save
A variety of AR devices with different operation energy with respect to specific requirements.
modes have been recently designed with multi- Acceptable Approximate Computing: Consis-
farious functions. Table 1 summarizes several well tently performing accurate calculation in AR appli-
known AR devices developed by major com- cations usually consumes much energy and results
panies. Among them, Google Glass, Microsoft in long processing delay. Therefore, approximate

IEEE Network • July/August 2019 163

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
Operation
Product CPU GPU Latency Battery autonomy Weight
mode

Google Glass Independent 1 GHz 300 MHz 700 ms for user interface (UI) response < 1 h when recording videos 42 g

Microsoft
Independent 1.04 GHz HoloLens Graphics Few seconds for gesture recognition 2–3 h for typical use 579 g
HoloLens

ODG R-9 Independent 2.45 GHz 710 MHz Not available 24 h for browsing and casual use 184 g

Meta 2 Dependent 3.4 GHz 1127 MHz 80 ms for gesture recognition Powered by PC 500 g

Recon Jet Dependent 1 GHz PowerVR SGX540 40 s for loading navigation map 4 h for typical use 85 g
TABLE 1. Major AR devices.

computing for AR tasks can be adopted accord- for example, tracking, mapping, and object recog-
ing to detailed computational accuracy require- nition, still remain for local computing, end-to-end
ments. For example, low-precision tasks, such as latency and energy consumption cannot be further
location sensing, can be approximately processed reduced and the anticipated user demand is still far
to balance computational accuracy and energy from reach. To overcome the resource shortage of
(delay) efficiency. mobile devices, MEC has been recently proposed
Partial Video Frame Updating: The frequent and is regarded as a new paradigm to deliver a bet-
movement of mobile AR devices would produce ter experience for mobile users by offloading com-
a large volume of video data, which is adverse putation-intensive tasks to the servers deployed at
for real-time processing. Therefore, partial video the edge of mobile networks [12]. Motivated by
frame updating can be exploited to avoid redun- this, in the following section we present a novel
dant computing by processing the newly-generat- computation framework for embedding MEC into
ed data only. AR applications.
Based on the above mentioned technologies,
the end-to-end latency and energy consumption of MEC-Based AR Framework
AR applications can be improved to some extent. In this section, we first present a novel hierarchi-
However, since the computation-intensive tasks, cal computation architecture for multi-user AR
systems and then introduce a detailed mechanism
to support AR applications.
Cloud Layer Hierarchical Computation Architecture
Considering a multi-user AR scenario where sev-
eral AR applications need to be executed simul-
taneously, we propose a hierarchical architecture
Edge Layer composed of three layers, namely, the user layer,
the edge layer, and the cloud layer. As illustrated
Operation Platform Virtualized Controller in Fig. 2, multiple AR devices are located in the
user layer and are connected to the edge layer
Database Network through wireless links. The cloud layer mainly con-
Controller sists of an enormous database for data storage
and abundant computational resources for data
AR Computing Platform processing. These two layers are quite similar to
Computing System the existing AR architectures and thus we do not
Computing Graphics devote special attention henceforth.
Controller Controller
Module Module The uniqueness of our proposed architecture
is that an edge layer is inserted between the con-
ventional cloud layer and user layer. The edge
Storage
layer can be equipped at the base stations (BSs)
Edge Cache Controller
of cellular networks or the access points (APs)
of WiFi networks. The merit of this design is that
MEC can be utilized to improve end-to-end laten-
cy and energy consumption performance of AR
Communication Unit applications. Furthermore, cloud computing and
edge computing can be potentially collaborated
Radio Remote Unit for further performance enhancement.
Decoder Encoder
(RRU) In the proposed hierarchical architecture, one
major issue is how to design the function mod-
ules in the edge layer to seamlessly collaborate
with the other two layers. To tackle this problem,
we introduce three functionalized components,
User Layer namely, communication unit, operation platform,
and virtualized controller in the edge layer on the
basis of software defined network (SDN) tech-
FIGURE 2. Hierarchical Computation Architecture. nology. Note that the combination of the three

164 IEEE Network • July/August 2019

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
modules also leads to the convergence of com-
munication, computation, and control. In what
follows, we will present a detailed description of
Cloud Tier
Cloud
these three modules. Database
Communication Unit: The communication
unit can be regarded as a “bridge” that enables
real-time data transmission between the edge Searching Feature
layer and the other two layers. In one aspect, Results Matching
when multiple AR devices offload the computa-
tional tasks, such as video steams, to the BS, the Core Network
inner radio remote unit (RRU) has the responsi-
bility to successfully receive these data. Further- No Edge Tier
more, these data are required to be delivered to
the operation platform for further processing. On Yes
Annotation
the other aspect, the computation results from Storage
Find Result?
the operation platform and cloud layer should be Image Database
multicasted to the corresponding users through
the communication unit.
Operation Platform: The operation platform is Tracker- Feature Feature
the core of this architecture, which processes the Mapper Matching Extraction
offloaded AR tasks from mobile users. The origi- Computing
nal data from the RRU is first stored in the edge Data
cache, and will be delivered to the computing Storage Video
platform for further processing. The critical AR Stream
computing platform is composed of a computing Cache
Video
module (resembling a computer’s CPU) and a User Clipping
Classification
graphics module (resembling a computer’s GPU). Instruction
The former is utilized to process computing-relat- Cache
ed tasks, such as tracking and mapping, while the Edge Storage Edge Cache
latter is utilized to process graphics-related tasks
such as object recognition. Since AR tasks usually
Wireless Communication
require additional data, such as 3D-models and
annotations of the recognized objects, we estab-
lish a small database at the edge layer for storing User Tier
the object information that is frequently accessed.
Downlink Cache Uplink Cache Local Cache Renderer
With this design, we do not have to invariably
fetch the requested information from the remote
cloud database, which can significantly reduce
end-to-end latency. Camera & Sensor Display
Virtualized Controller: The virtualized con-
troller serves as the “centralized coordinator” of
the whole edge layer, which is divided into four FIGURE 3. MEC-based AR operation mechanism.
specific components: the network controller,
the computing controller, the storage controller,
and the system controller. The network control- three tiers, which are consistent with the hierarchi-
ler manages all network activities among three cal architecture, that is, the user tier, the edge tier,
layers, such as network establishment and data and the cloud tier. The main idea of this mecha-
transmission. Accordingly, the computing control- nism is that the video source and the renderer of
ler supervises the entire process in the operation AR applications must be processed locally, where-
platform while it optimally allocates the available as the computation-intensive components, that
computational resources to each AR task with is, the tracker, mapper, and object recognizer,
specific requirements. Moreover, the inherent should be offloaded to the edge and cloud serv-
executive priorities and collaborative properties of ers. The detailed operations in each tier are gen-
AR tasks from different subscribers are also evalu- eralized into specific blocks and will be described
ated in the computing controller. In addition, the in the following.
storage controller aims at properly managing the The user tier is responsible for executing the
memory mechanism of the edge database for fast local operations. Multiple AR devices simultane-
data searching and updating. Finally, the system ously start with sensing the real environment, pro-
controller monitors the behaviors of the above ducing raw videos, and capturing users’ gestures
three controllers and coordinates them in a more via their cameras and sensors. Through identifica-
efficient way. tion analysis, this original information will be trans-
formed into two categories: the video streams
MEC-Based AR Operation Mechanism which contain the raw video data and the opera-
Thus far, our discussions have focused on the tion instructions which carry the specific require-
physical modules of the proposed hierarchical ments of mobile subscribers, such as object
architecture for AR applications. In the following, identification and sentiment analysis. Thereafter,
we further present some detailed operations to this information will be further duplicated into two
facilitate the AR implementation. copies with one stored in the uplink cache for
As depicted in Fig. 3, the proposed MEC-based data transmission and the other stored in the local
AR operation mechanism can be divided into cache for subsequent processing. Consequently,

IEEE Network • July/August 2019 165

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
and edge computing, and content-based image
By combining the advantages of the powerful computation capacity of cloud computing and the retrieval, to further improve energy efficiency and
proximity to mobile subscribers of edge computing, collaborative cloud and edge computing is reduce processing latency.
envisioned as a promising paradigm to achieve better AR performance. Joint Communication and
Computation Resource Allocation
In the proposed framework, the computational
the transceivers of mobile AR devices will trans- tasks of an AR application can be jointly pro-
mit the data in the uplink cache to the edge layer cessed by the local device, the edge server, and
through wireless channels. the cloud server. Meanwhile, the communication
The edge tier plays a critical role in comput- resources of the wireless network and the com-
ing AR applications. Upon receiving the offload- putational resources of the edge/cloud servers,
ed data from AR devices, the BS will first classify can be shared by all AR devices. Therefore, joint
them into two kinds: the raw video streams and communication and computation resource allo-
the user’s operation instructions, and then deliv- cation is essential to further improve the delay
er them to the edge cache for separate storage. and energy performance [13]. Generally, two
Thereafter, these two kinds of data are delivered schemes in current AR scenarios should be taken
to the tracker-mapper module and the video clip- into account, that is, the centralized and distribut-
ping module, respectively. The tracker-mapper ed resource allocation schemes.
module then tracks users’ locations and builds In centralized AR systems, both user priority
virtual coordinates to coincide with the real world and channel state information can be acquired
via some algorithms, such as simultaneous local- in the edge layer. Therefore, the network con-
ization and mapping (SLAM) and parallel track- troller and the computing controller can collabo-
ing and mapping (PTAM). Meanwhile, the video ratively design the resource allocation policy by
clipping module slices one representative frame optimizing a specific objective function, such as
(or image) from each raw video for subsequent minimizing end-to-end latency under a prescribed
processing. It should be noted that some users resource utilization constraint and maximizing
may observe the same object and require the mobile energy efficiency under an offloading
same information of this object. With this regard, latency constraint. Moreover, the network con-
classifiers can be leveraged to sort all frames from troller and the computing controller can monitor
different videos into several categories based on the specific requirements of different AR tasks in
the inherent information, such that each category real-time and adaptively adjust the resource allo-
contains those frames of the same object. This cation to meet the diverse user demands.
function can be implemented via some well known However, in distributed AR systems, the afore-
machine learning algorithms, such as convolu- mentioned information can no longer be obtained
tional neural networks (CNN) and support vector by the edge layer, making it difficult to centrally
machines (SVM). Afterward, one typical image of control the allocation of available communication
each category is picked out and utilized to match and computational resources. To tackle this prob-
the standard images pre-stored in the edge data- lem, game-theoretical techniques can be exploit-
base by some image retrieval methods. By this ed to develop distributed algorithms based on
means, the inherent collaborative properties of AR past network and channel information, whereas
applications can be fully utilized and redundant the obtained result is demonstrated to achieve
calculation can be effectively avoided. Then, if the the Nash equilibrium and can deliver satisfactory
matched standard image is found in the edge data- latency and energy consumption experiences for
base, the corresponding annotation information mobile AR subscribers.
will be fetched from the adjacent edge storage.
Otherwise, the related frame will be offloaded to Collaborative Cloud and Edge Computing
the cloud server for further searching. On the one hand, traditional cloud servers are
The cloud tier contains a large cloud database typically deployed at the remote cloud platform,
for storing the additional data that are not cached suffering from long propagation delay when trans-
in the edge tier due to its limited memory size. mitting onerous AR tasks in the core network. On
Once the retrieval operation in the edge tier is the other hand, although the adjacent MEC serv-
not realized, the corresponding image will be ers can be implemented at the nearby BS, their
offloaded to the cloud tier for further searching. computation and storage capacities are usually
Owing to the adequate computational resources limited. Therefore, by combining the advantages
at the cloud server and the sufficient capacity of of the powerful computation capacity of cloud
the cloud storage, image retrieval will be generally computing and the proximity to mobile subscrib-
achieved. Thereafter, the requested information ers of edge computing, collaborative cloud and
will be transmitted back to the edge tier and com- edge computing is envisioned as a promising par-
bined with the virtual map that is reconstructed adigm to achieve better AR performance.
by the tracker-mapper module. Also, the total In the proposed framework, due to the limited
computation results are multicasted to the corre- edge storage capacity, additional information such
sponding devices. Finally, these data are exhibited as 3D models and annotations of those unpopular
to the subscribers after mixing with the original objects should be stored in the cloud database
videos via the local renderers. to alleviate the storage burden at the edge data-
base. On the other hand, the edge database only
Key Implementation Technologies needs to store the frequently accessed informa-
In this section, we will develop three key tech- tion of the popular objects. This can be realized
nologies, that is, joint communication and com- by storing the historical access record of each
putation resource allocation, collaborative cloud object’s information in the edge layer and setting

166 IEEE Network • July/August 2019

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
an appropriate threshold to judge which object is
popular. By this means, massive object recogni- The object recognizer is a critical component of AR applications, which usually consumes a large
tion operations can be avoided at the edge layer, processing delay, especially in the image retrieval procedure. Therefore, advanced content-based image
and AR processing latency can be effectively
reduced. Moreover, when several edge servers retrieval technology should be developed to accelerate the searching speed.
are simultaneously served by the same cloud serv-
er, it is important to properly allocate the available
cloud computational resources to each edge serv- ognition. Consequently, end-to-end latency and
er based on their workload and local computation energy consumption can be effectively reduced.
capacities. The detailed resource allocation policy
can be derived by solving a specific optimization Performance Evaluation
problem through some convex and non-convex In this section, we present simulation results to
optimization tools. For instance, more cloud com- manifest the latency and energy performance
putational resources need to be allocated to assist enhancement of our proposals against two bench-
those edge servers with less computation capaci- mark schemes. We consider a scenario where
ties and heavier workload to balance the uneven multiple mobile devices in the coverage of the
resource and workload distribution over different same BS execute AR applications simultaneous-
edge servers. ly. The edge layer is implemented at the central
BS with a radius of 200 m. The cloud layer is
Content-Based Image Retrieval deployed in a remote cloud platform. The BS con-
The object recognizer is a critical component of nects the cloud platform through a backhaul link,
AR applications, which usually consumes a large whose uplink and downlink transmission capaci-
processing delay, especially in the image retrieval ties are both set to be 200 Mb/s. The edge and
procedure. Therefore, advanced content-based cloud computing capacities are set to be 5  1010
image retrieval technology should be developed CPU cycle/s and 3  1011 CPU cycle/s, respec-
to accelerate the searching speed. Motivated by tively [12]. Each AR device is randomly distributed
the work in [14], we present a typical pipeline for according to Poisson distribution within the BS
image retrieval within the proposed mechanism, coverage, adopting TDMA channel access with
which consists of three steps: feature extraction, a bandwidth of 15 MHz. The local computation
feature matching, and geometric verification. capacity of each device follows the uniform distri-
Feature Extraction: Once an image is input bution between [5  108, 2  109] CPU cycle/s.
into the graphics module of the edge layer, the The channel gains between mobile devices and
feature extraction algorithm will immediately the BS are generated according to indepen-
search its inherent salient interest points, which dent and identically distributed (i.i.d.) Rayleigh
are used to estimate the similarity between this random variables with unit variance. Since the
image and the standard images pre-stored in the cloud computing components are first offloaded
edge and cloud databases. Various robust fea- to the edge layer through wireless links and then
ture descriptors can be applied to obtain typical transmitted to the cloud server through the back-
features, such as scale invariant feature transform haul link equipped with high transmission band-
and speeded up robust features [15]. width, the distance between mobile devices and
Feature Matching: The current pairwise fea- the cloud server could be ignored in this article.
ture matching algorithm generally consumes a lot Other major parameters of AR applications can
of time since it directly matches the input image be found in [16].
with all standard images. To cope with this issue, In our test, we compare the performance of
we pre-construct a data structure to store the fea- our proposals with two benchmark schemes: the
tures of all standard images with particular indi- local computing scheme where the tracker and
ces. Then, by comparing the extracted features mapper modules are kept for local processing
with those in the data structure, a shortlist of can- while the object recognizer is offloaded for cloud
didate images can be efficiently filtered. Thereaf- computing, and the cloud computing scheme
ter, the slow pairwise feature matching method where the tracker, mapper, and object recognizer
can be applied between the input image and the are all offloaded to the remote cloud platform.
candidate images only, and a best-matched image Meanwhile, we name our proposed scheme the
can be eventually picked out from the shortlist. In edge computing scheme.
this way, image retrieval delay can be substantially Figure 4a presents the comparative results
reduced. of the end-to-end latency of each device in the
Geometric Verification: After obtaining the three schemes, where the transmission power of
best-matched standard image, further examination each mobile device is bounded by 24 dBm. In
is required to confirm whether the matching result this simulation, since both communication and
is correct. Geometric verification is commonly computation latency increase almost linearly with
utilized to test whether the input image and the the number of devices, the end-to-end latency of
best-matched standard image are similar with only each scheme has an approximately linear trend.
geometric and photometric distortions. If the veri- From the figure, the local computing scheme
fication result is correct, the matching relation will always delivers the worst latency performance
be established. Otherwise, the input image and its among the three schemes because of the insuf-
features will be offloaded to the cloud server for ficient computational resources of mobile devic-
further searching. es. Moreover, due to the limited communication
It should be noted that with the above three resources, the end-to-end latency increases with
steps, the computation and storage resources of the number of devices in all schemes. However,
the cloud server and edge server can be jointly the proposed edge computing scheme always
utilized to improve the accuracy of object rec- achieves the best latency performance since it

IEEE Network • July/August 2019 167

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
600 0.2
Local computing scheme Local computing scheme

Total mobile energy consumption (J)


Cloud computing scheme Cloud computing scheme
500
Edge computing scheme Edge computing scheme
End-to-end latency (ms)

0.15
400

300 0.1

200
0.05
100

0 0
4 12 20 28 36 4 12 20 28 36
AR user number AR user number
(a) (b) (c)

FIGURE 4. Simulation results: a) end-to-end latency of three schemes; b) energy consumption of three schemes; c) total mobile energy
consumption under different delay tolerances.

can collaborate the cloud computing and edge edge computing. Both processing latency and
computing. Specifically, for the case with 36 energy consumption can be significantly reduced,
devices, the edge computing scheme can reduce leading to better QoE, such as high image resolu-
latency by about 41.44 percent and 12.85 per- tion, long battery autonomy, convenient portabili-
cent as compared with local computing and ty, and real-time performance for AR subscribers.
cloud computing schemes, respectively. Note that
in the simulation, we set the computation capaci- Conclusion
ty of the cloud server almost 10 times of the edge This article presents a novel MEC-based compu-
server as a typical example. With the increase of tation framework for current AR applications to
the cloud computation capacity, the performance optimize energy efficiency and processing delay.
gap between the cloud computing scheme and A hierarchical computation architecture is first
the edge computing scheme will become smaller. developed, which is composed of three layers:
Figure 4b depicts the comparative results the user layer, the edge layer, and the cloud layer.
of the total energy consumption of all devices By seamlessly integrating the communication,
with the number of devices, where the maxi- computation, and control functions at the edge
mum delay tolerance of each AR application is layer and taking full advantage of both edge com-
450 ms. From this figure, the cloud computing puting and cloud computing, computation-inten-
scheme can always achieve better energy effi- sive tasks of AR applications can be intelligently
ciency than the local computing scheme because offloaded to both edge and cloud servers for col-
of the extra energy consumption for tracking and laborative computation. In accordance with this
mapping at local devices. Nevertheless, the pro- architecture, we then develop a novel mechanism
posed edge computing scheme always achieves to simultaneously support multiple AR applica-
the highest energy efficiency among all schemes. tions of different mobile subscribers. Three key
In the scenario with 36 devices, our proposed technologies are also discussed, which can coop-
scheme can save about 73.71 percent and 65.34 erate to further reduce the processing latency and
percent energy consumption as compared with energy consumption of mobile AR devices. Final-
local computing and cloud computing schemes, ly, our proposed architecture and mechanism are
respectively. Specifically, the overall energy con- tested by simulation results, which demonstrate
sumption of each scheme would grow rapidly the substantial performance improvement of our
when the number of devices becomes large, proposals over the existing schemes.
because the communication energy becomes
dominant with a large number of mobile devices, Acknowledgment
which increases exponentially with the number This work was supported in part by the Natu-
of devices. ral Science Foundation of China under Grants
To further show the performance advantage 61671407, 61471319, and 61831004; the Open
on energy saving of the proposed edge comput- Research Fund of the State Key Laboratory of
ing scheme, in Fig. 4c we depict the total energy Integrated Services Networks, Xidian University,
consumption of all mobile devices with different under Grant ISN18-13; the Zhejiang Provincial
delay tolerances, where the number of devices Natural Science Foundation for Distinguished
is set to be 30. From the figure, the mobile ener- Young Scholars under Grant LR19F010002; and
gy consumption of each scheme decreases with the Fundamental Research Funds for the Central
the maximum delay tolerance. The reason is that, Universities.
more computational resources should be utilized
to assist the task processing to meet the stricter References
delay requirement, leading to the higher mobile [1] D. Chatzopoulos et al., “Mobile Augmented Reality Survey:
From Where We Are to Where We Go,” IEEE Access, vol. 5,
energy consumption. Once again, the superiority Apr. 2017, pp. 6917–50.
of the edge computing scheme is demonstrated [2] M. Satyanarayanan et al., “The Case for VM-Based Cloudlets
by this test. in Mobile Computing,” IEEE Pervasive Comput., vol. 8, no. 4,
Overall, our tests demonstrate that the MEC- Oct.-Dec. 2009, pp. 14–23.
[3] F. Liu et al., “Gearing Resource-Poor Mobile Devices with
based AR framework is more efficient than the Powerful Clouds: Architectures, Challenges, and Applica-
other two conventional schemes as a result of tions,” IEEE Wireless Commun., vol. 20, no. 3, Jun. 2013,
the collaboration between cloud computing and pp. 14–22.

168 IEEE Network • July/August 2019

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.
[4] W. Li et al., “Mechanisms and Challenges on Mobility-Aug- sity, Xian 710071, China. His current research interests mainly
mented Service Provisioning for Mobile Cloud Computing,” include machine learning and mobile edge computing.
IEEE Commun. Mag., vol. 53, no. 3, Mar. 2015, pp. 89–97.
[5] T. X. Tran et al., “Collaborative Mobile Edge Computing in Yinghui He received his B.S.E degree in information engineer-
5G Networks: New Paradigms, Scenarios, and Challenges,” ing from Zhejiang University, Hangzhou, China, in 2018. He is
IEEE Commun. Mag., vol. 55, no. 4, Apr. 2017, pp. 54–61. currently pursuing the M.S. degree in information and communi-
[6] Y. Mao et al., “A Survey on Mobile Edge Computing: The cation engineering with the College of Information Science and
Communication Perspective,” IEEE Commun. Surv. Tut., vol. Electronic Engineering, Zhejiang University, Hangzhou, China.
19, no. 4, Aug. 2017, pp. 2322–58. His research interests include 5G wireless communications with
[7] T. Taleb et al., “Mobile Edge Computing Potential in Making an emphasis on mobile edge computing and device-to-device
Cities Smarter,” IEEE Commun. Mag., vol. 55, no. 3, Mar. communications.
2017, pp. 38–43.
[8] X. Masip-Bruin et al., “Foggy Clouds and Cloudy Fogs: A G uan H uang is currently pursuing the M.S. degree in infor-
Real Need for Coordinated Management of Fog-to-Cloud mation and communication engineering with the College of
Computing Systems,” IEEE Wireless Commun., vol. 23, no. 5, Information Science and Electronic Engineering, Zhejiang Uni-
Oct. 2016, pp. 120–28. versity, Hangzhou, China. His research interests mainly include
[9] L. Tong et al., “A Hierarchical Edge Cloud Architecture for mobile edge computing, compressive sensing, and sparse signal
Mobile Computing,” Proc. IEEE Int’l. Conf. Comput. Com- processing.
mun. (INFOCOM), San Francisco, California, Apr. 2016,
pp. 1–9. Guanding Yu (S’05-M’07-SM’13) is a professor at the Zhejiang
[10] T. Verbelen et al., “Leveraging Cloudlets for Immersive Col- Provincial Key Laboratory of Information Processing at Zhejiang
laborative Applications,” IEEE Pervasive Computing, vol. 12, University, China. He is also with the State Key Laboratory of Inte-
no. 4, Oct.-Dec. 2013, pp. 30–38. grated Services Networks, Xidian University, Xian 710071, China.
[11] S. Bohez et al., “Mobile, Collaborative Augmented Reali- His research interests include mobile edge computing and machine
ty Using Cloudlets,” Proc. Mobilware, Bologna, Italy, Nov. learning for wireless networks. He has served as an editor/guest
2013, pp. 45–54. editor for many IEEE journals, such as IEEE JSAC, IEEE Communica-
[12] C. You et al., “Energy-Efficient Resource Allocation for tions Magazine, IEEE Wireless Communications, IEEE Transactions on
Mobile-Edge Computation Offloading,” IEEE Trans. Wireless Green Communications and Networking, IEEE Wireless Communica-
Commun., vol. 16, no. 3, Mar. 2017, pp. 1397–1411. tions Letters, and IEEE Access. He received the 2016 IEEE ComSoc
[13] S. Sardellitti et al., “Joint Optimization of Radio and Com- Asia-Pacific Outstanding Young Researcher Award.
putational Resources for Multicell Mobile-Edge Comput-
ing,” IEEE Trans. Signal Inf. Process. Netw., vol. 1, no. 2, June Yunlong Cai (S’07-M’10-SM’16) received the Ph.D. degree in
2015, pp. 89–103. electronic engineering from the University of York, York, U.K., in
[14] B. Girod et al., “Mobile Visual Search: Architectures, Tech- 2010. From February 2010 to January 2011 he was a postdoc-
nologies, and the Emerging MPEG Standard,” IEEE Multime- toral fellow at the CNAM, Paris, France. Since February 2011 he
dia, vol. 18, no. 3, July-Sept. 2011, pp. 86–94. has been with the College of Information Science and Electronic
[15] E. Salahat and M. Qasaimeh, “Recent Advances in Features Engineering, Zhejiang University, Hangzhou, China, where he is
Extraction and Description Algorithms: A Comprehensive currently an associate professor. His research interests include
Survey,” Proc. IEEE ICIT, Toronto, Canada, Mar. 2017, pp. transceiver design for multiple-antenna systems, mmWave com-
1059–63. munications, and cooperative and relay communications.
[16] A. Al-Shuwaili and O. Simeone, “Energy-Efficient Resource
Allocation for Mobile Edge Computing-Based Augmented Zhaoyang Zhang (M’00) received his Ph.D. degree from Zhe-
Reality Applications,” IEEE Wireless Commun. Lett., vol. 6, jiang University in 1998. Since then he has been with Zhejiang
no. 3, June 2017, pp. 398–401. University and is currently a Qiushi Distinguished Professor.
His research interests are mainly focused on next generation
Biographies wireless communications, network intelligence with converged
Jinke Ren (S’17) received his B.S.E degree in information engi- computation and communication, etc. He is currently serving
neering from Zhejiang University, Hangzhou, China, in 2017. as an editor for IEEE Transactions on Communications and IET
He is currently working toward the Ph.D. degree with the Zhe- Communications, among other publications, and has served as
jiang Provincial Key Laboratory of Information Processing at General Chair, TPC Co-Chair or Symposium Co-Chair for WCSP
Zhejiang University, Hangzhou, China. He is also with the State 2013/2018, Globecom 2014 Wireless Communications Sympo-
Key Laboratory of Integrated Services Networks, Xidian Univer- sium, and VTC-Spring 2017 Workshop HMWC, etc.

IEEE Network • July/August 2019 169

Authorized licensed use limited to: University of Glasgow. Downloaded on November 25,2022 at 20:07:39 UTC from IEEE Xplore. Restrictions apply.

You might also like