You are on page 1of 59

Software

Digital
Transformation
Big Data
Mobile Computing

APRIL 2020 www.computer.org


IEEE COMPUTER SOCIETY computer.org

STAFF
Editor Publications Portfolio Managers
Cathy Martin Carrie Clark, Kimberly Sperka

Publications Operations Project Specialist Publisher


Christine Anthony Robin Baldwin

Production & Design Artist Senior Advertising Coordinator


Carmen Flores-Garvey Debbie Sims

Circulation: ComputingEdge (ISSN 2469-7087) is published monthly by the IEEE Computer Society. IEEE Headquarters, Three Park Avenue, 17th
Floor, New York, NY 10016-5997; IEEE Computer Society Publications Office, 10662 Los Vaqueros Circle, Los Alamitos, CA 90720; voice +1 714 821 8380;
fax +1 714 821 4010; IEEE Computer Society Headquarters, 2001 L Street NW, Suite 700, Washington, DC 20036.
Postmaster: Send address changes to ComputingEdge-IEEE Membership Processing Dept., 445 Hoes Lane, Piscataway, NJ 08855. Periodicals Postage
Paid at New York, New York, and at additional mailing offices. Printed in USA.
Editorial: Unless otherwise stated, bylined articles, as well as product and service descriptions, reflect the author’s or firm’s opinion. Inclusion in
ComputingEdge does not necessarily constitute endorsement by the IEEE or the Computer Society. All submissions are subject to editing for style,
clarity, and space.
Reuse Rights and Reprint Permissions: Educational or personal use of this material is permitted without fee, provided such use: 1) is not made for
profit; 2) includes this notice and a full citation to the original work on the first page of the copy; and 3) does not imply IEEE endorsement of any third-
party products or services. Authors and their companies are permitted to post the accepted version of IEEE-copyrighted material on their own Web
servers without permission, provided that the IEEE copyright notice and a full citation to the original work appear on the first screen of the posted copy.
An accepted manuscript is a version which has been revised by the author to incorporate review suggestions, but not the published version with copy-
editing, proofreading, and formatting added by IEEE. For more information, please go to: http://www.ieee.org/publications_standards/publications
/rights/paperversionpolicy.html. Permission to reprint/republish this material for commercial, advertising, or promotional purposes or for creating new
collective works for resale or redistribution must be obtained from IEEE by writing to the IEEE Intellectual Property Rights Office, 445 Hoes Lane,
Piscataway, NJ 08854-4141 or pubs-permissions@ieee.org. Copyright © 2020 IEEE. All rights reserved.
Abstracting and Library Use: Abstracting is permitted with credit to the source. Libraries are permitted to photocopy for private use of patrons,
provided the per-copy fee indicated in the code at the bottom of the first page is paid through the Copyright Clearance Center, 222 Rosewood Drive,
Danvers, MA 01923.
Unsubscribe: If you no longer wish to receive this ComputingEdge mailing, please email IEEE Computer Society Customer Service at help@
computer.org and type “unsubscribe ComputingEdge” in your subject line.
IEEE prohibits discrimination, harassment, and bullying. For more information, visit www.ieee.org/web/aboutus/whatis/policies/p9-26.html.

IEEE Computer Society Magazine Editors in Chief

Computer IEEE Intelligent Systems IEEE Pervasive Computing


Jeff Voas, NIST V.S. Subrahmanian, Marc Langheinrich, Università
Dartmouth College della Svizzera italiana
Computing in Science
& Engineering IEEE Internet Computing IEEE Security & Privacy
Lorena A. Barba (Interim), George Pallis, University David Nicol, University
George Washington University of Cyprus of Illinois at
Urbana-Champaign
IEEE Annals of the History IEEE Micro
of Computing Lizy Kurian John, University IEEE Software
Gerardo Con Diaz, University of Texas at Austin Ipek Ozkaya, Software
of California, Davis Engineering Institute
IEEE MultiMedia
IEEE Computer Graphics Shu-Ching Chen, Florida IT Professional
and Applications International University Irena Bojanova, NIST
Torsten Möller,
Universität Wien

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 1
APRIL 2020 � VOLUME 6 � NUMBER 4

Customer
and business
value

9 20 42
Digitization

The Innovations Digital Improving


of Open Source Transformation Performance
Hard power Soft power and Scalability of
(technology) (people and Next-Generation
businesses)
Cellular Networks
Software
9 The Innovations of Open Source
DIRK RIEHLE

15 How to Select Open Source Components


DIOMIDIS SPINELLIS

Digital Transformation
20 Digital Transformation
CHRISTOF EBERT AND CARLOS HENRIQUE C. DUARTE

26 Governing and Piloting Emerging Technologies


STEPHEN J. ANDRIOLE

Big Data
29 Big Data Solutions for Micro-, Small-, and Medium-Sized
Enterprises in Developing Countries
DIANA ROJAS-TORRES AND NIR KSHETRI

33 Analytics without Tears, or Is There a Way for Data to Be


Anonymized and Yet Still Useful?
JON CROWCROFT AND ADRIA GASCON

Mobile Computing
40 It Takes a Village to Secure Cellular Networks
ELISA BERTINO

42 Improving Performance and Scalability of Next-


Generation Cellular Networks
ALI MOHAMMADKHAN, K. K. RAMAKRISHNAN, UMA CHUNDURI, AND
KIRAN MAKHIJANI

Departments
4 Magazine Roundup
7 Editor’s Note: The Growth of Open Source
72 Conference Calendar

Subscribe to ComputingEdge for free at


www.computer.org/computingedge.
Magazine Roundup

T he IEEE Computer Society’s lineup of 12 peer-reviewed technical magazines covers cutting-edge topics rang-
ing from software design and computer graphics to Internet computing and security, from scientific appli-
cations and machine intelligence to visualization and microchip design. Here are highlights from recent issues.

simulation, and the result shows fitness-of-use for subsequent


that a new force sensor model can analysis. The transformation his-
Technology Predictions: work well. tory of a data-wrangling system
Art, Science, and Fashion is often insufficient for determin-
ing the usability of a dataset, lack-
Making accurate predictions ing information on how changes
isn’t easy. However, many peo- The Spring of Artificial affected the dataset. Capturing
ple have enjoyed making predic- Intelligence in Its Global workflow provenance along the
tions throughout history, and Winter wrangling process and combin-
even more individuals have taken ing it with descriptive information
part in learning about predictions. Against the dominant narrative as data provenance can enable
Read more in the December 2019 of the history of artificial intelli- users to comprehend how these
issue of Computer. gence (AI), this article from the changes affected the dataset,
October–December 2019 issue of and if they benefited data quali-
IEEE Annals of the History of Com- ty. The authors of this article from
puting shows how, during the pe- the November/December 2019 is-
Force Sensor Model Based on riod considered to be the second sue of IEEE Computer Graphics
FEA for the Electromagnetic global winter (1987–1993), the AI and Applications present DQProv
Levitation System field experienced spring in South Explorer, a system that captures
Korea. Putting the Korean lan- and visualizes provenance from
The traditional control strategy guage-processing problem at the data-wrangling operations.
of the electromagnetic levitation center, researchers began to lay
system (ELS) is a voltage or cur- the intellectual and social founda-
rent control strategy based on the tion of AI in South Korea in the late
gap sensor and the current sen- 1980s. Non-cooperative Target
sor. In this article from the Novem- Detection of Spacecraft
ber/December 2019 issue of Com- Objects Based on Artificial
puting in Science & Engineering, Bee Colony Algorithm
an electromagnetic force sensor Capturing and Visualizing
model is proposed and the map- Provenance from Data Although heuristic algorithms
ping relation between electromag- Wrangling have achieved state-of-the-art
netic force with current and levi- performance for object detec-
tation gap is obtained by Maxwell Data quality management and tion, they have not been demon-
computation. The new force sen- assessment play a vital role for strated to be sufficiently accu-
sor model is used in ELS’s control ensuring trust in data and its rate and robust for multi-object

4 April 2020 Published by the IEEE Computer Society 2469-7087/20 © 2020 IEEE
detection. To address this prob- independent developers still rep- reduced link distances and im-
lem, this article from the July/Au- resent a significant driving force. proved intra-router efficiency.
gust 2019 issue of IEEE Intelligent In this scenario, the availability of However, the sequential fabrica-
Systems incorporates the con- development tools for less-expe- tion methods utilized for M3D in-
cept of species into the artificial rienced developers and profes- tegration impose unique intercon-
bee colony algorithm and propos- sionals to reduce the time need- nect requirements for each of the
es a multi-peak optimization algo- ed to create prototypes is crucial. possible partitioning schemes at
rithm named Species-Based Ar- In this article from the July/Au- transistor, gate, and block gran-
tificial Bee Colony (SABC). The gust 2019 issue of IEEE Internet ularities. Furthermore, increased
authors apply SABC to detect the Computing, the authors first re- cell density introduces conten-
non-cooperative target (NCT) view the tools available to im- tion of available routing resourc-
from two aspects: multi-circle plement all the components of a es. Prior work on M3D NoCs has
detection and multi-template typical IoT architecture in differ- focused on the benefits of re-
matching. Experiments are con- ent programming languages, and duced distances, but has not con-
ducted using real cases of “Shen- then analyze how Python can be sidered these process-imposed
Zhou8” and “Apollo 9” space mis- used to implement all the com- circuit complications. In this ar-
sions, as well as the “Chang’e” ponents of a typical IoT archi- ticle from the November/Decem-
camera point system developed tecture. As a practical example, ber 2019 issue of IEEE Micro, NoC
by the Hong Kong Polytechnic they illustrate the implementa- topology decisions are analyzed in
University. Experimental results tion of a smart home system built conjunction with these M3D inter-
show that the proposed method exploiting low-cost off-the-shelf connect requirements to provide
is robust enough to detect NCT hardware and programmed only an equivalent architectural com-
under various kinds of noise, un- through Python. parison between M3D partitioning
der weak light, and in orbit, which schemes.
leads to accurate detection re-
sults more quickly than other
methods. Network-on-Chip Design
Guidelines for Monolithic 3D A Retrieval System of
Integration Medicine Molecules Based on
Graph Similarity
Rapid Prototyping of IoT Monolithic three-dimensional
Solutions: A Developer’s (M3D) integration is viewed as Medicine information retriev-
Perspective a promising improvement over al has grown significantly and is
through-silicon-via-based 3D in- based on the structural similarity
Many new Internet-of-Things tegration due to its greater inter- of medicine molecules. The chem-
(IoT) devices and solutions ap- tier connectivity, higher circuit ical structural formula (CSF) is a
pear in the market every day. Al- density, and lower parasitic ca- primary search target as a unique
though commercial IoT products pacitance. With M3D integration, identifier for each compound in
are the majority, Do-It-Yourself network-on-chip (NoC) commu- the research field of medical in-
(DIY) solutions implemented by nication fabric can benefit from formation. This article from the

www.computer.org/computingedge 5
MAGAZINE ROUNDUP

October–December 2019 issue accelerating the development of


of IEEE MultiMedia introduces a smarter vehicles, including au-
graph-based CSF retrieval sys- The Security Implications of tonomous cars. The authors of
tem, PharmKi, which accepts the Data Subject Rights this article from the November/
photos taken from smartphones December 2019 issue of IT Profes-
and the sketches drawn on the Data protection regulations give sional analyze the state-of-the-art
tablet PCs as inputs. To estab- individuals rights to obtain the in- results of autonomous cars, iden-
lish a compact yet efficient hy- formation that entities have on tifying the main stakeholders and
pergraph representation for mol- them. However, providing such in- their role in their success or fail-
ecules, the authors propose a formation can also reveal aspects ure. They also discuss some of
graph-isomorphism-based algo- of the entity’s underlying technical the challenges and opportunities
rithm for evaluating the spatial infrastructure and organization- of autonomous cars and identi-
similarity between graphical CS- al processes. This article from the fy some applications that might
Fs. An indexing strategy based on November/December 2019 issue justify their adoption in modern
the graph TF-IDF technology is al- of IEEE Security & Privacy explores society.
so introduced to achieve a high the security implications this rais-
efficiency for large-scale mol- es and highlights the need to con-
ecule retrieval. The results of a sider such rights in fulfillment
comparative study demonstrate processes.
that the proposed method out-
performs the existing methods
on accuracy, and performs well
on efficiency. From Art to Science: The
Evolution of Community
Development

Esports Athletes and Players: Community development in open-


A Comparative Study source ecosystems is increasing-
ly complex. This article from the
The authors of this article from November/December 2019 issue
the July–September 2019 issue of of IEEE Software focuses on the
IEEE Pervasive Computing pres- OpenShift and CNCF ecosystems
ent a comparative study of regu- and concludes that cross-commu-
lar players’ and professional play- nity collaboration analysis is chal- Join the IEEE
ers’ (athletes’) performance in lenging and that a more scientific Computer
Counter Strike: Global Offensive approach is required. Society
discipline. Their study is based
computer.org/join
on ubiquitous sensing, helping to
identify the biometric features sig-
nificantly contributing to the clas- Autonomous Cars: Challenges
sification of particular skills of the and Opportunities
players. The research provides a
better understanding of why the We are witnessing an evolution in
athletes demonstrate superior the automotive industry. Recent
performance as compared to oth- technological advances and fast
er players. proliferation of technologies are

6 ComputingEdge April 2020


Editor’s Note

The Growth of Open Source

O ver the past decade, the


programming community
has embraced open source soft-
Components” provides a practical
guide for choosing the right open
source projects to use in your
Enterprises in Developing Coun-
tries,” the authors show how uti-
lizing big data is helping busi-
ware (OSS), which is now used in work based on requirements such nesses in developing countries
everything from mobile apps to as functionality, licensing, docu- improve processes and access
defense systems. The sharing and mentation, and code quality. financial services. IEEE Internet
collaboration that OSS allows has Open source isn’t the only Computing’s “Analytics without
propelled its popularity among trend shaping the software indus- Tears, or Is There a Way for Data to
coders, researchers, and orga- try; digital transformation—which Be Anonymized and Yet Still Use-
nizations. GitHub recently sur- refers to organizations leverag- ful?” discusses privacy concerns
passed 40 million users and 100 ing the latest computing technol- related to big data analytics.
million repositories, while 89 per- ogy to increase their impact—will The final two articles in this
cent of IT leaders say that OSS is also affect the future of software. ComputingEdge issue address
important in their organization, IEEE Software’s “Digital Transfor- different aspects of 5G cellular
according to a Red Hat survey. mation” posits that it “will com- networks: security and perfor-
This ComputingEdge issue pletely reshape the landscape of mance. IEEE Security & Privacy’s
opens with two articles from software technologies and pro- “It Takes a Village to Secure Cel-
Computer that focus on the role cesses.” IT Professional’s “Govern- lular Networks” examines secu-
of OSS in modern software devel- ing and Piloting Emerging Technol- rity challenges and possible solu-
opment. In “The Innovations of ogies” advises business leaders on tions. IEEE Internet Computing’s
Open Source,” the author dis- which cutting-edge technologies “Improving Performance and
cusses the important legal, pro- to deploy in their companies. Scalability of Next-Generation
cess, tool, and business model Another article from IT Pro- Cellular Networks” investigates
innovations that the open source fessional covers digital trans- possible changes to system archi-
movement has produced in the formation with a focus on big tecture and protocols that could
software industry and beyond. data. In “Big Data Solutions for lead to lower latency and a better
“How to Select Open Source Micro-, Small-, and Medium-Sized user experience.

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 7
PURPOSE: The IEEE Computer Society is the world’s largest EXECUTIVE COMMITTEE
association of computing professionals and is the leading provider President: Leila De Floriani
of technical information in the field.
President-Elect: Forrest Shull
MEMBERSHIP: Members receive the monthly magazine Past President: Cecilia Metra
Computer, discounts, and opportunities to serve (all activities First VP: Riccardo Mariani; Second VP: Sy‐Yen Kuo
are led by volunteer members). Membership is open to all IEEE Secretary: Dimitrios Serpanos; Treasurer: David Lomet
members, affiliate society members, and others interested in the VP, Membership & Geographic Activities: Yervant Zorian
computer field. VP, Professional & Educational Activities: Sy-Yen Kuo
VP, Publications: Fabrizio Lombardi
COMPUTER SOCIETY WEBSITE: www.computer.org
VP, Standards Activities: Riccardo Mariani
OMBUDSMAN: Direct unresolved complaints to VP, Technical & Conference Activities: William D. Gropp
ombudsman@computer.org.
2019–2020 IEEE Division VIII Director: Elizabeth L. Burd
CHAPTERS: Regular and student chapters worldwide provide the 2020-2021 IEEE Division V Director: Thomas M. Conte
opportunity to interact with colleagues, hear technical experts, 2020 IEEE Division VIII Director-Elect: Christina M. Schober
and serve the local professional community.
AVAILABLE INFORMATION: To check membership status, report BOARD OF GOVERNORS
an address change, or obtain more information on any of the Term Expiring 2020: Andy T. Chen, John D. Johnson,
following, email Customer Service at help@computer.org or call
Sy-Yen Kuo, David Lomet, Dimitrios Serpanos,
+1 714 821 8380 (international) or our toll-free number, Hayato Yamana
+1 800 272 6657 (US): Term Expiring 2021: M. Brian Blake, Fred Douglis,
• Membership applications Carlos E. Jimenez-Gomez, Ramalatha Marimuthu,
• Publications catalog Erik Jan Marinissen, Kunio Uchiyama
• Draft standards and order forms Term Expiring 2022: Nils Aschenbruck,
• Technical committee list Ernesto Cuadros‐Vargas, David S. Ebert, William Gropp,
• Technical committee application Grace Lewis, Stefano Zanero
• Chapter start-up procedures
• Student scholarship information
• Volunteer leaders/staff directory EXECUTIVE STAFF
• IEEE senior member grade application (requires 10 years Executive Director: Melissa A. Russell
practice and significant performance in five of those 10) Director, Governance & Associate Executive Director:
Anne Marie Kelly
PUBLICATIONS AND ACTIVITIES Director, Finance & Accounting: Sunny Hwang
Director, Information Technology & Services: Sumit Kacker
Computer: The flagship publication of the IEEE Computer Society,
Director, Marketing & Sales: Michelle Tubb
Computer publishes peer-reviewed technical content that covers
Director, Membership Development: Eric Berkowitz
all aspects of computer science, computer engineering,
technology, and applications.
COMPUTER SOCIETY OFFICES
Periodicals: The society publishes 12 magazines and 18 journals. Washington, D.C.: 2001 L St., Ste. 700, Washington, D.C.
Refer to membership application or request information as noted 20036-4928; Phone: +1 202 371 0101; Fax: +1 202 728 9614;
above. Email: help@computer.org
Conference Proceedings & Books: Conference Publishing Los Alamitos: 10662 Los Vaqueros Cir., Los Alamitos, CA 90720;
Services publishes more than 275 titles every year. Phone: +1 714 821 8380; Email: help@computer.org
Standards Working Groups: More than 150 groups produce IEEE
MEMBERSHIP & PUBLICATION ORDERS
standards used throughout the world.
Phone: +1 800 678 4333; Fax: +1 714 821 4641;
Technical Committees: TCs provide professional interaction in Email: help@computer.org
more than 30 technical areas and directly influence computer
engineering conferences and publications. IEEE BOARD OF DIRECTORS
Conferences/Education: The society holds about 200 conferences President: Toshio Fukuda
each year and sponsors many educational activities, including President-Elect: Susan K. “Kathy” Land
computing science accreditation. Past President: José M.F. Moura
Certifications: The society offers three software developer Secretary: Kathleen A. Kramer
credentials. For more information, visit Treasurer: Joseph V. Lillie
www.computer.org/certification. Director & President, IEEE-USA: Jim Conrad
Director & President, Standards Association: Robert S. Fish
BOARD OF GOVERNORS MEETING Director & VP, Educational Activities: Stephen Phillips
Director & VP, Membership & Geographic Activities:
28 – 29 May: McLean, Virginia Kukjin Chun
Director & VP, Publication Services & Products: Tapan Sarkar
Director & VP, Technical Activities: Kazuhiro Kosuge

revised 14 January 2020


EDITOR: Dirk Riehle, Friedrich Alexander-University of Erlangen Nürnberg, dirk.riehle@fau.de

This article originally


appeared in
COLUMN: OPEN SOURCE EXPANDED
vol. 52, no. 4, 2019

The Innovations of Open Source


Dirk Riehle, Friedrich-Alexander-Universität, Erlangen-Nürnberg

FROM THE EDITOR


Welcome to this new column on open source! The “Open Source Expanded” column will aim to provide an
insightful article every two months. These articles will be written for the software practitioner by authors
from both academia and industry. Articles will be grouped by theme rather than appearing in arbitrary
order. Our first theme is about open source licenses and license compliance. Even decades after open
source was created, it is still a hot topic and unknown territory for many. Later themes will focus on using
open source, project communities, business models, interesting cases, and what might come after open
source. This issue’s article, the first in its series, provides an overview of what is to come by reviewing
the most important innovations that open source has provided for the software industry and beyond. If
you have comments or would like to suggest future themes and articles, feel free to contact me at dirk@
riehle.org. Computer will provide a discussion board for articles as well. — D. Riehle

Open source has given us many innovations. This article provides an overview of the most
important innovations and illustrates the impact that open source is having on the software
industry and beyond.

T
he main innovations of open source can be ›› the ability to pass on the source code and a
grouped into four categories: legal, process, binary copy.
tool, and business models. Probably the best
known innovations are open source licenses, which Before there was open source software, there was
also define the concept. Software becomes open free software. Richard Stallman defined the four free-
source if users receive it under an open source license. doms of software that make it “free” as:2
To be an open source license, it must fulfill 10 require-
ments set forth by the Open Source Initiative, the pro- the freedom to run the program as you wish, for any
tector and arbiter of what constitutes open source.1 purpose […], the freedom to study how the program
Most notably, the license must allow works, and change it so it does your computing as
you wish […], the freedom to redistribute copies so
›› free-of-charge use of the software you can help others […], the freedom to distribute
›› access to and modification of the source code copies of your modified versions to others […].

Open source software and free software, and the


DOI No. 10.1109/MC.2019.2898163 people behind them, have struggled with each other
Date of publication: 16 April 2019 at times. For all practical purposes, however, the

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 9
difference is irrelevant to users. What matters is incorrectly discredit open source software as “a virus”
the license under which a user receives a particular or “cancer” and a “communist” or “hippie undertaking.”
software. However, nobody forces anyone to use open source
software. In an amazing about-face, some of the most
LEGAL INNOVATION well-known companies that fought open source only
Licenses can be structured into permissions (the 15 years ago are now among its biggest supporters.
rights granted to a user), obligations (what is required The “Business Model Innovation” section of this article
to receive these rights), and prohibitions (what may explains some of this.
not be done; for example, claiming that using the soft-
ware implies an endorsement by its creator). The two PROCESS INNOVATION
legal innovations are The next innovation open source has brought us is en-
gineering process innovation.3 The open source ini-
1. the rights grant as introduced earlier tiative has this to say about open source software
2. a particular obligation called copyleft. development:1

The rights grant helped open source spread and Open source is a development method for software
succeed. As research has shown, it taps into the that harnesses the power of distributed peer review
human desire to help each other and collaborate on and transparency of process. The promise of open
interesting projects. source is better quality, higher reliability, more flexibil-
People sometimes ask why developers do not put ity, lower cost, and an end to predatory vendor lock-in.
their work into the public domain. This misses the
point: by putting something into the public domain, This is the other definition of open source, which
an author typically waives his or her rights, and most does not focus on licenses and intellectual property
authors do not want that. Rather, they want to be spe- but, rather, on collaborative development. There is
cific about which rights they grant and which obliga- no single open source software engineering process
tions they require. because each open source community defines its own.
The most famous license obligation is probably the Through his development of the Linux kernel, Linus
copyleft clause. Stallman invented this clause, and it Torvalds was the first to explore, at scale, a truly col-
became popular through GNU General Public License laborative open source process. His approach has no
v2 in 1991. It states that if you pass on copyleft-licensed particular name but is often identified with his moniker,
code, such as part of a product that you sell, you must BDFL (which stands for “benevolent dictator for life”),
also pass on your own source code if it modifies the implying a hierarchical structure. A core benefit of an
copyleft-licensed code. The specifics of this can get open collaboration process was named after Torvalds
complicated quickly, and they will be discussed in more and is called Linus’ law, which states, “Given enough
detail in future columns. Many companies worry that if eyeballs, all bugs are shallow.”4 The idea is that more
their source code is mixed with copyleft-licensed code, broadly used software matures more quickly since
they will lose their intellectual property and, hence, problems are found and solved more quickly.
their competitive advantage in the marketplace. The collaborative peer group, as explored by the
In the past, companies have used this clause to original Apache web server team (httpd) and codified

10 ComputingEdge April 2020


OPEN SOURCE EXPANDED

as The Apache Way (of open source software devel- TOOL INNOVATION
opment), is a similar but different approach that may Most of the tools used in open source software devel-
be more popular today. 5 The software industry owes opment are familiar to closed source programmers as
this group of developers as much as it owes Torvalds, well. However, the needs of open source processes
if not more. have led to two major tool innovations that have since
The Apache Way is a consensus-based, community become an important part of corporate software de-
driven governance approach to collaboration in open velopment as well: software forges and distributed
source projects. The Apache Software Foundation’s version control.
website explains it in detail. An important aspect is the A software forge is a website that allows the cre-
distinction between contributors, who submit work ation of new projects and provides developers with
for inclusion in an open source project, and commit- all of the tools needed for software development,
ters, who review and integrate the work. Committers such as a home page, an issue tracker, and version
are called maintainers in a Linux context, and they control. What makes software forges special is that
usually are developers, too, not just reviewers. Using they facilitate matchmaking between those who are
this contributor–committer interplay, nearly all open looking to find a useful software component and
source projects practice precommit code review to those who are offering one. They are an enterprise
ensure the quality of the software under development. software product category because, even within one
The principles of open source software develop- company, you want to have one place for all software
ment can be summarized as three principles of open being developed.
collaboration.6 Distributed version control is version control in
which you copy the original repository and work with
›› In open collaboration, participation is egalitarian your copy. Thus, you do not need commit rights or ask
(nobody is a priori excluded). for permission to start work. Git and Mercurial are the
›› Decision making is meritocratic (based on the two best-known examples of such software. Some
merits of arguments rather than status in a may argue that distributed version control is not an
corporate hierarchy). open source innovation because some of its roots are
›› People are self-organizing (they choose projects, in proprietary software. However, the open source
processes, and tasks rather than being assigned community developed and refined its own solutions,
to them). which work well with how open source software is
developed, and thereby popularized the concept.
Similarly, open source projects practice open Comparing distributed version control with branch-
communication. This form of communication is public ing misses the point. Having your own repository allows
(everyone can see it), written (so you don’t have to be developers to work using their own style, free of any
there when words are spoken), complete (if it wasn’t centralized decisions on how to use branches.
written down, it wasn’t said), and archived (so that Distributed version control was popularized by being
people can look up and review discussions later). the main version control software underlying a new gen-
Such open collaborative processes, which are not eration of software forges, most notably Github and
dominated by any single entity, lead to community Gitlab. As such, companies are adopting both forges
open source software, which is collectively owned, and distributed version control at a rapid pace.
managed, and developed by a diverse set of stakehold-
ers. These collaboration processes are not limited to BUSINESS MODEL INNOVATION
software but spill over into adjacent areas. For exam- Open source is changing the software industry by how
ple, they have brought forward many formal and de it makes new business models and breaks old ones.
facto standards that the software industry relies on.3 For instance, it lays the legal foundation for open col-
The methods for open source software development laboration between individuals and companies, de-
have also taken root inside companies, where they are fines more effective collaboration processes with
called inner source.7,8 higher productivity than closed-source approaches,

www.computer.org/computingedge 11
OPEN SOURCE EXPANDED

and invents the tools to support it. Open source itself are support and update services, but it often
may not be a business model, but it is a potent strategy also includes a copyleft license that users
and a tool to use in competitive environments. would like to replace with a proprietary one.
If done correctly, both the company and
For-profit models its products benefit from the help of the
There are different approaches for classifying busi- community of nonpaying users. The company
ness models enabled by open source, but I like to put typically does not get code contributions, but it
them into five categories. Three are for-profit business does get lively discussion forums, bug reports,
models, and two are nonprofit models. The for-profit feature ideas, and word-of-mouth marketing.
business models are as follows. The most well-known example of this model
was MySQL, the database company, but there
1. Consulting and support business models: In are many more recent ones, such as SugarCRM,
this conventional model, a company earns MongoDB, and Redis Labs.
money by providing consulting and support
services for existing open source software. The distributor and single-vendor models are
They do not sell a license, but they service the especially important because they enable returns
software anyway. The original open source on investment that are attractive to venture capital-
service company was Cygnus Solutions, which ists. Thus, they are the main conduit through which
serviced the GNU set of tools. More recent billions of dollars have been invested into open
examples are Cloudera and Hortonworks, source software.
which service Hadoop.
2. Distributor business model: In this business Open source foundations
model unique to open source, a company sells There are two more models that determine how the
subscriptions to software and associated development of open source software is being fund-
services that are partly or completely based ed. They are actually two variants of the same idea: the
on open source software. This model only open source foundation.
works for complex software that consists of An open source foundation is a nonprofit organiza-
tens or hundreds and sometimes thousands tion tasked with governing one or more open source
of possibly incompatible components that a projects, representing them legally, and ensuring their
customer wants to use. future. In the past, open source foundations were set
The most well-known examples are Linux up to ensure the survival of unsupported community
distributors like Red Hat and Suse, but many open source projects, but companies are increasingly
other smaller companies provide distributions coming together to set up a foundation with the goal of
of other kinds. The competitively differentiat- developing new open source software.
ing intellectual properties of a distributor are The two variants of open source foundations are
its test suites, configuration databases, and as follows.
compatibility matrices, which they typically do
not open source. 1. Developer foundations: This type of nonprofit
3. Single-vendor open source business model: foundation is run by software vendors (devel-
In this model, a company goes to market by opers) who decide to join forces to ensure
providing a sometimes reduced, sometimes the survival and health of the open source
complete, version of its product as open software they depend on. By ensuring broadly
source. The company never lets go of full shared ownership of the software, the vendors
ownership of the software and sets up various make certain that no one can monopolize this
incentives for users to move from the free particular type of component and reap all of
open source version to a paid-for, commercially the profits from software products that rely
licensed version. The most common incentives on it. This is why Linux was supported against

12 ComputingEdge April 2020


OPEN SOURCE EXPANDED

FIGURE 1.The economic logic of community open source software.

Microsoft Windows, Eclipse against Microsoft software development to increase by a couple of


Visual Studio, and, more recently, OpenStack orders of magnitude.
against Amazon Web Services.
2. User foundations: This type of nonprofit is REFERENCES
predominantly run by companies that are not 1. Open Source Initiative. 2018. [Online]. Available: https://
software vendors but rely on the software opensource.org
managed by the foundation, either as part of 2. Free Software Foundation, “GNU operating system.” 2018.
their operations or directly as part of a product [Online]. Available: https://www.gnu.org/philosophy
that is only partly software. Examples are the /free-sw.en.html
Kuali Foundation for software to run universi- 3. C. Ebert, “Open source drives innovation,” IEEE Softw.,
ties, the GENIVI foundation for automotive vol. 24, no. 3, pp. 105–109, 2007.
infotainment software, and the openKON- 4. E. Raymond, “The cathedral and the bazaar,” Knowledge,
SEQUENZ foundation for software for the Technol. Policy, vol. 12, no. 3, pp. 23–49, 1999.
(German) smart energy grid (the last of which I 5. The Apache Software Foundation. 2018. [Online].
helped create). Available: https://www.apache.org/foundation
/how-it-works.html
6. D. Riehle, “The five stages of open source volunteer-

F igure 1 shows how replacing a closed source


component in a product with an open source
component shifts profits between the different com-
ing,” in Crowdsourcing, W. Li, M. N. Huhns, W.-T. Tsai,
and W. Wu, Eds. New York: Springer-Verlag, 2015,
pp. 25–38.
ponent suppliers and generally leaves more profit for 7. J. Dinkelacker, P. K. Garg, R. Miller, and D. Nelson,
the vendor, which integrates the components and “Progressive open source,” in Proc. 24th Int. Conf.
sells the final product. Because of this economic logic, Software Engineering, 2002, pp. 177–184.
I expect to see more product vendors and service sup- 8. D. Riehle, M. Capraro, D. Kips, and L. Horn, “Inner source
pliers from outside the software industry get in on the in platform-based product engineering,” IEEE Trans.
game. They will fund the development of open source Softw. Eng., vol. 42, no. 12, pp. 1162–1177, Dec. 2016.
components they need, taking money out of the
market for such components and moving it to places DIRK RIEHLE is the professor for open source software at
where they can more easily appropriate it. Therefore, the Friedrich Alexander-University of Erlangen Nürnberg.
in the future, we can expect funding for open source Contact him at dirk@riehle.org.

www.computer.org/computingedge 13
IEEE Security & Privacy is a bimonthly magazine
communicating advances in security, privacy,
and dependability in a way that is useful to a
broad section of the professional community.

The magazine provides articles with both a


practical and research bent by the top thinkers in
the field of security and privacy, along with case
studies, surveys, tutorials, columns, and in-depth
interviews. Topics include:
• Internet, software, hardware, and systems security
• Legal and ethical issues and privacy concerns
• Privacy-enhancing technologies
• Data analytics for security and privacy
• Usable security
• Integrated security design methods
• Security of critical infrastructures
• Pedagogical and curricular issues in security education
• Security issues in wireless and mobile networks
• Real-world cryptography
• Emerging technologies, operational resilience,
and edge computing
• Cybercrime and forensics, and much more

www.computer.org/security

E-Currency and Fairness ■ Ransomware Defense ■ A National Cybersecurity Policy


IEEE SECURITY & PRIVACY

Software and Cybersecurity ■ Big Data: Privacy Versus Accessibility ■ Resiliency in Cloud Computing
AI ETHICS

Blockchain Technologies ■ The Fuzzing Revival ■ Cybersecurity for the Public Interest
IEEE SECURITY & PRIVACY

IEEE SECURITY & PRIVACY


DIGITAL FORENSICS, PART 2

PRIVACY AND AUTOMATED AIRPORT SCREENING


VOLUME 17

VOLUME 16
NUMBER 1
JANUARY/FEBRUARY 2019

NUMBER 3

January/February 2019 CYBERSECURITY AND


PRIVACY ISSUES IN BRAZIL
Vol. 17, No. 1
WWW.COMPUTER.ORG/SECURITY

VOLUME 17

November/December 2018
Vol. 16, No. 6
NUMBER 2
MAY/JUNE 2018

MARCH/APRIL 2019

Join the IEEE Computer Society May/June 2018


WWW.COMPUTER.ORG/SECURITY

Vol. 16, No. 3 March/April 2019


Vol. 17, No. 2
WWW.COMPUTER.ORG/SECURITY

for subscription discounts today!


www.computer.org/product/magazines/security-and-privacy
EDITOR: Dirk Riehle, Friedrich Alexander-University of Erlangen Nürnberg, dirk.riehle@fau.de

This article originally


appeared in
COLUMN: OPEN SOURCE EXPANDED
vol. 52, no. 12, 2019

How to Select Open Source


Components
Diomidis Spinellis, Athens University of Economics and Business

FROM THE EDITOR


Welcome back! Open source gives you high-quality software for free. What’s not to like about this? But
wait a second: You need to choose the right open source component. Making a poor choice for using
an open source component in your products or projects can create serious problems. In this article,
well-known open source expert Diomidis Spinellis takes us through the process of selecting the right
open source component for your needs. Significant thought should be spent on such a decision, because
using an open source component creates a dependency that needs to be managed, and some depen-
dencies are easier to manage than others. But more on this topic in one of the next columns. As always,
happy hacking! — Dirk Riehle

With millions of open source projects available on forges such as GitHub, it may be difficult
to select those that best match your requirements. Examining each project’s product and
development process can help you confidently select the open source projects required for
your work.

I
f most of the code comprising your product or protocol matches the one prevalent in your indus-
service isn’t open source software, it’s highly likely try and whether the system can scale in the future to
that you’re wasting effort and cash reinventing cover your projected needs. It is equally important to
the wheel. Yet with millions of open source projects evaluate whether the project’s functionality is egre-
available on forges such as GitHub, it may be difficult giously excessive compared to your needs. For exam-
to select those that best match your requirements. ple, if you simply want to compress data that you
Examining two facets of each candidate project, the store in a file, you may not want to use a multiformat
product and its development process (see Table 1), data archiving library. Selecting a small, focused proj-
can help you select with confidence the open source ect over a larger one has many advantages. In typi-
projects required for your work. cal cases, such a choice will offer a reduced storage
footprint for your system, fewer transitive third-party
PRODUCT dependencies, a lower installation complexity, and a
smaller surface vulnerable to malicious attacks.
Functionality If an open source project’s functionality nearly fits
Begin by assessing the functionality of the project your organization’s needs and no other project can
under consideration and determine whether it cov-
ers both current needs and future strategic direc-
tions. For instance, if you are selecting a message DOI No. 10.1109/MC.2019.2940809
queue, consider whether the underlying messaging Date of current version: 22 November 2019

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 15
TABLE 1. Judging the open source project selection criteria.

satisfy them completely, you can still use it and make Public License cannot be linked together with code
the required changes on your own. However, under this licensed under the GNU General Public License.
scenario, you must more stringently evaluate the ele-
ments I outline later on regarding source code changes Nonfunctional properties
and contributions. See the last column in Table 1. Evaluate the project’s fit with your requirements by
also looking at its nonfunctional properties. Is it com-
Licensing patible with your product’s processor architecture,
Narrow down your search by examining whether the operating system, and middleware? Will it accommo-
project’s licensing 1 is compatible with your business date your future expansion plans and directions? For
model, mission, or other software you are using. Within example, if your product works on macOS but you’re
your project’s source code, if you directly incorpo- also eyeing the Windows market, then you should be
rate elements licensed under the GNU General Public using open source libraries supported on both sys-
License, then you also must distribute your code under tems. Is the product’s performance compatible with
the same license. This may be undesirable if your busi- your requirements? This is especially important when
ness model depends on keeping your product’s source selecting a database or a big data analytics infrastruc-
code under wraps; in this case, you should be looking ture. If performance is critical, do not assume partic-
for projects that use more permissive licenses, such as ular performance outcomes; rather, benchmark with
the Berkeley Software Distribution and Apache ones. realistic workloads.
Similar concerns apply if you are offering software as
a service, and you plan to use software licensed under Popularity
the Affero General Public License. As another exam- Then consider the project’s popularity. Popularity is
ple, software released under version 1.1 of the Mozilla important because it can determine how likely it will be

16 ComputingEdge April 2020


OPEN SOURCE EXPANDED

for your questions to receive answers on public forums, and browse a few files. Are methods or functions short
for volunteers to contribute fixes and enhancements, and readable? Are identifiers well chosen? Is the code
and for the project to continue to evolve if its original reasonably commented? Is the formatting consistent
developers veer off course (namely, losing interest or with the language’s coding conventions? Again, seri-
steering the project toward an undesirable direction). ous deviations are often indicators of more important
Simple metrics, such as GitHub stars, the number of hidden flaws.
StackOverflow questions with the corresponding tag,
the download count, and the number of Google query Build process
results are all usually sufficient to discern the cases The quality of a project’s build process is important for
that really matter. two reasons. Some organizations reuse open source
code projects through binary distributions, as librar-
Documentation ies, that they link with their other code or as compo-
The project’s documentation is another aspect that nents that run on their infrastructure. If your organi-
should be examined. Although most answers regard- zation works like this, at some point you may need to
ing a software’s operation ultimately lie in the source build the binary from source code to fix a bug or add
code, resorting to such digging for everyday opera- a feature required by your organization. Other organi-
tions is undesirable. Therefore, judge how well the zations (mostly larger ones) have strict rules against
software is documented, both at the technical (instal-
lation and maintenance procedures) and user levels
(tutorials and reference manuals). Although nearly all SOFTWARE RELEASED UNDER
mature open source software projects are well docu- VERSION 1.1 OF THE MOZILLA PUBLIC
mented, some smaller ones suffer in this dimension. LICENSE CANNOT BE LINKED
There are Unix command-line utilities, for example, TOGETHER WITH CODE LICENSED
that lack the traditional manual page. I try to avoid UNDER THE GNU GENERAL PUBLIC
such projects, both to keep my sanity (life is too short LICENSE.
to waste on hunting down command-line options) and
because such a level of indifference toward the end
user is often a sign of deeper problems. using random binaries off the Internet and have pro-
cesses for building everything internally from source
SOURCE CODE (at least once).
This brings me to another product characteristic you Whatever the case, it’s sensible to check how
should check, namely the project’s source code and easy it is to perform a project build. Is the procedure
the code’s quality. If you anticipate adjusting the proj- documented? Does it work in your environment? Will
ect to your needs, then select projects written in pro- you need some rarely used build tools, an unsupported
gramming languages with which you are familiar. Even integrated development environment, or a compiler for
if you don’t plan to touch the project’s source code, an exotic programming language? For critical depen-
low code quality can affect you through bugs, secu- dencies, evaluate these requirements in the same way
rity vulnerabilities, poor performance, and mainte- that you’re evaluating the primary open source project
nance problems. Again, there’s no need to dig deeply under consideration.
to form a useful opinion. In most cases, your objective
is to avoid problematic projects, not to perform thor- PROCESS
ough due diligence of the code. Look at the project’s No matter how shiny the open source project appears
source code files. Are they named and organized into to your eyes, you also should invest some time to exam-
directories following the conventions of the project’s ine how it is produced and managed. This will affect
programming language? Is there evidence of unit test- your experience with it in the long term and also may
ing? Does the repository also contain elements that it uncover potential pitfalls that weren’t discernible from
shouldn’t, such as object and executable files? Open the product’s examination.

www.computer.org/computingedge 17
OPEN SOURCE EXPANDED

Development process deep-learning library), you want to see regular updates;


Start by evaluating the quality of the project’s devel- for more stable ones, you’re looking for evidence of
opment process. Does the project practice continu- maintenance releases. In some cases, frequently inte-
ous integration? You can easily determine this by look- grating new releases of an open source component
ing for corresponding configuration files (for example, into your code base can be disruptive, due to the risks
.travis.yml or Jenkinsfile) in the project’s root direc- and additional work of this process. To avoid these
tory. Examine what the continuous integration pipe- problems, check for a separate release channel for
line exercises. Does it, for example, include static anal- obtaining only security and other critical fixes. Addi-
ysis of the code as well as unit testing? Does it build tionally, to minimize the disturbance associated with
and spell-check the documentation? Does it calculate bringing in major updates, see if there are so-called
testing code coverage? Does it enforce coding stan- long-term support releases and determine whether
dards? Does it check for up-to-date dependencies? their time horizon matches your project’s pace.
A shortcut for answering these questions are badges
appearing in the project’s GitHub page, though their Support channels
significance is not always a given.2 Source code availability is an excellent insurance pol-
icy for obtaining support because it allows you to
resolve issues and fix bugs within your organization;
“Use the source, Luke,” to paraphrase a line from Star
A LACK OF FRESH COMMITS MAY Wars. Such measures, however, are typically extreme.
IMPLY THAT NOBODY WILL STEP IN TO When using open source software, a helpful support
ADDRESS NEW REQUIREMENTS OR forum is usually the most practical way to resolve
BUGS. such problems. Consequently, look for the project’s
available support channels. Is there an online forum,
a mailing list, or a chat group where you can ask ques-
tions? Do useful answers arrive quickly? Are respon-
Code commits dents supportive and friendly? In my experience,
Then look at code commits to the project’s revision the quality of a project’s technology and its support
management repository. Are commits regularly made are orthogonal. Some projects with mediocre qual-
by a diverse group of committers? Unless the proj- ity code offer excellent support and vice versa. For
ect is very stable and likely to remain so (consider a enterprise scenarios where it’s not prudent to rely
numerical library), a lack of fresh commits may imply on volunteer help for resolving critical issues, you
that nobody will step in to address new requirements may also wish to examine the quality of paid support
or bugs. Similarly, commits by a single author or very options offered through specialized companies, con-
few signal that the project suffers from a key person sultants, or products.
risk. Also known as a bus factor, this identifies the dan-
ger the project faces if, for example, a lead developer is Handling issues
hit by a bus.3 Also, look at the details of a few commits. Inevitably, at some point, you’re likely to encounter a
Are they clearly labeled and appropriately described? bug in the open source project you’re using. Therefore,
Do they reference any documented issues that they it’s worth examining how the project’s volunteers han-
have addressed using a standard convention? Is there dle issues.4 Many open source projects offer access
evidence that code changes and additions have been to their issue management platform, such as GitHub
reviewed and discussed? Issues, Bugzilla, or Jira, which allows you to look
under the hood of issue handling. Are issues resolved
Project releases quickly? How many issues have been left rotting open
Down the road, see how these commits translate for ages? Does the ratio between open and closed
into complete project releases. Are these sufficiently issues appear to be under control, that is, in line with
recent and frequent? For cutting-edge projects (say, a the number of project contributors?

18 ComputingEdge April 2020


OPEN SOURCE EXPANDED

Contributing fixes and enhancements 2. A. Trockman, S. Zhou, C. Kästner, and B. Vasilescu,


Another scenario down the road concerns the case “Adding sparkle to social coding: An empirical study of
where you make some changes to the project’s source repository badges in the npm ecosystem,” in Proc. 40th
code, either to fix a bug or add a new feature that Int. Conf. Software Engineering, 2018 , pp. 511 – 522.
your organization requires. Although you can keep 3. K. Yamashita, S. McIntosh, Y. Kamei, A. E. Hassan,
your changes to yourself, integrating them into the and N. Ubayashi, “ Revisiting the applicability of the
upstream project safeguards their continued availabil- Pareto principle to core development teams in open
ity and maintenance alongside new releases (in addi- source software projects,” in Proc. 14th Int. Workshop
tion to it being the proper thing to do). Principles of Software Evolution, 2015, pp. 46 – 55.
Evaluate how you’ll fare in this case by examining 4. T. F. Bissyandé, D. Lo, L. Jiang , L. Réveillère, J. Klein, and
how easy it is to contribute fixes and enhancements. Y. L. Traon, “Got issues? Who cares about it? A large
Is there a contributor’s guide? If you’re using the proj- scale investigation of issue trackers from GitHub,” in
ect as a binary package, is it easy to build and test the Proc. IEEE 24th Int. Symp. Software Reliability Engineer-
project from its source code? Through what hoops do ing (ISSRE), 2013, pp. 188 –197.
you have to jump to get your contribution accepted?
Is there an efficient method by which to submit your DIOMIDIS SPINELLIS is a professor of software engineering
changes, for example, through a GitHub pull request? and head of the Department of Management Science and
Does the project regularly accept third-party contribu- Technology, Athens University of Economics and Business.
tions? Note that some organizations release projects His most recent book is Effective Debugging: 66 Specific
with an open source code license but allow little or no Ways to Debug Software and Systems. He is a Senior Member
code to be contributed back to their code base. of the IEEE and ACM. Contact him at dds@aueb.gr.
All 13 evaluation criteria I’ve outlined in Table 1 are
important, and taking them into account can spare
you unpleasant surprises and the cost of switching
from one project to another. Furthermore, you can use
Table 1 as guidance on how crucial some criteria are in

specific contexts. Specifically, those identified in the


first colored column can be deal breakers. In addi-

Cutting Edge
tion, if you identify problems with the yellow-marked
criteria in the second column, this means that you’ll stay
need to build in-house capacity to support the cor- on the
responding open source project. Finally, if you decide
to support the project with in-house resources, then
the green-marked components in the third column
of Artificial Intelligence
become more important. Ultimately, all of these J a n ua ry/ f E b r ua ry 2 016

IEEE Intelligent Systems provides peer-


IEEE

checks will help to ensure a long, happy, and prosper-


Also in this issue:
aI’s 10 to Watch 56
real-Time Taxi Dispatching 68
IEEE
January/FEBruary 2016

from flu Trends to Cybersecurity 84

ous relationship with the open source components


P U T T I N G A I I N T O P R A C T I C E

reviewed, cutting-edge articles on the


you’re selecting for your work. theory and applications of systems
Online BehAviOrAl AnAlysis

that perceive, reason, learn, and


ACKNOWLEDGMENTS act intelligently.
VOLuME 31

I thank Zoe Kotti and Alexios Zavras, who made many


nuMBEr 1

www.computer.org/intelligent

IS-31-01-C1 Cover-1 January 11, 2016 6:06 PM

helpful suggestions in an earlier version of this article.

REFERENCES The #1 AI Magazine


1. A. Morin, J. Urban, and P. Sliz , “A quick guide to
www.computer.org/intelligent
IEEE

software licensing for the scientist-programmer,” PLOS


Comput. Biol., vol. 8 , no. 7, pp. 1 –7, 2012.

www.computer.org/computingedge 19
EDITOR: Christof Ebert, Vector Consulting Services, christof.ebert@vector.com
This article originally
appeared in
DEPARTMENT: SOFTWARE TECHNOLOGY
vol. 35, no. 4, 2018

Digital Transformation
Christof Ebert and Carlos Henrique C. Duarte

FROM THE EDITOR


Digital transformation is a technology-driven continuous change process of companies and our entire
society. Its cornerstone is ubiquitous embedded computing, connectivity, and flexible value streams.
Obviously, there’s no digital transformation without software. But what are the underlying platforms?
What lessons can we take from actual projects? In this instalment of Software Technology, Carlos Duarte
and I highlight lessons learned from transformation projects and the underlying technologies and plat-
forms. I look forward to feedback from both readers and prospective article authors. —Christof Ebert

D
igital transformation (DX) is about adopt- partnerships with customers.5,6 The widespread
ing disruptive technologies to increase pro- implementation of DX will profoundly affect the indus-
ductivity, value creation, and the social try business environment—for example, by providing
welfare. Many national governments, multilateral orga- better value-chain integration and new-market exploi-
nizations, and industry associations have produced tation, with competitive-advantage gains.
strategic-foresight studies to ground their long-term DX is driven by a flood of software technologies.
policies. By proposing the implementation of public Embedded electronics such as microdevices with sen-
policies regarding DX, such groups expect to achieve sors and actuators connected through the IoT facilitate
the goals listed in Table 1. ubiquity. Data analytics, cloud storage and services,
DX is forecasted to have high annual growth and convergent interactivity and cognition, augmented
fast penetration.1–3 But there are barriers slowing its reality with visualization and simulation, pattern recog-
dissemination, such inadequate or overly heteroge- nition, machine learning, and AI are facilitating a con-
neous company structures or cultures, the lack of DX vergence of IT and embedded systems.2,5 Underlying
strategies and ROI (return on investment) visibility, these, we’ve identified enabling methods, techniques,
and even the perception of cannibalization of existing and tools, such as agile development for flexible sys-
businesses (the “innovator’s dilemma”4). External bar- tems, blockchains and Hyperledger to ensure security
riers also exist, such as the lack of recognition of how and trust in distributed transactions, and microser-
DX will benefit all of society, a shortage of skills and a vices and open APIs supporting software architectures.
qualified labor force, lacking or insufficient infrastruc- Let’s look at automotive technologies, where digi-
ture, missing or inadequate regulation and consumer tization is ramping up fast. A modern car incorporates
protection, and poor access to funding, particularly for 50 to 120 embedded microcontrollers and is connected
small and medium businesses. over various external interfaces to a variety of cloud and
infotainment technologies. Onboard software is in the
THE INDUSTRY PERSPECTIVE range of hundreds of millions of lines of code (MLOC)
ON DX and is still growing exponentially. Automotive software
Industry is moving to adopt holistic business models, product lines and variants are some of the largest and
completely redesign products and services, and estab- most complex in industry. It’s said that the automobile
lish closer interactions with suppliers and long-term is rapidly becoming a “computer on wheels.”

20 April 2020 Published by the IEEE Computer Society  2469-7087/20 © 2020 IEEE
SOFTWARE TECHNOLOGY

Automotive original-equipment TABLE 1. Digital transformation (DX) goals.


manufacturers (OEMs) are equip-
ping next-generation production
processes and vehicles with con-
nected embedded sensors and
actuators to obtain better intel-
ligence and control. They adapt
information and communication
technology workflows from their IT
systems to each car. Vertical inte-
gration is attained by ensuring that
product-lifecycle-management sys-
tems, enterprise-resource-planning
systems, production-planning-and-
control systems, and manufacturing-
execution systems work in coordi-
nation with capital goods on plant
floors. Concerning horizontal integration, vehicle customers formulate requests. Ten million sessions
parts are delivered with RFID tags to guarantee pro- have already been opened, and 94 percent of them
duction traceability. have been solved in an automated way.
OEMs work with suppliers that have the same Hospital Samaritano and Hospital Sírio-Libanês,
focus, to ensure that the acquired parts come with self- two leading São Paulo institutions, have consistently
or distance-monitoring facilities. Examples include invested in DX to improve the patient experience and
highly interconnected electronic control units from operational performance. Both keep integrated secure
companies such as Bosch, Continental, Denso and ZF; electronic health records of patients, which are used
mechatronic systems from Aptiv, Magna, Mahle, and in procedures, treatments, prevention, and healthcare
Schaeffler; and head units and infotainment systems planning and decisions.
from companies such as Harman, Valeo, Panasonic,
and Visteon. In factories, robots from ABB, Denso, DX IMPACTS
Kuka, and Yaskawa assemble complete vehicles from DX has been a source of continuous entrepre-
parts with exact monitoring and logging of, for exam- neurship and business dynamism, particularly in
ple, screw load torque to ensure compliance with pro- technology-intensive industries. These companies
duction and safety standards. All software is individ- have reorganized themselves to operate simultaneously
ually configured for each car by modern IT systems, in two distinct modes. The standard mode keeps tradi-
both in production and after sales with over-the-air tional businesses and operations running, while a dis-
upgrades. These movements toward a digital automo- ruptive mode seeks additional opportunities to exploit
bile world have already rationalized costs and invest- new markets and innovate in technologies, processes,
ments. For example, according to David Powels, former products, or services. Figure 1 illustrates that value is
CEO of Volkswagen’s Latin American operations, in the now created not only in traditional ways (the yellow
three-year period ending in 2016, the group obtained arrows) but also through digitization (the green arrows).
30 percent productivity gains in some factories, with a Software technology today is both the driver and
focus on digital process and competencies. 7 effect of disruption. The market leaders are ahead
Other industries are following fast toward DX. Vivo, of their competitors because they develop and com-
a company in the Spanish Telefonica group, is adopt- mercialize new technologies to address customers’
ing the agile-squad model and open innovation as the future performance needs. However, these compa-
bases of its DX implementation. The company devel- nies don’t want to cannibalize their current cash cows.
oped a social software robot called Vivi, which helps So, they’re rarely in the forefront of commercializing

www.computer.org/computingedge 21
SOFTWARE TECHNOLOGY

considered disruptive because they haven’t obtained


value recognition outside their initial markets (and
therefore aren’t included in Table 2). We structured our
findings in the form of a knowledge map,6,8 but here
we presented just some branches in textual, pictorial,
and tabular form. We hope this compilation will help
software engineering (SE) practitioners and research-
ers develop and implement DX.

THE MUTUAL INFLUENCES OF DX


AND SE
With software being key to any DX, mutual influences
between DX and SE must exist. But DX disruptions due
to SE innovations might emerge at any time and are
almost impossible to predict. So, we can only specu-
late about the implications for SE.
To perform initial verification and validation of our
ideas, we organized a debate panel during the 2016
IEEE Requirements Engineering Conference.5 During
FIGURE 1. Digital transformation is a convergence of hard the panel, researchers and practitioners discussed the
and soft forces and movements from which additional value impact and relationship of requirements engineering
emerges. Value is now created not only in traditional ways (the and DX in industry and research institutions. The pan-
yellow arrows) but also through digitization (the green arrows). elists came from companies such as Intel (US), Nokia
(Finland), Denso (Japan), Civic (China), and CI&T (Brazil).
The participants agreed that software technolo-
new technologies that don’t initially meet the needs of gies at the core of DX disruptions have been around for
mainstream customers and that appeal to only small some time. These technologies have caused disrup-
or emerging markets. tions because of
So, disruptive companies explore the occupation
gaps left by the market leaders. This is a source of inno- ›› early or timely value delivery (agile methods),
vation and market change, which Clayton Christensen ›› usage at larger scales (APIs, microservices, and
illustrated using price and performance data from the IPv6),
hard-disk-drive industry.4 ›› applications in new domains (3D modeling and
We performed a systematic classification of the DX printing, control software, and blockchains), and
technology offerings; Table 2 presents some details. ›› unpredicted technology combinations (cog-
Although we haven’t compiled quantitative evidence of nitive computing, which combines computer
the disruptions caused by the analyzed hardware tech- vision, voice recognition, natural-language pro-
nologies, Organisation for Economic Co-operation and cessing, and machine learning).
Development studies have recognized that robots, 3D
printing, and connected devices have disrupted pro- So, DX has not led to the development of radically
ductivity in their respective markets.1,2,5 new software technologies. Instead, it has given rise
Likewise, some of the software technologies we to new software technology applications, owing to the
studied have been disruptive, but this is due only to additional requirements that must be satisfied.
their strategic significance, resulting from their pene- Technological solutions’ complexity and scale have
tration, adoption, and perceived value in distinct mar- increased substantially, leading to software systems
ket segments vis-à-vis the initial markets. Finally, the with many MLOC, usually binding together old and
remaining software technologies we studied can’t be new, in-house and third-party developments. However,

22 ComputingEdge April 2020


SOFTWARE TECHNOLOGY

TABLE 2. DX technologies.

Technology Inherent nature and Distribution and Early-adopter Adopted Ease of No. of
type attributes significance experience technology adoption alternatives

URLs

www.computer.org/computingedge 23
SOFTWARE TECHNOLOGY

the industry objectives for DX—an improved customer completely reshape the landscape of software tech-
experience and operational excellence—have placed nologies and processes. (For more on DX and systems
time to market, quality, and affordability at the fore- engineering, see the sidebar.)
front. So, practical DX problems have become tractable With industry, home, healthcare, and automotive
with software only with effective development man- applications being major drivers, IT will converge with
agement, reusability, and requirements-engineering embedded systems such as the IoT and Industry 4.0. At
methods, techniques, and tools. the same time, embedded industries will evolve toward
The SE branches we’ve described have many inter- IT with cloud solutions and dynamic over-the-air
faces, which deal with unproven metrics, hard com- upgrades. Critical industries such as the automo-
plexity bottlenecks, and imprecise artifacts. The tive industry involve practically all the quality require-
human factor is central to addressing these issues, ments, such as safety, cybersecurity, usability, per-
but the required key competences for problem solving, formance, and adaptability. The underlying software
managing complexity, and dealing with high abstrac- components cover anything from embedded real-time
tion levels are often lacking or insufficient. firmware to complex secured cloud solutions. Failure
DX requires software engineers to organize their to meet any of those quality requirements results in
work efficiently, act on their own initiative, have excel- expensive callback actions and legal lawsuits. These
lent communication skills, and successfully perform challenges will soon reach across industries.
tasks involving emotion, intuition, creativity, judg- DX is opening the doors for technology innovation,
ment, trust, empathy, and ethics.2 new business models, and cross-industry collabora-
At higher organizational levels, SE managers are tion. The future is arriving while some are just running
expected to change their mind-sets and abandon com- in their hamster wheels. Thus, we should be cautious,
mand and control, moving to more leadership-oriented, along the lines of what technology strategist Herman
risk-taking, and mistake-tolerant approaches. Cor- Kahn already observed several decades ago: “Every-
porate leaders need to motivate, direct, support, and body can learn from the past. Today it’s important to
inspire their autonomous teams, while learning along learn from the future.”9
with them. They must be prepared to face business
environments in which hyperawareness, informed ACKNOWLEDGMENTS
decision making, and fast execution rule. The assumptions, views, and opinions in this article
How can people obtain such skills? DX challenges are solely the authors’ and don’t necessarily reflect
traditional SE education systems to change their meth- the official policy, strategy, or position of any Brazilian
ods and content to a digitally transformed reality. Apart government entity.
from the classroom and learning-by-doing approaches,
continuous, just-in-time, and innovative learning meth- REFERENCES
ods—such as massive open online courses, gamifica- 1. H. Demirkan, J.C. Spohrer, and J.J. Welser, “Digital Inno-
tion, and simulation—will be increasingly demanded. vation and Strategic Transformation,” IEEE IT Profes-
To meet DX demands, SE will transform completely, sional, vol. 18, no. 6, Nov. 2016, pp. 14–18.
leading to changes in how SE education treats human 2. C. Ebert, “Looking into the Future,” IEEE Software, vol.
factors. In this new scenario, human resources will be 32, no. 6, 2015, pp. 92–97.
extremely valuable, possibly becoming more impor- 3. M. Gebhart, P. Giessler and S. Abeck, “Challenges of
tant than the underlying technologies. the Digital Transformation in Software Engineering,”
Proc. 11th Int’l Conf. Software Eng. Advances (ICSEA 16),

D X today is the megatrend across industries. How-


ever, DX is challenging because it demands a new
set of competences, combining embedded-systems
2016, pp. 136–141.
4. C.M. Christensen, The Innovator’s Dilemma: When New
Technologies Cause Great Firms to Fail, Harvard Business
development with IT and cybersecurity. Software thus Rev. Press, 2013.
is the cornerstone of DX. In its convergence of clas- 5. C. Ebert and C.H.C. Duarte, “Requirements Engineer-
sic IT with embedded-systems engineering, DX will ing for the Digital Transformation: An Industry Panel,”

24 ComputingEdge April 2020


SOFTWARE TECHNOLOGY

DIGITAL TRANSFORMATION AND SYSTEMS ENGINEERING

Christof Ebert

A t Vector Consulting Services, we’ve supported


many companies in their journey toward digital
transformation (DX), with both software technologies
fusion, picture recognition, and data analytics for
automated processes such as autonomous driving,
medical-surgery support, and industry-scale predic-
and the necessary agile business and development tive maintenance; and
techniques. One key observation from projects is the »» system-level modeling, testing, and simulation with
fast-growing relevance of requirements engineering models in the loop, and ensuring quality requirements
and systems engineering. such as safety.
Requirements engineering, in its link with systems
engineering, is the decisive success factor for the effi- As complexity and scale increase, quality must be
cient development of hardware and software systems. ensured end-to-end at the system level. This demands
Agile systems engineering supports the continuous professional methods and tools to ensure robustness,
development of requirements up to validation. It creates dependability, functional safety, cybersecurity, and
understanding of critical dependencies and provides usability. Security and robustness tremendously affect
methodological support to address growing complex- business models and potential liability. The more we
ity. Because systems engineering provides an increased share and network, the more we’re exposed to attacks of
number of abstraction levels and consistent handling of all kinds. Usability is an interesting example of how qual-
dependencies across a system, specifications become ity factors are increasingly crucial. Across industries
clearer, simpler, and less redundant. This not only and applications, insufficient usability is a major source
increases development speed but also ensures clearly of hazards, operational failures, and critical failures.
understood domain concepts in a project. Because quality, deadlines, and cost are pivotal across
Systems engineering for DX includes industries, the push for even better processes and project
management is continuing at a fast pace. The demand for
»» cloud services for environment awareness, location- more agility and flexibility is rising. System users expect
based services, online apps, remote diagnosis, con- the same adaptive behaviors and continuous-delivery
tinuous software updates, and emergency functions; models that they have with their mobile devices. Rapid
»» cybersecurity, usability, and performance for modern advances toward autonomous driving and open vehicle
software systems; communication demand short-cycle recertification after
»» service-oriented advanced OSs with secure commu- over-the-air software updates. We’ve observed that only
nication platforms; novel development paradigms such as service-oriented
»» machine learning and AI—for instance, in multisensor architecture can cope with the growing need for flexibility.

Proc. 24th IEEE Int’l Requirements Eng. Conf. (RE 16), Speculation on the Next Thirty Years, Macmillan, 1967.
2016, pp. 4–5.
6. C.H.C. Duarte, “Patterns of Cooperative Technology Devel- CHRISTOF EBERT is the managing director of
opment and Transfer for Software-Engineering-in-the- Vector Consulting Services. He is on the IEEE
Large,” Proc. 2nd Workshop Software Eng. Research and Software editorial board and teaches at the
Industrial Practice (SER&IP 15), 2015, pp. 32–38. University of Stuttgart and the Sorbonne in
7. C. Silva, “‘Na crise, nossa produtividade cresceu 30%’” Paris. Contact him at christof.ebert@vector.com.
(“‘In the crisis, our productivity grew 30%’”), Estado de
São Paulo, 9 Dec. 2017 (in Portuguese); http://economia CARLOS HENRIQUE C. DUARTE is a senior
.estadao.com.br/noticias/negocios,na-crise-nossa member of the technical staff at the National
-produtividade-cresceu-30,70001976321. Bank of Social and Economic Development in
8. M.E. Fayad et al., Software Patterns, Knowledge Maps, Rio de Janeiro, and an IEEE Computer Society
and Domain Analysis, CRC Press, 2014. Distinguished Visitor. Contact him at cduarte@bndes.gov.br
9. H. Kahn and A.J. Weiner, The Year 2000: A Framework for and carlos.duarte@computer.org.

www.computer.org/computingedge 25
EDITOR: Stephen J. Andriole, Villanova University School of Business, steve@andriole.com
This article originally
appeared in
COLUMN: LIFE IN THE C-SUITE
vol. 20, no. 3, 2018

Governing and Piloting


Emerging Technologies
Stephen J. Andriole, Villanova University School of Business

T
he governance of emerging technologies is processes and simulations of how modified processes
different than the governance of operational can lead to more efficiency, lower costs, and digital
technologies. No one governs plumbers the transformation. You cannot improve what you cannot
way they govern architects, which is not to derisively model or optimize. B/RPA is the first step to technol-
equate databases, enterprise resource systems, or ogy optimization.
communications networks with plumbing, but rather
to equate architects with creativity. Artificial intelligence (AI)
C-suiters must first identify the emerging tech- and machine learning.
nologies that might disrupt their businesses. Then These technologies can shorten, improve, and replace
they must decide which ones to pilot to determine routine and important processes and tasks, and
just how disruptive each might be to their business can save you a lot of money by reducing headcount
processes and whole business models. The gover- and improving throughput. The technology matters
nance of emerging technologies thus consists of because it represents alternative methods, tools, and
“identify-and-pilot,” rather than the traditional gover- techniques for collecting, organizing, and analyzing
nance of “identify-deploy-manage.” data and information, and generating deductive and
inductive knowledge. Put another way, AI can mimic—
EMERGING TECHNOLOGIES AND and improve—your best salespersons, underwriters,
WHY THEY MATTER and brokers, among other professionals in your orga-
There are at least 10 emerging technology clusters nization. AI also enables conversational speech, which
C-suiters should track (and possibly pilot). means you will soon be able to augment intelligent
bots, which you’re probably already using. AI can also
Business and robotic process diagnose problems and suggest solutions in real time
automation (B/RPA). across vertical industries. The results of your BPA and
These are essential to identifying, describing, and RPA projects will suggest exactly where “intelligence”
improving the processes that run your business. If you might be leveraged.
don’t know what your processes look like and how they
behave, you cannot improve them by leveraging exist- Blockchain.
ing or emerging technologies. B/RPA is a blueprint for This technology enables trusted, verifiable transac-
change—change is sometimes good and sometimes tions without the cost or complexity of transactional
bad, but you must know the details of your processes intermediaries. According to the Forbes Technol-
and how they might be modified to make you more ogy Council, “blockchain is a public register in which
efficient and more competitive. BPA maps the key pro- transactions between two users belonging to the
cesses and RPA mimics the processes in simulated same network are stored in a secure, verifiable and
software functions. They are both enabled by auto- permanent way. The data relating to the exchanges
mated tools that facilitate the description of existing are saved inside cryptographic blocks, connected in

26 April 2020 Published by the IEEE Computer Society  2469-7087/20 © 2020 IEEE
a hierarchical manner to each other. This creates an engine monitoring and maintenance, and refrigera-
endless chain of data blocks—hence the name block- tors that automatically fill your Instacarts. The key is
chain—that allows you to trace and verify all the trans- two-way connectivity and the data exchanges enabled
actions you have ever made … in the case of Bitcoin, by this connectivity. If your company has opportuni-
the blockchain serves to verify the exchange of cryp- ties to connect people, processes, devices, and com-
tocurrency between two users, but it is only one of the ponents in real time with diagnostic information, then
many possible uses of this technological structure. In the IoT matters.
other sectors, the blockchain can certify the exchange
of shares and stocks, operate as if it were a notary and
‘validate’ a contract or make the votes cast in online AUGMENTED ANALYTICS IS
voting secure and impossible to alter.”1 POTENTIALLY OF HUGE IMPORTANCE
TO ALL ANALYTICS EFFORTS ACROSS
Cryptocurrency. ALL VERTICAL INDUSTRIES.
This is a virtual (digital) currency that relies on cryp-
tography to make the currency safe and secure. Cryp-
tocurrency is not issued by a central bank or a gov- Augmented analytics.
ernment that guarantees its authenticity. You likely This is analytics 2.0, where “regular” analytics
already know about Bitcoin and probably Ethereum. (advanced statistical analysis focused on describ-
You might also know that cryptocurrency is gaining ing, explaining, predicting, and prescribing events and
transactional traction: you can buy houses with Bit- conditions) is augmented by AI tools, such as machine
coin.2 Cryptocurrency can also be used to execute learning, natural language processing, and neural net-
transactions you might want to hide. Depending on work modeling. This augmentation promises to accel-
the business you’re in, cryptocurrency might already erate descriptive, explanatory, predictive, and pre-
be an optional currency. Watch adoption rates closely. scriptive analytics. It also enables the processing of
If they rise sharply, you might have to respond accord- massive databases and the production of heretofore
ingly. That said, it’s important to track methods, tools, undiscovered deduction and induction. Augmented
and techniques that enable transactional autonomy. analytics is potentially of huge importance to all ana-
lytics efforts across all vertical industries. If you’re
Internet of Things (IoT). investing in analytics, you will find value in augmenting
There’s value in connecting everything in your com- your investment with AI. If you’re investing in both ana-
pany (and supply chain) with sensors that can “talk” lytics and AI, then the offspring of the marriage should
to everything else. The data, information, and knowl- be obvious.
edge collected and exchanged among active sen-
sors in real time (if necessary) can be as diagnostic Virtual and augmented reality.
as you need. The IoT enables all sorts of applications These technologies enable you to change the per-
like smart buildings, smart cities, smart power grids, spective and the experience of your customers and

www.computer.org/computingedge 27
LIFE IN THE C-SUITE

all those who comprise your value chain. In simple PILOTING


terms, augmented reality (AR) facilitates the overlay- C-suiters should select the technologies most likely
ing of aspects of the digital world onto the real world. to impact their businesses and industries and deter-
Virtual reality (VR), on the other hand, simulates a mine which ones to pilot. The first step is to conduct
completely virtual world. The opportunities for sales some industry intelligence to see what others in your
and marketing are extensive, especially if you’re sell- industry are doing. Vendors are a rich source of infor-
ing physical objects whose features can be enhanced mation, as are professional technology forecast-
with augmented reality or represented digitally in vir- ers like Gartner, Forrester Research, and the Cutter
tual reality. Remember that consumers of AR and VR Consortium.
are demographically available and that the form factor Executives should also structure pilots narrowly
will eventually include conventional glasses, making to determine the impact deployments might have on
the technology much more accessible to consumers. the targeted business processes and models. Vendors
should be asked to sponsor pilots (with appropriate
Mobile everything. expectations about future work should the pilots
If you understand value-chain demographics, you succeed).
already know that mobile computing is ubiquitous. It The first pilot should be a B/RPA project designed
will continue to grow as more and more data is trans- to map corporate processes and whole business
mitted across multiple devices, though most of the models, which will identify the best opportunities for
growth will be with smartphones (and future incarna- leveraging emerging technologies for corporate gain.
tions of today’s smartphones). The exploding adoption After that, C-suiters can pilot the most promising tech-
of mobile applications is driving all this growth. Many nologies informed by rigorous process mapping.
business processes already exist on your customers’
and suppliers’ phones. It’s safe to say that more and REFERENCES
more of your business processes will be delivered and 1. J. Giordani, “ Blockchain—What Is It And What Is It
optimized on mobile devices. For?,” Forbes, blog, 28 March 2018 ; www.forbes
.com/sites/forbestechcouncil/2018/03/28/blockchain
Wearables. -what-is-it-and-what-is-it-for/#57f3e49f1a16.
Wearables come in various forms, including clothes, 2. S. Mishkin , “ You Can Buy This Stunning Southern
jewelry, shoes, wristwatches, and hats, not to men- California Lake House For 32 Bitcoins,” Money, blog,
tion all the fitness trackers out there. Some wearables 19 December 2017; http://time.com/money/5055888
are embedded under the skin of humans and animals. /real-estate-bitcoin.
Many transmit data continuously. The possibilities
here are endless. Any person, pet, or object can be STEPHEN J. ANDRIOLE is the Thomas G. Labrecque Profes-
“activated” with a wearable. sor of Business Technology in the Villanova School of Busi-
ness at Villanova University, where he teaches courses in
Cybersecurity and privacy. strategic technology and innovation and entrepreneurialism.
First and foremost, you have no choice but to contin- Contact him at steve@andriole.com.
uously and heavily invest in cybersecurity. Cybersecu-
rity spans your networks, databases, and applications.
Breaches are expensive and can severely damage your
brand. Privacy is an expectation, though here demo-
graphics are on your side. Millennials are less con-
cerned with privacy than their parents and grandpar-
ents. That said, C-suiters need to watch what happens
with the rollout of the General Data Protection Regu-
lation (GDPR) across Europe, which went into effect in
WWW.COMPUTER.ORG/COMPUTINGEDGE
May 2018.

28 ComputingEdge April 2020


DEPARTMENT: IT ECONOMICS This article originally
appeared in

Big Data Solutions for Micro-, vol. 21, no. 5, 2019

Small-, and Medium-Sized


Enterprises in Developing Countries
Diana Rojas-Torres, Universidad de La Sabana, Colombia
Nir Kshetri, University of North Carolina at Greensboro

T
he use of big data solutions to make better and Crop Accurate system uses data from diverse sources
faster business decisions is no longer limited such as satellite, drone, and sensors for smart farm-
to large organizations. In recent years, micro-, ing. The system analyzes soil condition, weather, and
small-, and medium-sized enterprises (MSMEs) in growth progress to give farmers advice regarding the
developing countries are increasingly benefiting from best time to plant, fertilize, and use pest control. Farm-
big data solutions (see Table 1). Such solutions have ers can make more efficient use of fertilizer and pesti-
helped developing world-based MSMEs to improve cides.3 CI-Agriculture also learned about local farming
business processes and market intelligence. Big data practices and supply chains to develop the system.4
solutions have also helped them to increase access to The technology is scalable, which means that it is pos-
financial services, such as loans, credit, and insurance.1 sible to use sensors for a large area. Agricultural data
MSMEs make a significant contribution to national are collected and analyzed on a regular basis to pre-
economic development in developing countries. The dict crop yields. At the end of each season, smart farm-
developing world is estimated to have 365–445 mil- ing system analyzes the data and provides recommen-
lion formal and informal MSMEs (https://tinyurl.com dation to improve farming in the next season.
/y5vffr3p). Formally registered MSMEs are estimated CI-Agriculture's another solution, Agritrack sys-
to contribute up to 45% of total job creation and 33% tem, links farmers with supply chain partners such as
of gross domestic product in developing economies. distributors, market, and end customers. 5 Each party
These proportions would dramatically increase when of the supply chain provides data via an app. Real-time
informal MSMEs are included. MSMEs are likely to play information on key indicators, such as commodity
especially important roles in reducing rural poverty prices, is provided, which can help predict prices and
among women and other disadvantaged groups.2 Big demand of farmers’ produces.
data diffusion among MSMEs is, thus, likely to bring To take another example, China's Alibaba has
tremendous economic and social benefits to develop- attracted vendors to its e-commerce websites Tao-
ing countries. bao Marketplace and Tmall.com by promoting big
data-based advertising and other services. These solu-
BIG DATA SOLUTIONS TO IMPROVE tions provide deep insights into shoppers’ preferences.6
BUSINESS PROCESSES AND Taobao has 666 million monthly active users (https://
MARKET INTELLIGENCE fortune.com/longform/ping-an-big-data) and over nine
An important use of big data solutions has been in million vendors (https://www.azoyagroup.com/blog
improving MSMEs’ business processes and market /index/view/chinas-new-e-commerce-law-bad-for
intelligence. To take an example, Collective Intelli- -daigou-good-for-cross-border-e-commerce/).
gence Agriculture (CI-Agriculture), a subsidiary of Indo-
nesia's big data analytic firm Mediatrac, has developed
precision farming techniques for the Indonesian con- DOI No. 10.1109/MITP.2019.2932236
text. Small holder farmers can use the technique. Its Date of current version 11 September 2019.

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 29
IT ECONOMICS

TABLE 1. Some examples of big data solutions from MSMEs in developing countries.

Launched by Big data solution Key functions

Alibaba has also developed a big data-based behavioral analytics (https://tinyurl.com/y66awka5).


retail-management platform known as Ling Shou Tong The information is used to generate credit scores and
for small physical stores in China. The solution aims to assess their creditworthiness for loans (https://tinyurl
help store owners in making decisions related to prod- .com/yyun3mes). Banks can use this information to
uct procurement and sales. In 2017, Alibaba started pro- provide loans to farmers and customize a farmer's pay-
viding the platform to Chinese retail shops. The shops back timeline in order to match with harvests. As of
get the platform for “free” but they are required to use 2018, over 200,000 farmers were using DigiFarm on a
their storefronts as Alibaba's fulfillment-and-delivery daily basis and 7,000 had successfully received loans
centers. They also need to provide data on their cus- to buy seeds, fertilizers, and pesticides (https://tinyurl
tomers’ shopping habits and patterns (https://tinyurl .com/yyx4u8sz).
.com/y5mpzgle). Likewise, big data solutions of Indonesia's CI Agri-
culture discussed above are expected to reduce loan
BIG DATA SOLUTIONS TO IMPROVE costs for small-holder farmers. Data from satellite,
ACCESS TO KEY RESOURCES drone, and sensors are used to calculate a field's pro-
Big data solutions can also help to improve MSMEs’ duction potential with a higher level of accuracy. These
access to key resources. Especially access to credit is data can also be utilized to make more efficient use
extremely difficult for smallholder farmers in the devel- of fertilizer and pesticides.3 CI-Agriculture provides
oping world. For instance, less than 1% of farmers in insurance to farmers, which is based on calculations
Kenya are reported to have access to formal credit.7 To and schemes on smart farming technology, sensor sys-
address this, the Kenya-based social enterprise Farm- tems, and analysis of other categories of data (https://
Drive's big data solution DigiFarm helps unbanked tinyurl.com/yxrtgn2o). Insurance models are based on
and underbanked smallholder farmers to receive an analysis of weather data for up to 10 years.4
credit. The process is simple. Smallholder farmers
keep a record of their revenues and expenses. An app KEY CHALLENGES AND
installed in the phone tracks these records. This infor- OPPORTUNITIES
mation is combined with data generated from other MSMEs in developing countries have a number of
sources such as satellite, agronomic data such as crop options available to utilize big data solutions. MSMEs
yields, pests and diseases, and local economic data. 8 can benefit from open-source software such as Hadoop
In addition to agriculturally relevant data, DigiFarm and Spark. For instance, Hadoop-based applications
also uses know your customer data to identify and help MSMEs take advantage of real-time analytics
verify the identity of the farmer as well as advanced from diverse sources and types of data. These include

30 ComputingEdge April 2020


IT ECONOMICS

data from external sources such as social media, to utilize big data effectively. Indeed, even large enter-
machine generated data, as well as data from video, prises face such challenges in developing countries.
audio, email, and sensors. Many global technology Especially there has been a severe lack of big data
companies such as Microsoft, IBM, EMC, Google, and manpower with high-level strategic thinking capabili-
Amazon Web Services provide Hadoop-as-a-Service ties in developing countries. For instance, compared
(HaaS) to MSMEs in many developing countries. The to many other developing countries, China has a rich
HaaS providers help MSMEs in the management, endowment of big data human resources, thanks to an
analytics, and storage of data (https://tinyurl.com abundant supply of engineers. The country, however,
/y5vagmu7). MSMEs are increasingly adopting HaaS lacks experts at the executive level (https://tinyurl
(https://tinyurl.com/y23hhj4r). These and other big .com/y5ayxkwl). Likewise, the lack of strategic leader-
data solutions targeted at MSMEs are getting more ship and the lack of idea of where to start the imple-
user-friendly. This is a rich community of users. Tutori- mentation of solutions are noted as a main reason why
als, tools, and other services are more easily accessi- Colombian companies have not taken advantage of
ble to MSMEs (https://tinyurl.com/yyun3mes). big data.11
MSMEs obviously experience a number of bar- More broadly, the big data labor market in develop-
riers in the adoption of big data solutions. Big data ing economies faces challenges on two fronts.5 First,
solutions provided by big companies are unaffordable there is a severe lack of engineers and scientists in
and out-of-reach for many MSMEs in developing coun- order to perform analytics. Second, many analytics
tries. For instance, it was reported that due primarily consultants lack skills and capabilities to understand,
to Alibaba's high advertising rates, most vendors on interpret, and put the data to work. Some estimates
Taobao were making losses.9 According to an article suggest that India would experience a shortage of 1 mil-
published in the Chinese language newspaper Enter- lion data consultants (http://www.techrepublic.com
prise Observer in August 2013, over 80% of sellers on /article/indias-high-demand-for-big-data-workers
Taobao did not make a profit. It was also reported that -contrasts-with-scarcity-of-skilled-talent/).
thousands of shops on Taobao close down every day.10
In many cases the benefits to MSMEs of so-called SUMMARY
free big data solutions provided by technology giants The reliance on big data to make better and faster
are not clear. The providers of such solutions tend to decisions is, thus, no longer limited to large compa-
use them as a useful instrument to promote their own nies. While only a tiny fraction of MSMEs in the devel-
interests rather than those of MSMEs’. For instance, oping world are currently taking advantage of big data
while the use of Ling Shou Tong may make it easier solution, such solutions are getting popular among
to run stores, many small stores worry about unfair these enterprises. They increasingly depend upon big
competition from Alibaba's online marketplace, which data. Data-driven decisions are gradually becoming
has a huge selection of products to choose from. the norm among these enterprises.
This means that these small stores’ customers may There are growing and encouraging signs of big
decide to take advantage of the convenience of online data's positive impacts on MSME in the developing
shopping on Alibaba's online marketplace and pick up world. Big data-based innovations such as low-cost
the products from these stores (https://tinyurl.com crop insurance and low-cost loans have benefitted
/y5mpzgle). Furthermore, Alibaba is in a position to micro enterprises and small holder farmers in the
make a better utilization of data on these stores’ cus- developing world. In addition to access to these stra-
tomers that these stores are required to provide. tegically valuable resources such as finance and insur-
Effective utilization of big data requires orga- ance, big data-based solutions have also increased
nizational capability to handle cooperation across the quality of their entrepreneurial activities with
different units and departments. Organizations in improved business processes and market intelligence.
many developing countries may lack capabilities to Since MSMEs are critical for job creation and
organize and manage such multidisciplinary teams. A economic growth, it is important to deploy policy
further challenge is MSMEs' lack of human resources measures to facilitate the adoption of big data by

www.computer.org/computingedge 31
IT ECONOMICS

these enterprises. For instance, broad national policy ecommerce to China,”Financial Times, Blog. 2014.
measures directed toward increasing competition [Online]. Available: https://www.ft.com/content
in the big data industry may lead to the availability of /11022ce8-a61a-11e3-8a2a-00144feab7de
affordable solutions to MSMEs. This can be done by #axzz3V2tjJstS
attracting foreign big data companies and facilitating 10. R. Lu, “Tea leaf nation: For Alibaba’s small business
the growth of local companies in this sector. army, a narrowing path,”Foreign Policy, Blog. 2014.
[Online]. Available: https://foreignpolicy.com/2014
REFERENCES /09/10/for-alibabas-small-business-army-a-narrowing
1. N. Kshetri, et al., Big Data and Cloud Computing -path/
for Development: Lessons from Key Industries and 11. B. Mario Augusto, “En Colombia apuestan por una
Economies in the Global South. New York, NY, USA: mejor utilización del big data,”Feb. 1, 2016. [Online].
Routledge, 2017. Available: http://www.cioal.com/2016/02/01/en
2. “SMEs and SDGs: Challenges and opportuni- -colombia-apuestan-por-una-mejor-utilizacion-del
ties, ”Apr. 23, 2019. [Online]. Available: https:// -big-data/
oecd-development-matters.org/2019/04/23/smes
-and-sdgs-challenges-and-opportunities/ DIANA ROJAS-TORRES is an Assistant Professor of entre-
3. F. Kuo, “Can technology save ASEAN’s food supplies preneurship and innovation with the International School
from climate change?” The Interpreter, Blog. 2019. of Economic and Administrative Sciences, Universidad De
[Online]. Available: http://www.lowyinterpreter.org La Sabana, Chia, Colombia. Contact her at diana.rojas7@
/the-interpreter/can-technology-save-asean-s-food unisabana.edu.co.
-supplies-climate-change
4. N. Freischlad, “Drones over the rice paddy: NIR KSHETRI is a Professor of management with the Bryan
Ci-Agriculture brings smart tech to the field,” Techina- School of Business and Economics, University of North
sia, Blog. 2015. [Online]. Available: https://www Carolina at Greensboro, Greensboro, NC, USA. Contact him
.techinasia.com/ci-agriculture-precision-farming at nbkshetr@uncg.edu.
-indonesia
5. N. Kshetri, Big Data’s Big Potential in Developing
Economies: Impact on Agriculture, Health and
Environmental Security. Wallingford, U.K., Centre
for Agriculture and Biosciences International (CABI)
Publishing, 2016.
Call for Articles
6. L. Lorenzetti, “Alibaba’s first public earnings reveal
IEEE Software seeks
major revenue growth,” Fortune, Blog. 2014. [Online]. practical, readable articles
Available: https://fortune.com/2014/11/04/alibabas that will appeal to experts

-first-public-earnings-reveal-major-revenue-growth/ and nonexperts alike. The


magazine aims to deliver
7. S. Maina, “Safaricom’s DigiFarm aims to put more coins
reliable information to software
in farmers’ pockets through technology,”Techweez,
developers and managers to
Blog. 2018. [Online]. Available: https://techweez help them stay on top of rapid
.com/2018/07/23/ technology change. Submissions

safaricom-digifarm-more-coins-farmers/ must be original and no more


than 4,700 words, including 250
8. L. Burwood, “FarmDrive raises funding to help Africa’s
words for each table and figure.
smallholder farmers get finance with credit scoring
algorithm,”Agfundernews, Blog. 2017. [Online]. Avail-
able: https://agfundernews.com/farmdrive Author guidelines:
www.computer.org/software/author
-raises-funding-to-help-africas-smallholder-farmers Further details: software@computer.org
-get-finance-with-credit-scoring-algorithm.html www.computer.org/software
9. C. Clover, “Alibaba has almost single-handedly brought

32 ComputingEdge April 2020


EDITOR: Yong Cui, cuiyong@tsinghua.edu.cn
This article originally
appeared in
COLUMN: STANDARDS
vol. 22, no. 3, 2018

Analytics without Tears or Is


There a Way for Data to Be
Anonymized and Yet Still Useful?
Jon Crowcroft, University of Cambridge and The Alan Turing Institute
Adrià Gascón, Warwick University and The Alan Turing Institute

In this article, we discuss the new requirements for standards for policy and mechanism to
retain privacy when analyzing users’ data. More and more information is gathered about all
of us, and used for a variety of reasonable commercial goals—recommendations, targetted
advertising, optimising product reliability or service delivery: the list goes on and on. However,
the risks of leakage or misuse also grow. Recent years have seen the development of a number of
tools and techniques limit these risks, ranging from improved security for processing systems,
through to control over what is disclosed in the results. Most of these tools and techniques
will require agreements on when and how they are used and how they inter-operate.

HISTORICAL AND TECHNICAL where. User data might be valuable in many contexts,
CONTEXT especially when aggregated from several sources.
We have always kept data about ourselves—maybe This created a market for data, making suitable for
household accounts to divvy up the food bill among the Internet a point made in the 70s regarding televi-
students in a shared flat each month, or maybe our sion1: “If you are not paying for the product, you are the
baby’s weight and height. It’s often easier to let some- product.”
one else look after that data, a bank or our doctor for In this context, we’ve experienced a shift in how
example, since they can use it for our benefit and keep it decision-making processes are approached by enter-
safe. They then store many peoples’ records, and need prises and governments: crucial decisions in areas as
a way to find our particular one, via some primary key, diverse as policy making, medicine, law enforcement,
an identifier that uniquely fishes out our information— banking, and the workplace are informed to a great
perhaps a mix of our name, birthday and postcode. extent by data analysis. For example, decisions related
The two-sided market of cloud analytics emerged to which advertisements and promotions we see
almost accidentally, initially from click-through associ- online, which of our incoming emails are discarded,
ated with users’ response to search results, and then what type of TV series are produced, what conditions
adopted by many other services, whether webmail or are attached to our insurance, or what new drugs are
social media. The perception of the user is a free ser- developed, are all informed to a great extent by the
vice (storage and tools for photos, video, social media, analysis of electronic data.
etc.) with a high-level of personalization. The value to This trend is far from stabilizing. As digital sur-
the provider is untrammeled access to the users’ data veillance grows apace, the advent of personal data
over space and time, allowing upfront income from the gleaned not just from social media and online services
ability to run recommenders and targeted adverts, to but from sensors in smart homes, cars, cities, and
background market research about who is interested health devices, becomes more and more intrusive.
in what information, goods, and services, when and Something has to give.

2469-7087/20 © 2020 IEEE Published by the IEEE Computer Society April 2020 33
There are both technical and socio-economic evidence-building community to combine data and
reasons why this has to change, and this has been rec- conduct analysis without directly accessing or storing
ognized amongst regulators and in industry. Privacy information.”
failures risk personal and corporate wealth and safety. In general, similarly to general security, privacy
Theft of credit card data, identity and trade secrets is a has proven to be a slippery concept, that requires a
real and present danger, increasing with the extended robust, mathematically rigorous approach. Neverthe-
“attack surface” presented in the surveillance society. less, very promising advances have been made in the
The volume of detail available admits of inference last decade, both from a practical and a theoreti-
about people and institutes in ways not recognized cal perspective. In this scenario, privacy-preserving
or necessarily intended. In some cases, this even can analytics has emerged as a very active research topic
lead to threats to personal safety. As a result, new laws simultaneously in several fields such as machine learn-
and new technology have been proposed and enacted, ing, databases, cryptography, hardware systems, and
which might go some way toward alleviating this. statistics. The main challenges include several appli-
However, the roadmap is quite complex and involves cations currently being simultaneously pursued within
choices, as well as agreements between parties. Some research communities in these fields, such as finding
of the technical choices will have implications for secure ways of providing public access to private
standards, which may in turn reflect back on how regu- datasets, securely decentralizing services that rely on
lation and legislation will evolve. One thing appears private data from individuals, enabling joint analyses
clear, that organizations are keen to retain the value on private data held by several organizations, and
of all that big data now being gathered and analyzed, securely outsourcing computations on private data.
whether for entertainment, health, security, or profit. There are several alternative directions that this
Hence there is a need for careful harmonization among will evolve in the future.
the new regulations and the new technology that can
support the old value chains. ASPECTS OF PRIVACY-
The General Data Protection Regulation (GDPR) PRESERVING ANALYTICS:
and the ongoing e-privacy regulation effort are sig- COMPUTATION AND DISCLOSURE
nificant steps in regulating the protection of sensitive The general goal of research into privacy-preserving
information by placing obligations on data control- data analysis is to develop techniques that get the
lers and data processors, as well as specifying user’s best utility out of a dataset without violating the pri-
rights. However, no specific algorithms are mentioned, vacy of the individuals represented in it. However,
and hence we are far from effective standardization there are several interpretations of what we may mean
guidelines. This is justified, as the technology is not by privacy in this context.
quite there, and more research is needed both from First of all one has to realize that if you belong to a
a theoretical and applied perspective. However, its certain population, and an analysis on that population
potential has been recognized both in industry and is disclosed, then your privacy has been breached and
government. The recent report by the US Commission there is nothing you can do—or could have done—
on Evidence-based Policymaking 2 describes differen- about it. For example, assume that a predictive model
tial privacy and multi-party computation as emerging about mobility in London is made public. Let’s say that
technologies and states, “New privacy-protective the model is able to accurately predict the location of
techniques [...] may allow individuals in the Federal London tube users, given some of their characteristics.

34 ComputingEdge April 2020


STANDARDS

Regardless of whether that model was trained with user, together with a proof that it is indeed the result of
their data or not, the privacy of all London tube users is the intended computation.
breached to some extent. This point might sound obvi- It is important to remark that this approach relies
ous, but it is important: technical advances in general on trusted hardware, which is in general hard to patch
do not solve all ethical issues, and privacy is not an if vulnerabilities are found. Moreover, there are some
exception to that. Every data analysis has some ethi- limitations to its security guarantees, as it does not
cal issues regarding privacy associated with it, which protect against cache-timing and physical attacks,
must be approached as such. as well as limitations in terms of scalability because
However, there are many crucial privacy issues in the amount of available RAM within a container is
data analysis that technology can help to over-come. often limited.
Two aspects for which we have in principle satisfac- Examples of this technology are Intel’s SGX and
tory technical solutions are privacy of stored data, ARM TrustZone, which are evolving and being adopted
i.e., encryption of data at rest (on disk), and privacy of quickly. A recent instance of such adoption are the
data as it is being transmitted, i.e., encryption of data Azure Confidential Computing capabilities.
in transit. Current basic research challenges have
to do with preserving privacy even during process- Homomorphic encryption.
ing, and generally correspond to two orthogonal but An encryption scheme is said to be homomorphic
tightly-related aspects: privacy-preserving computa- with respect to a given operation if one can perform
tion and privacy-preserving disclosure. that operation on the encrypted data by just manip-
ulating the corresponding ciphertext. For example, if
Privacy-preserving computing: an encryption scheme is homomorphic with respect
The result and nothing but the result! to addition, two encryptions of arbitrary values, say 23
Let’s say you upload your data encrypted to the cloud, and 19, can be combined—without prior decryption—
but still allow for some concrete computations to be to produce the encryption of their sum. Asymmetric
performed on it by service providers, such as train- key encryption schemes that are homomorphic with
ing machine learning models, or selecting ads tailored respect to either addition, e.g., Paillier, or multiplica-
for you. This would certainly keep the service provid- tion, e.g. ElGamal, have been known for a while, but it
ers happy, while protecting your private data from data wasn’t until 2009 that Gentry described the first fully
breaches. homomorphic encryption scheme,4 namely a scheme
So now how do we execute software on machines that’s homomorphic with respect to both addition and
owned and maintained by an untrusted party? Or, more multiplication. Note that, if we operate on a binary
generally, how do we compute on private data held by domain, i.e., mod 2, addition and multiplication is all
mutually untrusted parties? There are several emerg- that one needs to do anything a modern processor
ing techniques to do this that could be combined in can do. This enables secure outsourced computation
principle, and come from the areas of hardware secu- relying solely on encryption, as opposed to the secure
rity and cryptography. enclave approach, as a user can encrypt all their data
and share it with the cloud encrypted, together with
Secure enclaves. the public key of the encryption and a description of
The idea behind secure enclaves is based on new tech- the computation. Then the cloud provider can com-
nology (not so new on the iPhone but new to servers) pute on it in encrypted form—as if it was computing
called a Trusted Execution Environment.3 Such trusted blindfolded—and return the encrypted result.
hardware provides a secure container into which Fully homomorphic encryption is a remarkable
the secure cloud user can upload encrypted private breakthrough, as before Gentry’s contribution, it was
data, securely decrypt it, and compute on it. Both the not even clear whether such kind of encryption could
decryption and the computation are run in a proces- even exist. However, although several alternative
sor, which, in principle, not even its owner can break improved schemes have been proposed since Gen-
into. The result is again securely transmitted to the try’s, homomorphic encryption is currently far from

www.computer.org/computingedge 35
STANDARDS

scaling to the secure cloud computing application, hence incurring some computation and commu-
and in particular, data analysis tasks involving mas- nication cost. This is in contrast with secure cloud
sive input sizes. Nevertheless, several homomorphic approaches, where the encrypted data only has to
encryption libraries are available, and a limited notion be uploaded once. This motivates architectures that
of fully homomorphic encryption supporting a fixed include sets of noncolluding untrusted parties that
number of nested multiplications called somewhat are used to simulate a secure cloud. It is important to
homomorphic encryption might be enough for some remark that MPC techniques provide high-assurance
data science applications. cryptographic guarantees.

Multiparty Computation (MPC). Edge computing.


Another alternative is to use secure Multiparty Com- There are performance advantages to edge comput-
putation (MPC), an area of cryptography kicked off by ing in some new scenarios, especially in the smart
Andrew Yao in the 80s.5 There are a number of proto- home and Internet-of-Things use cases. At the least,
cols, including the lovely Yao’s garbled circuits, that we can remove the burden of sending large amounts of
revolve around the idea of sharing secrets without detailed data from very large numbers of edge devices
actually giving them away, and then computing on into the cloud. Instead, we retain the data in local hubs
them by transforming their shares, and still keeping (e.g., smart home hubs), and send analytics software
them secret. An example is the way to find out who is to execute there, rather than on the central cloud. This
the richest person in the room, without revealing how is recognized in the IoT hub work by Microsoft and in
much each person actually possesses. These are hard several other IoT platforms. We still need to retain all
to reason about for the layperson, but can be verified the same approaches to supporting privacy concern-
in design, and probably therefore are a promising addi- ing the data, but the very distributedness of the data
tional technique. Moreover, MPC techniques are quite and computation reduces the risk of a mass-leak of
efficient and, due to a sequence of theoretical and information because the attack-surface of the whole
engineering breakthroughs, have become of practi- system is now fragmented. Techniques for decentral-
cal interest, with many available libraries and applica- ized analytics work quite well and can adopt many of
tions, and even commercial products. the techniques used for large-scale analytics in data
MPC technologies allow for moving away from centers. Moreover, one can in principle enhance such
the trusted aggregator model for analysis of distrib- approaches with the cryptographic techniques men-
uted data. Instead of moving all the data to a single tioned above to yield high-assurance guarantees.
server (where it might be leaked), we can leave data
in peoples’ devices (smart homes, smart TVs, cars, IoT Tailored approaches.
devices, tablets, etc,) and distribute the programs that For a concrete problem, for example, logistic regres-
do the analytics in a privacy-preserving way. This then sion on distributed data, custom “hybrid” protocols
moves the results (e.g., market segment statistics) that combine several of the techniques above are
to businesses that wish to exploit them without ever likely to give the best results by sacrificing generality.
moving the raw personal data anywhere at all. Hence, Research prototypes that follow this hybrid approach
the parties interested in an aggregated model learn have been proposed for private training and classifi-
such a model, and nothing but the model. In principle, cation in models such as neural networks, ridge and
this permits reversing the business models’ direction logistic regression, nearest neighbors, text classifica-
of value—the subject (user) can now charge for their tion, and random forests, among others.
data! In the distributed approach, since there is no
central data center/cloud anymore, there’s no need to Privacy-preserving disclosure:
cover its cost, so the change adds up. How much does the result
The caveat here is that while MPC techniques actually disclose?
allows keeping the data with the parties owning it, Although the techniques above can be used to com-
such parties must get involved in the computation, pute a statistical model in a privacy-preserving way,

36 ComputingEdge April 2020


STANDARDS

namely not disclosing any unnecessary information, You will not be affected, adversely or otherwise,
they do not address the problem of quantifying how by allowing your data to be used in any study
much is disclosed by such a model. This (vague) ques- or analysis, no matter what other studies, data
tion regarding “how much is disclosed” has many sets, or information sources are available.”
aspects. For example, one might be interested in quan-
tifying to what extent a sensitive feature of the training How it is done is a detail that varies across appli-
dataset is disclosed by the model. Alternatively, one cations, and can involve several approaches to filter-
could try to address whether the model would allow ing data collected, or fuzzing features of the data, or
deanonymizing a public, in principle, unrelated data- analyzing and blocking overly intrusive or too-frequent
set, or whether a given individual can be identified as questions. Given that most market research style ana-
part of the training dataset. Each of these goals cap- lytics is concerned with identifying groups (segments
tures something about our intuition regarding privacy or bins) in the data, this may not lose any value at all.
and they may be more or less suitable in different con- Users with really obscure or rare features don’t repre-
texts. As with issues such as bias and fairness in sta- sent significant market opportunities.
tistical models, mathematical definitions of privacy There are several aspects that make DP an appeal-
are important even if they only capture part of what we ing definition. First of all, DP neutralizes linkage
intuitively mean by privacy preservation. attacks, as it is defined as a property of the analysis,
One thing is clear among information security not the data on which the analysis is run. For the same
experts: simply removing the primary keys (names, reason, DP is also immune to post-processing: running
birthday, postcode, etc.) of a database and replac- subsequent analysis on the result of a differentially
ing with some pseudo-random numbers, so called private analysis cannot result in a less private result.
“de-identification,” won’t work in general. There are DP also has nice composition properties, that allow
too many diverse holders of records to prevent trivial the building of provably private analysis from simpler
re-identification (sometimes called triangulation) by building blocks. Finally, one of the main characteristics
linking data from different sources and inferring who of DP is that it allows for privacy quantification, as it
the subject is. Another defense, often referred to as defines privacy in terms of real-valued parameters ε
k-anonymization, consists on “fuzzing” the dataset so and δ. Setting these parameters properly is in general
that any allowed query includes data from at least k indi- an open problem that corresponds to the tension
viduals, hence providing some uncertainty that should between privacy and utility in data analysis: smaller
protect privacy. However this also does not account values of the parameters provide better privacy, but
for the above mentioned linkage attacks. In summary, might render the analysis useless.
what makes privacy difficult is dimensionality: a some- To illustrate the ideas behind DP to provide pri-
times surprisingly small number of features is enough vacy by a randomized mechanism, it is useful to con-
to make a database record essentially unique. Hence, sider the very related idea of randomized response.
an attacker with a bit of background knowledge about Randomized response is a technique developed in
a given individual can use it to obtain the additional the 60s to collect statistics about illegal or embar-
information about that person present in a database. rassing behavior. Every participant of the study is
A particularly successful mathematical definition instructed to (a) (privately) flip a coin before replying
of privacy, as it is receiving lots of attention from both to the question, if the coin comes up heads then (b)
academia and industry, is Differential Privacy (DP). answer truthfully, and if the toss comes up tails (c)
Intuitively, DP allows us to design analyses in a way randomly answer “yes” or “no,” using a second coin
that quantifies how much they give away in terms of toss. The surveyor can then correct the result using
whether a record was part of the database or not. As their knowledge about the surveying mechanism to
Dwork and Roth put it,6 get an approximate count. Note that privacy for the
participants here comes in the form of plausible deni-
“Differential privacy describes a promise, made ability: they can always claim “I did not vote for Brexit!,
by a data holder, or curator, to a data subject: the coin tosses made me report yes.”

www.computer.org/computingedge 37
STANDARDS

DP has been shown to be a very rich concept with tackled from several areas such as hardware/sys-
interesting connections to several aspects of infor- tems security, cryptography, statistics, and machine
mation theory and learning. There have also been learning. Several privacy-enhancing techniques have
some applications of it by big data controllers such evolved significantly in the last decade from being
as Google and Apple, but a clear path to standardiza- mainly theoretical to becoming academic prototypes
tion does not yet exist. The main challenges have to and even commercial products and, as recognized by
do with modelling and parameterization choices, to both governments and industry, have the potential to
which DP is very sensitive not only in terms of privacy revolutionize the field. These techniques have differ-
but also utility. ent tradeoffs, maturity levels, and privacy guarantees,
and in some cases solve slightly different problems.
TOWARD STANDARDS FOR LARGE A fully fledged approach to privacy-preserving data
SCALE DATA ANALYTICS analysis would still require significant interdisciplinary
The technologies mentioned above and more impor- effort, some of which have to do with issues such as
tantly, their interplays, are not mature enough to be effective personal data management and consent,
completely standardized. Moreover, the complexity which we did not address in this paper.
of secure data analysis will require several kinds of The need for robust privacy-preserving data
standards, related not only to the different aspects analysis technologies has been recognized by both
of privacy-preserving analytics discussed above, but regulators and industry. This would not only mitigate
also related issues like personal data management the growing risks of privacy failures, but also enable
and consent. First, just like there are standard virtual- opportunities based on computing on private data.
ization APIs, we need standard APIs for trusted execu- This is analogous to how encryption revolutionized
tion. Moreover, one has to address choices regarding secure communications, enabling a huge economic
cryptographic protocols, the architectures on which development, mainly through secure payments. While
they are deployed (possibly involving semi-trusted regulation and standardization would apparently
parties), and with which security guarantees in terms accelerate this process, the technology is not quite
of key sizes and similar parameters. Even if we agree on there, and more research is needed before the field
which protocols to use and how to instantiate them, as a whole is mature enough to yield precisely defined
there are always a set of services that are required to good practices and regulation, capable of, for example,
deploy such protocols in practice, and a reasonable enabling audits to ensure compliance.
incentive system for parties to provide such services
must be in place. This includes tasks such as key dis- REFERENCES
tribution, attestation, and verification, which might 1. R. Serra, Television Delivers People, 1973; https://www
potentially involve actors focused on these tasks. A .moma.org/collection/works/118185.
major challenge to overcome is that of “how much pri- 2. Report of the Commission on Evidence-based Policy-
vacy is enough?” Privacy, unlike secrecy or security, is making, September 2017; https://www.cep.gov
in some cases not a binary predicate, as it undermines /content/dam/cep/report/cep-final-report.pdf.
utility in many applications. For example, establish- 3. V. Costan and S. Devadas, “Intel SGX explained,” IACR
ing a “safe” differential privacy modelling and parame- Cryptology ePrint Archive, 2016; https://eprint.iacr
ters for recurrent analysis on sensitive census data is a .org/2016/086.pdf.
major challenge. What one means by “safe” would have 4. C. Gentry, “Fully homomorphic encryption using ideal
to be not only rigorously established, but also effec- lattices,” Proceedings of the forty-first annual ACM
tively communicated by, for example, something like symposium on Theory of computing (STOC 09), 2009,
kitemarks for safety, but instead for privacy level. pp. 169–178.
5. A.C. Yao, “How to generate and exchange secrets,”

P rivacy-preserving data analysis is an emerg-


ing discipline within data science, which posts
several challenges currently being simultaneously
Proceedings of the 27th Annual Symposium on Founda-
tions of Computer Science, 1986, pp. 162–167.
6. C. Dwork and A. Roth, “The algorithmic foundations of

38 ComputingEdge April 2020


STANDARDS

differential privacy,” Foundations and Trends in Theoreti- ADRIÀ GASCÓN a Turing Senior Research Fellow at Warwick
cal Computer Science, vol. 9, no. 3-4, 2014, pp. 211–407. University and a Research Fellow at The Alan Turing Institute.
He received a PhD in Computer Science from the Polytechnic
JON CROWCROFT is a professor at the University of Cam- University of Catalonia (UPC). His research interests include
bridge and a Fellow at The Alan Turing Institute. He received program synthesis, distributed algorithms, applied cryp-
a PhD in Computer Science from UCL. His research interests tography, and machine learning. Contact him at agascon@
include Internet Protocols, Cloud Computing, Social Media turing.ac.uk.
Analytics and Privacy Technologies . Contact him at jon
.crowcroft@cl.cam.ac.uk. Contact editor Yong Cui at cuiyong@tsinghua.edu.cn.

ADVERTISER INFORMATION

Advertising Coordinator Central US, Northwest US, Southeast US, Asia/Pacific:


Eric Kincaid
Debbie Sims Email: e.kincaid@computer.org
Email: dsims@computer.org Phone: +1 214-553-8513 | Fax: +1 888-886-8599
Phone: +1 714-816-2138 | Fax: +1 714-821-4010 Cell: +1 214-673-3742

Midwest US:
Advertising Sales Contacts Dave Jones
Email: djones@computer.org
Mid-Atlantic US: Phone: +1 708-442-5633 Fax: +1 888-886-8599
Dawn Scoda Cell: +1 708-624-9901
Email: dscoda@computer.org
Phone: +1 732-772-0160
Cell: +1 732-685-6068 | Fax: +1 732-772-0164 Jobs Board (West Coast and Asia), Classified Line Ads

Southwest US, California: Heather Bounadies


Mike Hughes Email: hbuonadies@computer.org
Email: mikehughes@computer.org Phone: +1 623-233-6575
Cell: +1 805-208-5882

Northeast, Europe, the Middle East and Africa: Jobs Board (East Coast and Europe), SE Radio Podcast
David Schissler
Email: d.schissler@computer.org Marie Thompson
Phone: +1 508-394-4026 Email: marie.thompson@computer.org
Phone: +1 714-813-5094

www.computer.org/computingedge 39
COLUMN: LAST WORD
This article originally
appeared in

vol. 17, no. 5, 2019


It Takes a Village to Secure
Cellular Networks
Elisa Bertino, Purdue University

T
he 4GLTE technology has tremendously involved in the protocol into finite-state machines,
increased the bandwidth available for smart- which must be at a suitable abstraction level. The
phones, essentially delivering broadband second is that the verification of the protocols is
capacity to them. The most recent 5G technology often performed with respect to properties of inter-
is poised to further enhance transmission capac- est, which requires considerable domain knowledge.
ity and latency through the use of the unused radio Extracting finite-state machines from standardiza-
spectrum bands (e.g., millimeter waves) and devices, tion documents is challenging. These documents are
such as massive multiple-input, multiple-output written in natural language and, thus, have ambigui-
antennas. 5G will enable a host of new applications ties. In addition, they are lengthy, and their structure
and permit the massive deployment of the Internet makes reading them a cumbersome task.
of Things systems. Of course, we cannot expect writers of standard-
Cellular networks are and will increasingly be one ization documents to produce finite-state machines
of the most critical infrastructures, and their security out of their documents. Perhaps a community-based
is obviously crucial. However, securing cellular net- effort involving researchers from academia that aims
works is a challenging task. These networks consist of to formalize those protocols and make these formal-
multiple layers: physical, radio resource control, non- izations publicly available would be the way to go. The
access stratum, and so on. Each layer, in turn, has its availability of such formal specifications and relevant
own protocols, such as those for attaching/detaching properties would enable researchers and developers
devices to/from the network and paging devices for to systematically analyze cellular network protocols.
notifications of incoming calls and SMS. It would allow the community to test the limits of cur-
Research in recent years has succeeded in iden- rently available tools for formal methods and promote
tifying a few security and privacy vulnerabilities. new research directions for enhanced approaches
However, we are far from having systematic and and, perhaps, of tools based on tools other than formal
comprehensive approaches to identifying such vul- methods. These specifications would also be relevant
nerabilities. In some cases, past work has relied on for industries implementing those protocols in that,
network traces, in other cases, on cryptographic veri- on the one hand, they would provide a more precise
fiers, and in still other cases, on researchers’ intuition. description, and, on the other hand, they would allow
As a result, past analyses are limited. the comparison of protocol implementations against
More recently, systematic methods have emerged, the standard.
such as those combining cryptographic verifiers and Let’s now talk about vulnerabilities. When we look
model checkers. However, the use of such methods is at those so far identified, it is striking that most of
far from trivial. The first challenge is that, to use those them are due to the lack basic and well-known secu-
methods, we must encode the behaviors of the parties rity practices. Notable examples include lack of integ-
rity verification or lack of replay protection for certain
messages and lack of authentication for certain
DOI No. 10.1109/MSEC.2019.2922867 broadcast messages. An important example of the
Date of publication: 3 September 2019 latter is the messages periodically broadcast by base

40 April 2020 Published by the IEEE Computer Society  2469-7087/20 © 2020 IEEE
LAST WORD

stations to advertise their presence in corresponding complex and have even more stringent real-time con-
geographical cells and provide parameters so that straints. Testing those solutions and devising further
devices can connect through them. These messages requirements will require the engagement of network
are broadcast with high frequency, irrespective of providers and device manufacturers. It will take a vil-
any device’s presence in the cell area: every 40 ms lage to secure cellular networks!
for the master_info_block message and every 80 ms
for the system_info_block message. Because those ELISA BERTINO is a professor with Purdue
messages are not digitally signed, the devices have University. Contact her at bertino@purdue
no assurance that they are connecting to a legitimate .edu.
base station. As a result, it is possible for malicious
parties to spoof legitimate base stations.
At first glance, we may think that the deployment
of well-known approaches, such as a public-key
infrastructure-based authentication mechanism,
would easily fix this problem. However, a closer analy-
sis of how such an approach would work to authenti-
cate those two types of broadcast messages shows
that there are many different requirements for such a
method. For example, sending certificates along with
those messages would impose high communication
overhead; bandwidth is a precious resource, and net-
work providers may be reluctant to spend bandwidth
for sending certificates.
Therefore, certificate size should be minimized
as much as possible, with careful selection of which
IEEE MultiMedia

messages to sign. The overhead for signature gen-


July–September 2016

http://www.computer.org
july–september 2016

eration and verification is also critical and needs to


be minimized. Base stations need to reduce the time
required for generating the signatures because of
the high frequencies of those broadcast messages;
❚ Quality Modeling

approaches, such as signature aggregation, should


be devised to be deployed at base stations and
highly optimized.
Mobile devices, on the other hand, need to mini-
Volume 23 Number 3

mize the signature verification times, as saving energy


is critical. This would require mobile-device manufac- mult-22-03-c1 Cover-1 July 12, 2016 4:40 PM

turers to come up with energy-efficient implementa- IEEE MultiMedia serves the community of
tion of signature verifications. Additional constraints scholars, developers, practitioners, and students
include revocation of certificates, backward compati- who are interested in multiple media types and
bility, and certificate management for devices roaming
work in fields such as image and video processing,
audio analysis, text retrieval, and data fusion.
across different providers (for example, when travel-
ing abroad). The research community has developed
many interesting and novel solutions to the problem
Read It Today!
of digital signatures. However, we need to test these www.computer.org/multimedia
solutions and the solutions to other security problems
in the complex scenario of cellular networks that,
especially in the case of 5G networks, are increasingly

www.computer.org/computingedge 41
COLUMN: STANDARDS This article originally
appeared in

Improving Performance and vol. 23, no. 1, 2019

Scalability of Next Generation


Cellular Networks
Ali Mohammadkhan and K. K. Ramakrishnan, University of California, Riverside

Uma Chunduri and Kiran Makhijani, Future Networks, Huawei Technologies

The 5G cellular network’s packet core architecture has adopted concepts of software-
based networking to improve scale and flexibility. In this paper, we investigate
potential improvements to the current architecture, the protocols for the 5G
control plane and backhaul network to achieve signaling efficiencies, improve user
experience, performance, scalability, and support low-latency communications.

5
G networks promise to revolutionize cellular that conserves energy by quickly transitioning to an
communications, with a substantial increase idle state, turning off the radio. A service request event
in per-user bandwidth and low latency through (when the UE transitions from idle to active to exchange
improvements in the wireless radio technology. 5G packets), requires between 13 and 32 messages (Figure
networks are being proposed as an alternative not only 3).1 This long sequence of messages introduces unde-
for traditional smart-phone based data and telephony sirable latency in initiating a data transfer after the idle
applications but also for Internet-of-Things (IoT) and period. The overhead (in messages exchanged) and
even for residential Internet service. While the use of latency may nullify the purpose and goal of transition-
improved radio technology will help tremendously, ing to an idle state.
challenges remain because of the complexity of the We suggest a careful re-examination of the 5G
cellular network protocols. Of particular concern is the architecture and control plane protocol to improve per-
complexity of the control plane protocol and the use formance. There are three aspects we explore (a) rede-
of the GPRS tunneling protocol (GTP). Tunnels carry signing the control plane signaling protocol and 5GC
traffic between the end user equipment (UE) and the system architecture, (b) an optimized traffic engineer-
cellular packet core network. With the increased use ing (TE) path selection in the backhaul network, and (c)
of small cells (potentially more frequent handovers) an enhanced programmable data plane. We begin with
and the need to support a large number of IoT devices an overview of 5GC architecture, its control plane pro-
(which switch between idle and active more frequently tocol and approaches to simplify them, thus reducing
to save battery power), the need for efficiency of the latency, improving efficiency, throughput, and scalabil-
control plane is even more important. ity. Second, we propose a simplification of the backhaul
The 5G Core (5GC) consists of several different network, which is usually treated as an opaque entity.
components that carry out individual tasks. When an We explore alternatives currently being considered in
event for a user (e.g., attach, handover, service request) the Internet Engineering Task Force (IETF). Finally, a pro-
occurs, a large number of messages are exchanged grammable data plane is discussed to enable additional
between these components for notification and syn- network level functions necessary for 5G applications
chronizing state. Consider, for example, an IoT device that require high reliability or low latency.

BACKGROUND
DOI No. 0.1109/MIC.2018.2884882 Conceptually, the 5G architecture follows the prin-
Date of current version 6 March 2019. ciples of the Control and User Plane Separation of

42 April 2020 Published by the IEEE Computer Society  2469-7087/20 © 2020 IEEE
STANDARDS

FIGURE 1. 5G system architecture.3

cellular core architecture introduced in Release 14 of involving compute nodes, the data plane needs to
the 3 GPP specification thus simplifying the function- perform high-speed simple forwarding and is built
ality needed to be supported by each component.2 with hardware accelerated forwarding engines. How-
However, the separation between the control and user ever, with the advent of virtualization, common
plane components significantly increases the num- off-the-shelf server (COTS) systems with a large num-
ber of messages needed to coordinate a user session ber of processor cores, software libraries such as the
state across these components. There are five main “Data Plane Development Kit” and high-performance
entities, apart from the UE and cellular base station network interface cards, this separation of func-
(called the gNB in 5G new radio) as shown in Figure 1. tionality is no longer necessary.4 For example, a sin-
These are access and mobility management function gle server running the OpenNetVM platform can pro-
(AMF), Service Management Function (SMF), authen- cess and forward 10 s of millions of packets per second
tication server function (AUSF), unified data manage- with software-based network functions (NFs) han-
ment (UDM) and user plane function (UPF). The UPF dling both complex control plane functions and high
is a data plane entity, while the others are control and rate data plane workloads.5 This has led the cellular
management plane entities. The AMF is main control industry to evolve into a software-based packet core
plane orchestrator, managing UE mobility, session (5GC) system architecture. However, the 5G architec-
establishment (through SMF), and handling service ture continues to emphasize the separation between
requests. Additionally, entities such as the NSSF, PCF, the control plane and data plane as one of the goals
and AF also play important roles, the details of which even as software-based systems are able to elegantly
can be found in 3 GPP TS 23.501.3 support multiple classes of functions running on the
same system.6 The main requirement of efficiency
POTENTIAL ARCHITECTURAL and high performance can, in fact, be achieved by hav-
ENHANCEMENTS ing the 5G control and data plane functions coresi-
Traditionally, hardware components are purpose-built dent on the same COTS system. Coresident NFs can
and customized for distinct functions. While the con- share state information more easily, and where pos-
trol plane requires the capability to handle complex sible take advantage of shared packet processing.
processing and has more sophisticated capabilities Use of software-based NFs should be viewed as an

www.computer.org/computingedge 43
STANDARDS

FIGURE 2. Example architecture using software-based NFV for implementing core control and data functions.

opportunity to carefully take stock of how the func- the NFV-based 5GC platform allows inherent scale-out
tional architecture of the 5GC should be implemented. of functions on-demand. CleanG can also conveniently
By doing so, we see the potential for reidentifying the support slicing of the 5G network into multiple logical
functional components, but not require separation into slices (whether it is for having different logical planes
different physical entities. Instead, these components for distinct services or for different virtual network
can be implemented as submodules or NFs in a service operators) by having distinct instances of the core NF
chain (or multiple service chains) on a single system. for each slice.
We propose an architecture called CleanG,7 where A direction currently being pursued in the industry
the data plane and control plane are supported by dis- is to have the 5G control and data planes separated
tinct NFs collocated on the same physical system as by having an SDN controller as an intermediary.8
depicted in Figure 2 (i.e., a single component with two Unlike IP networks where the timescales for control
submodules). For example, supporting them on Open- plane updates (infrequent, of the order of seconds or
NetVM platform, each submodule can be assigned more) are very different from data plane operations
resources (CPU and buffering) dynamically as needed (frequent, of the order of microseconds or less), the
based on the control and data plane workloads. CleanG cellular control plane and data plane are much more
remains true to 5G system architecture (release 16), tightly coupled (e.g., when a UE transitions from idle
both based on NFV. In CleanG 5GC functions remain to active, data packets can only flow after control
logically and functionally decoupled. However, by vir- plane operations for processing the service request
tue of being coresident with other 5GC user and con- are completed).9 Having a controller to mediate the
trol functions, several messages between functions updates between the control and user plane adds sub-
are unnecessary and communications overheads are stantial delay. Additionally, the controller may become
reduced. There is no need to distribute and synchro- a bottleneck under heavy control traffic. Because of
nize state information. These are major contributors the need to minimize the delays between control and
to performance improvements in CleanG. Additionally, user plane operations in the cellular environment, we

44 ComputingEdge April 2020


STANDARDS

TABLE 1. Approximate number of control plane messages received (R) and sent (S) for different events in 5G. (B = baseline, O =
optional messages).

believe it is highly desirable to have them coresident IMPROVING CELLULAR CONTROL


on the same node wherever possible. PLANE PROTOCOL
Supporting software-based NF offers additional As the number of components in 5G increased com-
opportunities for simplifying complex functions such pared to LTE, additional messages are needed to keep
as roaming. With the ability to copy over the state of state synchronized among them. Table 1 shows the
a user session, an NF can be initiated in the visiting number of messages exchanged for the 5G for each
network in a short time (less than a second). This user event.
enables the user to be served by an NF in the visiting Most of these messages are exchanged sequen-
network while maintaining information for the home tially, or on occasion, after a timer expiration. The
network, avoiding extra packet exchanges with the completion time for control plane actions for an
home network. This approach is more efficient than event is the cumulative time for exchanging these
normal roaming or local breakout approaches as both messages, thus contributing to high delays. This
data and control plane components are closer to the delay includes time to process each control message
roaming user. and the propagation and queuing delays for sending
Our proposed CleanG architecture for the 5G cel- packets between different components, especially if
lular core exploits logically separate but physically the control plane and data plane components are far
consolidated core control (CCF) and core data plane apart. Based on the CleanG NFV-based architecture,
functions (CDF), running on the OpenNetVM platform multiple core network processing components may
is shown below.7 The CCF supports functionality pro- be consolidated into one or more NFs running on the
vided by the SMF and AMF, while CDF implements the same node. This facilitates reducing the number of
functionality of the 5G network's UPF. A primary goal control plane messages, lowering completion time.
is minimizing delay for an update from the cellular A second major 5G overhead is in using GTP-U
control plane resulting in changes to the data plane. tunneling to carry data packets between different
This is achieved by the CDF and CCF sharing data and user plane components. The latency consuming task
state using the OpenNetVM shared memory. Finally, is the setting up of the tunnel. The receiving end
multiple instances of this NFV-based 5GC may be cre- assigns a tunnel ID (TEID) to a flow and notifies the
ated to scale-out based on traffic, and to dynamically sender. Because the control plane components (AMF
adapt to the ratio of control to data plane traffic. and SMF) are involved in initiating and mediating the

www.computer.org/computingedge 45
STANDARDS

tunnel setup, a number of messages are exchanged, Background on Current


which is time-consuming. In the CleanG architecture, TE Mechanisms
we use simple Generic Routing Encapsulation tunnel- In the current cellular architecture, traditional
ing that does not require explicit setup or exchang- approaches like MPLS with RSVP-TE determine the
ing TEIDs. Different classes of services can be used label switched path and manage resources along
to meet simple application requirements using the the path.10,11 However, they are not dynamic and
differentiated service code point (DSCP) field in the require provisioning steps along the path, involv-
outer IP header. ing out-of-band signaling. Another mechanism being
Another challenge for future cellular networks considered is to use segment routing (SR) which is a
is from new types of workload, e.g., from IoT devices. source routing technology. SR (RFC 8402), uses path
Exchanging a large number of control messages for an segments computed offline by a controller for a par-
idle-active transition not only adds delay but results in ticular flow or a service.12 The path is then decom-
the overhead on 5GC control components (e.g., AMF, posed into a sequence of network segments along
SMF). One option proposed in 3GPP standardization which packets of a flow are routed. A sequence of
is to piggyback data packets with the first control segments is carried in the packet, essentially using
message to an AMF, to reduce delay. However, this can source routing, with either MPLS labels or IPv6
cause the AMF to become the bottleneck (excessive address formats.13,14 While SR allows packet steering
load from large numbers of IoT devices), contributing on a specified path, it does not have any notion of QoS
to the additional delay. The consolidation of control or resources being reserved along the path. Further-
and data plane components in CleanG enables imme- more, SR also has the well-known overhead of increas-
diate notification of the control plane while avoiding it ing the packet header by encoding the segments into
having to process data packets. packets for routing purposes.15
Consider, for example, a service request user A recent alternative being proposed, called Pre-
event in the current 3GPP specification (see Figure ferred Path Routing (PPR),15,16 seeks to overcome some
3) for 5G networks. The AMF updates the SMF about of the challenges in above-mentioned approaches.
user sessions and receives responses. The SMF then PPR is an innovative path routing architecture to
updates the UPF, enabling forwarding of packets by signal explicit paths from sources and per-hop pro-
the data plane. In an NFV environment where the con- cessing including QoS awareness from the controller
trol and data plane components are coresident on the to network nodes. PPR builds on the SDN paradigm
same system and can share state, the need for 6a,b, and utilizes various central path computation engine
7a,b, 18a,b, 21a,b can be eliminated. Consolidation of features to create the TE paths/graphs) with a cer-
control entities in the architecture can eliminate mes- tain bandwidth,17,18 latency and/or allowable jitter
sages 4, 11, 15, and 19. For the common cases when constraints. When deployed in 5G backhaul, PPR can
an intermediate UPF is not used and dynamic policy combine TE and QoS guarantees of the path without
is not enforced, 13 out of 15 core message exchanges significant packet overheads.
are not essential. Figure 4 shows a typical backhaul network with
cell site routers (CSRs) at gNB and virtualized UPF. At
CARRYING TRAFFIC EFFICIENTLY CSR, PPR encapsulates user packet with destination
OVER BACKHAUL NETWORKS IP address as PPR-ID, which is a preprogrammed for-
With the higher speeds of new radio technologies, the warding identifier along with QoS attributes through
capacity demands and Traffic Engineering require- the underlying routing protocol/IGP. The packet is
ments on backhaul networks between gNB and 5GC forwarded along the preferred path associated with
increase rapidly. Efficient use of the network, its PPR-ID from gNB to UPF. PPR can support new ser-
resources and proper use of available paths are essen- vices that require low and/or bounded latencies, along
tial for both cost and performance reasons. We exam- with traditional bandwidth guarantees.
ine current technologies trends for path selection, PPR flexibly supports a wide range of existing
routing, and TE in the backhaul network. forwarding planes including native IP (IPv4/IPv6) user

46 ComputingEdge April 2020


STANDARDS

FIGURE 3. Messages exchanged for service request event in 5G among 5GC components and UE & (R)AN.

www.computer.org/computingedge 47
STANDARDS

FIGURE 4. Backhaul transport with PPR.

planes. PPR also brings much-needed optimizations to transit through multihop networks. Specifically, when
SR defined user planes (SR-MPLS, SRv6). TE is complex faced with congestion, packets of a service offering
and 5G applications need these paths to be available ultra-reliability may be arbitrarily dropped because
or withdrawn on-demand. The PPR solution simplifies the data plane has little knowledge about the need
this in the 5G backhaul by enabling different types of to provide ultra-reliable service to a flow. To keep
TE paths in a scalable manner and allows for rapid traf- data planes more aware of service constraints, SDN
fic switchover to backup paths. frameworks and programmable switches may be used
to provide dynamic treatment of flows within a switch
IMPROVING THE BACKHAUL or router via the control plane; nevertheless, such
DATA PLANE out of band programmability alone is insufficient to
GTP-U is the legacy 4G user plane protocol that is also respond to changing network conditions in a timely,
used in 5G. The new N9 interface between UPFs and service-aware manner. A contextual data plane is
its encapsulation is currently under study in both 3 desired that can discriminate packets based on their
GPP and IETF in order to determine whether retaining service guarantees at runtime. In this regard, the “big
GTP-U or using other IP approaches over the backhaul packet protocol” (BPP) data plane solution providing a
network would address concerns about supporting higher degree of customization of flows across a net-
a diverse set of services, such as having low-latency work has been suggested.
applications as well as supporting high bandwidth The BPP data plane is a generic framework that
flows in 5G networks. enables carrying different service-specific param-
In this section, we discuss challenges in support- eters and constraints along with the packets and
ing such a wide range of applications with traditional processing them on each hop. With this additional
IP and recommend further refinements to the 5G information, routers know precisely what to do for
user plane. each packet, going beyond just the coarse-grained
Supporting deterministic service guarantees for indication from the DSCP bits of an IP packet. A
5G applications, such as a vehicle to infrastructure typical BPP packet on the wire is shown in Figure 5(a).
communication, and industrial automation and The BPP block provides “commands,” for example,
control is a nontrivial task in modern data planes. basic forwarding actions such as next hop, redirect,
IP offers coarse-grained control over how packets inspect, drop, etc., as well as new scheduling or shap-
of such flows are scheduled and processed as they ing actions such as bounded end-to-end latency or

48 ComputingEdge April 2020


STANDARDS

FIGURE 5. (a) BPP Packet format. (b) BPP aware packet network to enable low-latency networking.

acceptable jitter. Additionally, the meta-data block In Figure 5(b), two latency sensitive paths are
may carry information about the packet such as what shown, a) from NR to edge compute node shown as
service it is part of, and/or mappings to the path that green dotted path, which needs low latency guaran-
packet may use. Although the protocol itself does not tees of the order of 10 ms, b) from edge compute to
preclude inserting the BPP Block by a host where the data center in core network with bounded latency
packet originates, it is best if inserted at the edge of shown as yellow dotted path that should support
the service network thereby remaining under net- bounded latency in the order of 30 ms. For simplicity,
work operator's control. We discuss below how BPP assume these are BPP enabled IP networks, capable
may be useful for providing very low latency, and of processing data packets per BPP instructions and
highly reliable delivery (the so-called “ultra-reliable, appropriate scheduling.
low-latency” class of traffic, URLLC) over the cellular A BPP packet carries “residual latency” instruc-
backhaul next. tion as it traverses the BPP aware backhaul network,
the value of which gets adjusted at each hop. When
Supporting Low-Latency Traffic a packet arrives on a particular hop, the forwarding
Over the Cellular Backhaul pipeline on the node reads the instruction, sched-
The example used to demonstrate BPP capabilities ules the packet based on the residual latency for the
is that of the vehicle to infrastructure (V2X) commu- packet but also updates the value of latency for next
nication in a 5G network. In V2X applications, receiv- hop. If the latency bound (specified in latency instruc-
ers do not wish to receive traffic that is already delayed tion metadata) cannot be met because of the current
beyond a latency bound. For example, a vehicle that queueing at the router, the packet will be dropped
already crossed intersection does not benefit from early as there is no value in late arrival, thus avoid-
receiving traffic information for that intersection. ing wasted network bandwidth further downstream.

www.computer.org/computingedge 49
STANDARDS

Similar instructions can be incorporated for reliabil- EPC nodes (CUPS),” 2017. Available at: http://www
ity (backup path) but are not discussed here due lack .3gpp.org/cups
of space. 3. 3 GPP TS 23.501, “System architecture for the 5G
Admittedly, such data planes would require new system,” 2018.
capabilities in the forwarding engines, which are well 4. [Online]. Available at: https://www.dpdk.org
served with programmable software-based NFV plat- 5. W. Zhang et al., “OpenNetVM: A platform for high per-
forms like OpenNetVM. formance network service chains,” in Proc. Workshop
Hot Topics Middleboxes Netw. Funct. Virtualization,

5 G cellular networks promise to provide low


latency and high bandwidth to meet emerging,
demanding, performance-sensitive applications. A
pp. 26–31.
6. 3 GPP TR 23.799, “Study on architecture for next
generation system (Release 14),” V14.0.0, 2016.
key enabler is the use of NFV that offers flexibility 7. A. Mohammadkhan et al., “CleanG: A clean-slate
from being software based. We note that the use of EPC architecture and control plane protocol for next
NFV allows data plane and control plane function- generation cellular networks,” in Proc. ACM Workshop
ality to be supported on the same platform, unlike Cloud-Assisted Netw., 2016, pp. 31–36.
the traditional packet core network of multiple dis- 8. R. Shah et al., “Cuttlefish: Hierarchical SDN control-
tributed components. Scalability is enabled by the lers with adaptive offload,” in Proc. Int. Conf. Netw.
dynamic instantiation of the NFV platform support- Protocols, 2018, pp. 198–208.
ing the CDF and CCF. However, architectural and 9. A. Mohammadkhan et al., “Considerations for
implementation changes alone with NFV squander re-designing the cellular infrastructure exploiting
the opportunity for truly improving the cellular net- software-based networks,” in Proc. 24th Int. Conf.
work's performance if the protocols do not properly Netw. Protocols, 2016, pp. 1–6.
take advantage of the ability to consolidate the 10. D. Berger et al., “RSVP-TE: Extensions to RSVP for LSP
tightly interdependent cellular control and data tunnels,” IETF RFC 3209, Dec. 2001.
plane. We rethink the design of the control protocol 11. RFC 3031, MPLS “Multiprotocol label switching
to achieve low latency and high throughput by simpli- architecture,” 2001.
fication and using fewer messages. Complementing 12. C. Filsfils et al., “The segment routing architecture,” in
this, the backhaul network and protocols must also Proc. IEEE Global Commun. Conf., 2015, pp. 1–6.
be designed to judiciously utilize capacity and achieve 13. A. Bashandy et al., “Segment routing with MPLS user
low latency. The PPR protocol is a key enhancement plane,” draft-ietf-spring-segment-routing-mpls-14, IETF,
for the cellular backhaul. Low-latency applications Jun. 2018.
are enabled by having a more flexible plane, using the 14. C. Filsfils et al., “IPv6 segment routing header (SRH),”
ideas of BPP. draft-ietf-6man-segment-routing-header-14, IETF, Jun.
In this paper, we sought to analyze several of the 2018.
complexities of cellular networks without completely 15. U. Chunduri et al., “Preferred path routing – A
disrupting the 5G system architecture, and proposed next-generation routing framework beyond segment
carefully thought-out approaches to enhance the routing,” in Proc. IEEE Global Commun. Conf., 2018.
architecture and protocols of 5GC, the 5G backhaul, 16. T. Eckert, Y. Qu, and U. Chunduri, “Preferred path
and using a new backhaul transport data plane. routing (PPR) graphs beyond signaling of paths to
networks,” in IEEE Hi-Precis. Netw., 2018.
ACKNOWLEDGMENTS 17. U. Chunduri et al., “Preferred path routing (PPR) in
This work was supported in part by the NSF under Grant IS-IS,” Available at: https://tools.ietf.org/html
CNS-1618344 and in part by a grant from Futurewei Inc. /draft-chunduri-lsr-isis-preferred-path-routing
-01, 2018.
REFERENCES 18. U. Chundur et al., “Transport network aware mobility
1. 3 GPP TS 23.502, “Procedures for the 5G system,” 2018. for 5G,” Available at: https://tools.ietf.org/html/draft
2. P. Schmitt et al., “Control and user plane separation of -clt-dmm-tn-aware-mobility-01, 2018.

50 ComputingEdge April 2020


STANDARDS

19. “Optimized Mobile User Plane Solutions for 5G, draft K. K. RAMAKRISHNAN is a Professor of Computer Science
-bogineni-dmm-optimized-mobile-user-plane-01,” with the University of California, Riverside, Riverside, CA,
Available at: https://www.ietf.org/id/draft-bogineni USA. Contact him at kk@cs.ucr.edu.
-dmm-optimized-mobile-user-plane-01.txt, 2018.
20. R. Li, A. Clemm, U. Chunduri, L. Dong, and K. Makhijani, UMA CHUNDURI is a System Architect, with Future Networks
“A new framework and protocol for future networking team at America Research Center, Huawei Technologies,
applications,” in Proc. Workshop Netw. Emerg. Appl. Santa Clara, CA, USA. Contact him at uma.chunduri@huawei
Technol., 2018, pp. 21–26. .com.

ALI MOHAMMADKHAN is currently working toward the KIRAN MAKHIJANI is a Principal Engineer with Future
Ph.D. degree in computer science with the University of Networks, America Research Center, Huawei Technologies,
California, Riverside, Riverside, CA, USA. Contact him at Santa Clara, CA, USA. Contact her at Kiran.Makhijani@
amoha006@ucr.edu. huawei.com.

Rejuvenating Binary Executables Visual Privacy Protection Communications Jamming


■ ■
Policing Privacy Dynamic Cloud Certification Security for High-Risk Users
■ ■
Smart TVs Code Obfuscation The Future of Trust
■ ■

IEEE Symposium on
Security and Privacy

January/February 2016 March/April 2016 May/June 2016


Vol. 14, No. 1 Vol. 14, No. 2 Vol. 14, No. 3

IEEE Security & Privacy magazine provides articles


with both a practical and research bent by the top
thinkers in the field.
• stay current on the latest security tools and theories and gain invaluable practical and
research knowledge,
• learn more about the latest techniques and cutting-edge technology, and computer.org/security
• discover case studies, tutorials, columns, and in-depth interviews and podcasts for the
information security industry.

www.computer.org/computingedge 51
IEEE TRANSACTIONS ON
SUBMIT
TODAY 6867$,1$%/(&20387,1*
SCOPE
The IEEE Transactions on Sustainable Computing (T-SUSC) is a peer-reviewed journal devoted to
SXEOLVKLQJ KLJKTXDOLW\ SDSHUV WKDW H[SORUH WKH GLƬHUHQW DVSHFWV RI VXVWDLQDEOH FRPSXWLQJ 7KH
notion of sustainability is one of the core areas in computing today and can cover a wide range
of problem domains and technologies ranging from software to hardware designs to application
GRPDLQV6XVWDLQDELOLW\ HJHQHUJ\HƱFLHQF\QDWXUDOUHVRXUFHVSUHVHUYDWLRQXVLQJPXOWLSOHHQHUJ\
sources) is needed in computing devices and infrastructure and has grown to be a major limitation
to usability and performance.

Contributions to T-SUSCPXVWDGGUHVVVXVWDLQDELOLW\SUREOHPVLQGLƬHUHQWFRPSXWLQJDQGLQIRUPDWLRQ
SURFHVVLQJ HQYLURQPHQWV DQG WHFKQRORJLHV DQG DW GLƬHUHQW OHYHOV RI WKH FRPSXWDWLRQDO SURFHVV
These problems can be related to information processing, integration, utilization, aggregation, and
generation. Solutions for these problems can call upon a wide range of algorithmic and computational
frameworks, such as optimization, machine learning, dynamical systems, prediction and control,
decision support systems, meta-heuristics, and game-theory to name a few.

T-SUSC covers pure research and applications within novel scope related to sustainable computing,
such as computational devices, storage organization, data transfer, software and information
SURFHVVLQJ DQG HƱFLHQW DOJRULWKPLF LQIRUPDWLRQ GLVWULEXWLRQSURFHVVLQJ $UWLFOHV GHDOLQJ ZLWK
hardware/software implementations, new architectures, modeling and simulation, mathematical
models and designs that target sustainable computing problems are encouraged.

SUBSCRIBE AND SUBMIT


For more information on paper submission, featured articles, calls for papers,
and subscription links visit:

www.computer.org/tsusc
Pa p e rs :
Call for c t i o n s
E Tra n s a
IE E
p u te r s
on Com

Publish your work in the IEEE Computer Society’s flagship journal, IEEE
Transactions on Computers (TC). TC is a monthly publication with a wide
distribution to researchers, industry professionals, and educators in the
computing field.

TC seeks original research contributions on areas of current computing


interest, including the following topics:

• Computer architecture • Security and reliability


• Software systems • Machine learning
• Mobile and embedded systems • Quantum computing

All accepted manuscripts are automatically considered for the monthly


featured paper and annual Best Paper Award.

Learn about calls for papers and submission details at


www.computer.org/tc.
Get Published in the New IEEE Open
Journal of the Computer Society

Submit a paper today to the


premier new open access
journal in computing
and information technology.

Your research will benefit from


the IEEE marketing launch and
5 million unique monthly users
of the IEEE Xplore® Digital Library.
Plus, this journal is fully open
and compliant with funder
mandates, including Plan S.

Submit your paper today!


Visit www.computer.org/oj to learn more.
Conference Calendar

I EEE Computer Society conferences are valuable forums for learning on broad and dynamically shifting topics
from within the computing profession. With over 200 conferences featuring leading experts and thought lead-
ers, we have an event that is right for you. Questions? Contact conferences@computer.org.

MAY Boston, USA Smart Internet of Things), Bei-


3 May 29 June jing, China
• FCCM (IEEE Int’l Symposium • DSN (IEEE/IFIP Int’l Conf. on 16 August
on Field-Programmable Cus- Dependable Systems and Net- • Hot Chips, Palo Alto, USA
tom Computing Machines), works), Valencia, Spain 17 August
Fayetteville, USA 30 June • ACSOS (IEEE Int’l Conf. on
18 May • MDM (IEEE Int ’ l Conf. on Autonomic Computing and
• SP (IEEE Symposium on Secu- Mobile Data Management), Self- Or ganizing S ys tems),
rity and Privacy), San Fran- Versailles, France Washington, DC, USA
cisco, USA 19 August
• IPDPS (IEEE Int’l Parallel and JULY • RTCSA (IEEE Int’l Conf. on Em-
Distributed Processing Sym- 6 July bedded and Real-Time Comput-
posium), New Orleans, USA • ICME (IEEE Int’l Conf. on Multi- ing Systems and Applications),
30 May media and Expo), London, UK Gangnueng, South Korea
• ISCA (ACM/IEEE Int’l Sympo- 8 July 24 August
sium on Computer Architec- • ICDCS (IEEE Int’l Conf. on Dis- • FiCloud (Int’l Conf. on Future
ture), Valencia, Spain tributed Computing Systems), Internet of Things and Cloud),
Singapore Rome, Italy
JUNE 13 July 31 August
7 June • COMPSAC (IEEE Computer • RE (IEEE Int’ l Require -
• ARITH (IEEE Int’l Symposium Sof tware and Applications ments Eng. Conf.), Zurich,
on Computer Arithmetic), Conf.), Madrid, Spain Switzerland
Portland, USA 28 July
14 June • CBMS (IEEE Int’l Symposium SEPTEMBER
• CVPR (IEEE Conf. on Computer on Computer-Based Medical 7 September
Vision and Pattern Analysis), Systems), Rochester, USA • EuroS&P (IEEE European Sym-
Seattle, USA posium on Security & Privacy),
15 June AUGUST Genova, Italy
• ICHI (IEEE Int’l Conf. on Health- 1 August 14 September
care Informatics), Oldenburg, • JCDL (ACM/IEEE Joint Conf. • CLUSTER (IEEE Int’l Conf. on
Germany on Digital Libraries), Xi’an, Cluster Computing), Kobe,
22 June China Japan
• CSF (IEEE Computer Secu- 14 August 21 September
rity Foundations Symposium), • SmartIoT (IEEE Int’l Conf. on • ASE (IEEE/ACM Int’l Conf. on

72 April 2020 Published by the IEEE Computer Society 2469-7087/20 © 2020 IEEE
Automated Sof tware Eng.), 25 October • AIKE (IEEE Int’l Conf. on Arti-
Melbourne, Australia • VIS (IEEE Visualization Conf.), ficial Intelligence and Knowl-
24 September Salt Lake City, USA edge Eng.), Irvine, USA
• BigMM (IEEE Int’l Conf. on Mul- 14 December
timedia Big Data), New Delhi, NOVEMBER • BCD (IEEE/ACIS Int’l Conf. on
India 2 November Big Data, Cloud Computing,
27 September • ICFEC (IEEE Int’l Conf. on Fog and Data Science Eng.), Macao
• ICSME (IEEE Int’l Conf. on Soft- and Edge Computing), Mel- • CloudCom (IEEE Int’l Conf. on
ware Maintenance and Evolu- bourne, Australia Cloud Computing Technol-
tion), Adelaide, Australia 6 November ogy and Science), Bangkok,
28 September • SmartCloud (IEEE Int’l Conf. Thailand
• SecDev (IEEE Secure Develop- on Smart Cloud), Washington, 16 December
ment), Atlanta, USA DC, USA • BIBM (IEEE Int’l Conf. on Bio-
8 November informatics and Biomedicine),
OCTOBER • NAS (IEEE Int’l Conf. on Net- Seoul, South Korea
3 October working, Architecture, and
• PACT (Int’l Conf. on Parallel Storage), Riverside, USA
Architectures and Compila- 9 November
tion Techniques), Atlanta, USA • ICTAI (IEEE Int’l Conf. on Tools
5 October with Artificial Intelligence),
• EDOC (IEEE Int’l Enterprise Baltimore, USA
Distributed Object Comput- • ISMVL (IEEE Int’l Symposium
ing Conf.), Eindhoven, the on Multiple-Valued Logic),
Netherlands Miyazaki, Japan
• ICSE (IEEE/ACM Int’l Conf. on 15 November
Software Eng.), Seoul, South • SC, Atlanta, USA
Korea 16 November
17 October • FOCS (IEEE Symposium on
• MICRO (IEEE/ACM Int’l Sym- Foundations of Computer Sci-
posium on Microarchitecture), ence), Durham, USA
Athens, Greece • LCN (IEEE Conf. on Local Com- Learn more
18 October puter Net works), S ydney,
about IEEE
• MODELS (ACM/IEEE Int’l Conf. Australia
on Model-Driven Eng. Lan- Computer
guages and Systems), Mon- DECEMBER Society
treal, Canada
21 October
9 December
• CC (IEEE Int’l Conf. on Conver-
conferences
• FIE (IEEE Frontiers in Educa- sational Computing), Irvine, computer.org/conferences
tion Conf.), Uppsala, Sweden USA
T
EN
EV
EW
N
IEEE
QUANTUM
WEEK

20

,
D
EL
20

FI
ER

SA M
B

U O
TO

O O
D BR
C
O

R R—
6
–1

LO E
12

A
O V
C EN
D
IEEE Quantum Week 2020 Is Open for Submissions
Participation opportunities are available for IEEE Quantum Week includes the following
the inaugural IEEE International Conference on technical paper tracks:
Quantum Computing and Engineering (QCE
• Quantum Communications, Sensing,
2020) to be held 12–16 October 2020, in Denver—
Cryptography
Broomfield, CO.
• Quantum Photonics and Optics
IEEE Quantum Week aims to be a leading venue • Quantum Computing
for presenting high-quality original research, • Quantum Algorithms & Information
ground-breaking innovations, and compelling • Quantum Applications and Simulating Nature
insights in quantum computing, engineering, • Quantum Engineering
and technologies. • Quantum Benchmarks & Measurements
• Quantum Education
Authors are invited to submit proposals for
technical papers, posters, tutorials, workshops, Papers accepted by IEEE QCE will be submitted
and panels. Submission schedules are available at to the IEEE Xplore Digital Library. The best papers
qce.quantum.ieee.org/important-dates. will be invited to the journals IEEE Transactions
on Quantum Engineering (TQE) and ACM
Transactions on Quantum Computing (TQC).

Submission instructions and details:


qce.quantum.ieee.org/callforcontributions

Council on Superconductivity

You might also like