Professional Documents
Culture Documents
net/publication/350382496
CITATIONS READS
0 4,236
1 author:
Abdelghny Orogat
Carleton University
8 PUBLICATIONS 6 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Abdelghny Orogat on 25 March 2021.
KEYWORDS
Data Lakes, Metadata, Open data
1 ABSTRACT
This is a report that summarizes the book "Databases and big data
set Laurent, Anne Laurent, Dominique Madera, Cedrine - Data lakes
(2020, ISTE, Ltd., Wiley) - libgen.liFile" and on existing Data Lakes
and their current features and limitations.
2 INTRODUCTION
James Dixon, a Penthao CTO, coined the word data lake for the
first time [4]. Dixon predicted that data lakes would be massive
collections of raw data, structured or unstructured, that users could
use for sampling, mining, or analytical purposes in this seminal
work. The idea of a data lake, according to Gartner [2] in 2014,
was nothing more than a modern way of storing data at low cost.
This point was revised a few years later, based on the fact that data
lakes are now considered important in many businesses [7]. As a
result, Gartner now finds the data lake model to be the holy grail
of knowledge management when it comes to innovating through
data value.
The characteristics of data lakes are described as storing data
in its original state at low cost, according to one of the earliest
academic papers about data lakes [1]. Since (1) data servers are
inexpensive and (2) no data transformation, cleaning, or planning
is needed, the cost is kept low. In 2016, Bill Inmon published a
Figure 1: Data Lake Architecture
book on a data lake architecture [3] in which the issue of storing
useless or impossible to use data is addressed. More specifically, Bill
Inmon argues in this book that the data lake architecture should
move towards information systems, rather than storing only raw 3.1 Ingestion Layer
data, in order to avoid storing “prepared” data via a method like
The Ingestion Layer is in charge of bringing data into the data lake
ETL (Extract-Transform-Load), which is commonly used in data
system from various sources. One of the main features of the data
warehouses.
lake concept is the ease with which any type of data can be ingested
Following other works, the most influential work on data lake ar-
and loaded. Data lakes, on the other hand, have been repeatedly
chitecture, modules, and positioning is discussed in IBM [6], since
reported as requiring governance in order to avoid becoming data
the focus is on data governance, specifically the metadata cata-
swamps. Administrators of data lakes are in charge of this critical
log. The authors of [6] pointed out that the metadata catalog is a
topic.
key component of data lakes that prevents them from being data
The metadata extractor is the most key component of the in-
"swamps."
gestion layer. It should make it easier for data lake administrators
to configure new data sources and make them accessible in the
3 DATA LAKES ARCHITECTURES data lake. To accomplish this, the metadata extractor should extract
Data Lake, as defined in the IBM Redbook, is a set of centralized as much metadata as possible from the data source (for example,
repositories containing vast amounts of raw data (either structured schemas from relational or XML sources) and store it in the data
or unstructured), described by metadata, organized into identifi- lake’s metadata storage. The raw data, in addition to the meta-
able datasets, and available on demand. Figure 1 shows a standard data, must be absorbed into the data lake. As noted in [9], since
proposal for an architecture of a data lake system [9]. The system the raw data are kept in their original format, this is more like
consists of four layers which are 1) Ingestion Layer, 2) Storage Layer, a “copy” operation, which is certainly less complex than an ETL
3) Transformation Layer, and 4) Interaction Layer. (Extract-Transform-Load) process in data warehouses.
3.2 Storage Layer
The metadata repository (discussed in Section 4) and the raw data
repositories are the two key components of the storage layer. Since
raw data repositories must be stored in their native format, data
lake environments must support a variety of storage systems for
relational, graph, XML, and JSON data. Hadoop appears to be a
strong candidate for the storage layer’s basic platform. To support
data fidelity, however, additional components such as Tez or Falcon
are needed.
Figure 4: Tool of Record Metadata Architecture 4.1.5 Federated Metadata Architecture. The federated repository
can be considered a virtual enterprise metadata repository. An
architecture for the federated repository would look something
like Figure 6. The bottom of the diagram is a series of separate
attained due to the following limitations. sources of metadata and other metadata repositories, some located
Limitations: on premise and some located in the cloud. Virtualization is achieved
• Cost and complexity of implementation. through a series of connectors, with each connector designed to
• It relies on a repository tool with extensive metadata ex- access the metadata held within the source system or repository.
change bridges. These connectors provide a means of abstraction to the integrated
• It requires extensive customization, requiring several years metadata store. This system comprises of automated services to
of effort and cost. ingest new metadata or analyze new data stores as they are added
to the overall integrated metadata system. Finally, a single-user
4.1.3 Tool of Record Metadata Architecture. It also uses a central- experience allows both business and IT users to find, access and
ized repository, but this centralized repository stores only unique view metadata. The portal provides different search methods such as
metadata from metadata sources of record as shown in Figure 4. SQL, free text and graph, in order to allow different communities to
The tool of record centralized repository is like building a data search for the metadata they are most interested in. The portal also
warehouse for metadata. provides collaboration to allow users to tag or provide additional
Although this architecture has less complexity to implement descriptions and commentary on the metadata.
because it only collects unique metadata, not all., it still has some Although this architecture focuses on automated curation and
limitations. providing a single view of metadata to the user and leverages an
Limitations: organization’s current tools, it still has some limitations.
Limitations:
• It relies on customization.
• It faces the same challenges of any data warehouse imple- • It is not ideally suited for manual metadata curation.
mentation (e.g. semantic and granularity issue resolution • Limited market tool set to provide the virtualization.
before metadata loading). • Architectural approach and software still maturing.
3
Figure 7: Functional Data Lake Architecture