Professional Documents
Culture Documents
Full Chapter Building Machine Learning Systems With A Feature Store Batch Real Time and LLM Systems Early Release Jim Dowling PDF
Full Chapter Building Machine Learning Systems With A Feature Store Batch Real Time and LLM Systems Early Release Jim Dowling PDF
https://textbookfull.com/product/biota-grow-2c-gather-2c-cook-
loucas/
https://textbookfull.com/product/interpretable-ai-building-
explainable-machine-learning-systems-meap-v02-ajay-thampi/
https://textbookfull.com/product/mastering-kafka-streams-and-
ksqldb-building-real-time-data-systems-by-example-1st-edition-
mitch-seymour/
https://textbookfull.com/product/machine-learning-and-security-
protecting-systems-with-data-and-algorithms-first-edition-chio/
Machine Learning with PySpark: With Natural Language
Processing and Recommender Systems 1st Edition Pramod
Singh
https://textbookfull.com/product/machine-learning-with-pyspark-
with-natural-language-processing-and-recommender-systems-1st-
edition-pramod-singh/
https://textbookfull.com/product/learning-tensorflow-a-guide-to-
building-deep-learning-systems-1st-edition-tom-hope/
https://textbookfull.com/product/embedded-systems-vol-3-real-
time-operating-systems-for-arm-cortex-m-microcontrollers-
jonathan-valvano/
https://textbookfull.com/product/hands-on-machine-learning-with-
scikit-learn-and-tensorflow-early-release-2nd-edition-aurelien-
geron/
https://textbookfull.com/product/feature-engineering-for-machine-
learning-and-data-analytics-first-edition-dong/
Hopsworks
Building Machine Learning Systems
with a Feature Store
Batch, Real-Time, and LLM Systems
With Early Release ebooks, you get books in their earliest form—the
author’s raw and unedited content as they write—so you can take
advantage of these technologies long before the official release of
these titles.
Jim Dowling
Building Machine Learning Systems
with a Feature Store
by Jim Dowling
The views expressed in this work are those of the author and do not
represent the publisher’s views. While the publisher and the author
have used good faith efforts to ensure that the information and
instructions contained in this work are accurate, the publisher and
the author disclaim all responsibility for errors or omissions,
including without limitation responsibility for damages resulting from
the use of or reliance on this work. Use of the information and
instructions contained in this work is at your own risk. If any code
samples or other technology this work contains or describes is
subject to open source licenses or the intellectual property rights of
others, it is your responsibility to ensure that your use thereof
complies with such licenses and/or rights.
[TO COME]
Brief Table of Contents (Not Yet
Final)
Preface
Introduction
Charles Fyre, developer of the Full Stack Deep Learning course, said
the following of ID2223:
In 2017, having a shared pipeline for training and prediction
data that updated automatically and made models available as
a UI and an API was a groundbreaking stack at Uber. Now it’s
stanard part of a well-done (ID2223) project.
With Early Release ebooks, you get books in their earliest form—the
author’s raw and unedited content as they write—so you can take
advantage of these technologies long before the official release of
these titles.
This will be the introduction of the final book. Please note that the
GitHub repo will be made active later on.
In the early 2010s, when both machine learning and deep learning
exploded in popularity, many companies became what are now
known as “hyper-scale AI companies”, as they built the first ML
systems using massive computational and data storage
infrastructure. ML systems such as Google translate, TikTok’s video
recommendation engine, Uber’s taxi service, and ChatGPT were
trained using vast amounts of data (petabytes using thousands of
hard drives) on compute clusters with 1000s of servers. Deep
learning models additionally need hardware accelerators (often
graphical processing units (GPUs)) to train models, further increasing
the barrier to entry for most organizations. After the models are
trained, vast operational systems (including GPUs) are needed to
manage the data and users so that the models can make predictions
for hundreds or thousands of simultaneous users.
Naturally, many organizations have not had the same resources that
were available to Uber to build equivalent ML infrastructure. Many of
them have been stuck at the model training stage. Now, however, in
2024, the equivalent ML infrastructure has become accessible, in the
form of open-source and serverless feature stores, vector databases,
model registries, and model serving platforms. In this book, we will
leverage open-source and serverless ML infrastructure platforms to
build ML systems. We will learn the inner workings of the underlying
ML infrastructure, but we will not build that ML infrastructure—we
will not start with learning Docker, Kubernetes, and equivalent cloud
infrastructure. You no longer need to build ML infrastructure to start
building ML systems. Instead, we will focus on building the software
programs that make up the ML system—the ML pipelines. We will
work primarily in Python, making this book widely accessible for
Data Scientists, ML Engineers, Architects, and Data Engineers.
Figure I-1. A feature is a measurable property of an entity that has predictive power for the
machine learning task. Here, the fruit’s color has predictive power of whether the fruit is an
apple or an orange.
In figure 2, you can also see there are a small number of oranges
that are not correctly classified by the decision boundary. Our model
could wrongly predict that an orange is an apple. However, if the
model predicts the fruit is an orange, it will be correct - the fruit will
be an orange. We can say that the model’s precision is 100% for
oranges, but is less than 100% for apples.
Let’s complicate this simple model. What if we add red apples into
the mix? Now, we want our model to classify whether the fruit is an
apple or orange - but we will have both red and green apples, see
figure 3.
Figure I-3. The red apple complicates our prediction problem because there is no longer a
linear decision boundary between the apples and oranges using only color as a feature.
We can see that red apples also have zero for the blue channel, so
we can ignore that feature. However, in figure 4, we can see that the
red examples are located in the bottom right hand corner of our
chart, and our model (a linear decision boundary) is broken—it
would predict that red apples are oranges. Our model’s precision and
recall is now much worse.
Figure I-4. When we add red apples to our training examples, we can see that we can no
longer use a straight line to classify fruit as orange or apple. We now need a non-linear
decision boundary to separate apples from oranges, and in order to learn the decision
boundary, we need a more complex model (with more parameters), more training
examples, and m.
NATURAL HISTORY
SCORPION-PROOF PLATFORM.