rar2024 AHistory of AL
A History
of Al
We chose key moments in the history of AI and Machine
Learning. The first was "Deep Learning" that launched a decade
ago that made it accessible and easy to use.
The others were "Visual Machine Learning” and "Machine
Learning" to create new methods that could have a better
understanding of the neural network.
It is exciting that this is the case. It is a new and innovative
approach that will eventually be used on the public-health
market.
Note: This intro text is generated with the GPT-2 Model by
OpenAl, based on the first sentence: ‘We chose key moments in
the history of AI and Machine Learning.’
hitpssiwwuchistoryfa.n! sr9raro2 Astor of AL
1800 ~ 1850
Linear Regression -
First Machine
Learning?
It is difficult to exactly pinpoint
where in history the first Linear
Regression
ia.org/wiki/Linear_regression)
and who was
For sure, names such as Legendre
(https://en.wikipedia.org/wiki/Adrien-
Marie_Legendre), Gauss
://en.wikipedia.org/wiki/Carl_Friedrich_Gauss), ‘uf
|.wikipedia.org/wiki/Francis_Galton),
htpssiwwuchistoryfa.n! 209riar02t A tistory of
anid) Pearson
(https://en.wikipedia.org/wiki/Karl_Pearson)
are involved and contributed in
making Linear Regression Analysis
famous, arguably the first Machine
Learning Model.
‘An examination of publications of Sir
Francis Galton and Karl Pearson
revealed that Galton's work on
inherited characteristics of sweet
peas led to the initial
conceptualization of linear
regression.
Galtor's first regression line was
presented at a lecture in 1877. Only
under Pearson's later treatment did 'r’
come to stand for the correlation
coefficient (Pearson 1896).
In 1896, Pearson published his first
rigorous treatment of correlation and
regression in the Philosophical
Transactions of the Royal Society of
London. In this paper, Pearson
credited Bravais (1846) with
ascertaining the initial mathematical
formulae for correlation. Pearson
noted that Bravais happened upon
the product-moment (that is, the
“moment” or mean of a set of
products) method for calculating the
correlation coefficient but failed to
prove that this provided the best fit
to the data
Is it really Machine Learning, when a
computer/algorithm learns to fit a
line through points on a flat surface?
Or maybe if its dimension is 3, 10, or
100?
hitpssiwouchistoryfa.n!
329riar02t A tis of
Many times we call systems
intelligent when itis able to provide
us the solution, in a large space of
possible answers (search). The right
move in a game, the right object on a
photo, the right answer to some
question formed in natural language
Fitting the best line through a set of
points, is perhaps the most clean
and simple version of this, and
maybe therefore the reason many
Machine Learning courses start with
Linear Regression.
READ LESS
hitpssiwwuchistoryfain!
1941
First Electronic
Statistical Computer
The first prototype of an electronic
device to perform statistics
calculation was built by John
Atanasoff
(https://en.wikipedia.org/wiki/John_Vince
Statistical computing was mostly
done in statistical labs, using
mechanical tabulators with punched
cards.
429rar2024
1950
Can machines think?
| propose to consider the question,
‘Can machines think? This should
begin with definitions of the meaning
hitpssiwwuchistoryfain!
Atistory of AL
This first electronic device had a lot
of similarities with modern
computers. It was electronic and had
a memory unit, a central processor,
and binary arithmetic. The same as
your smart-phone!
Atanasoff built this machine to solve
linear equations. An often told story
is that Atanasoff set off on along
drive across lowa to think about this
problem sometime during the winter
of 1937-1938.
Several hundred miles later, at a
roadside bar in Illinois, the basic
elements for a machine to solve
systems of linear equations were
conceived.
Did you know that Linear Regression
comes down to solving a linear
equation? This machine is the first
Linear Regression Solver Machine.
READ LESS
929Astor of AL
tA h cha
subject to guess if he in
a machine, or a real human.
Ever doubted if the person on the
other side is a human or machine?
Try teach it something new, because
learning something new with little .
information is something humans ~
still excel at, at least, for now... #
READ LESS “a
1959 ea
hitpssiwwuchistoryfain!‘nar2024 Astor of AL
ld of stu t gives
the ability to =
being
ic arithmetic and math to
jut so far no human was
le to lay down simple rules that S
ne a computer see the ee
ae. acat log. The
peter learned s herself.
To less extend the same goes fre
complex games, like Chess, Go, and
Poker, where the impossible task of
transferring some masters intuition
to rules into a computer rises.
see ion is something we cannot
and the same time the very
hy \g we want a computer be able to
to. Humans can do so much,
can explain so Leer ot
being capable of far more than you
can explain is Polyani paradox.
Arthur's definition is still valid today, ae
many times the computer's learning
ability amazes us. Games like Poker,
Go, and complex computer games
all are no more a challenge.
Fortunately there is en
h oe
hitpssiwwuchistoryfa.n! 728Astor of AL
Can this be considered the firs
(commercial) Machine-Learning
software?
The point where statistics becomes.
machine learning is not exact, but for
sure ML and A| is built on statistical
concepts, hence we want to mention
the first commercial sot
packages. Before the 70's! Before
the internet!
SAS was developed at North
Carolina State University from 1966
until 1976, when SAS Institute was
incorporated. SAS was further in the *
1980s and 1990s with the addition
of new statistical procedures,
additional components and the
introduction of JMP.
SPSS released its first version in
1968 as the Statistical Package for
the Social Sciences (SPSS) after
being developed by Norman H. Nik
and C. Hadlai Hull.
“ee =
hitpssiwwuchistoryfa.n! 229raro2 Astor of AL
Back Prépagating!
Artificial Neural
Networks can finally
learn!
eural Networks have a long history,
ack to before 1950. Mostly
os :
{- oe jenting, and creating the
mathematical foundation to make
them work and learn.
1986 was the year David Rumelhart,
Geoffrey Hinton, and Ronald
Williams published their paperon
Back propagation. Back propagation
is the most popular way to make
artificial neural networks learn, and
will probably be for the next coming
concepts for net
med, the |
pa
.
hitpssiwwuchistoryfa.n! 929raro2
hitpssiwwuchistoryfa.n!
Astor of AL
similar learning al
paper mentioned above is. pethaps
once networks to
scognize p de char
with an extreme Soren
fern Neural Networks (often
ous andkey pointin. =
Neural Networks (CNN), "
called Deep Learning), contain much
more layers, and come in many
variations; VGG, Inception, ResNet,
ResNeXt, DenseNet, the list is
endless! They are all examples of
st 2010 architectures. All have 2
things in common, they are trained
he ick propagations, and in the
‘many layers you can find CNN layers.
LSTM, a special kind of neural
network which had the notion of a
memory, was discovered about 10
years later by Sepp Hochreiter and
Jiirgen Schmidhuber. Notice how all
important foundations for modern
Deep Learning were done before
2000.
Yann LeCun and Yoshua Benjio, and
Geoffrey Hinton
Award in 2018,
é as recognition fraro2 Astor of AL
computer explore and min
data itself, usually by crawling il
subsets in databases. What will she
find?!
The step from a classical methods
like linear regression, to data mining,
can be seen as the first dimension 4 ast a
explosion, the space thecomputer “
needed to explore and find — oe
parameters for got bigger, and our a
hardware and software could _
support the needed resources.
Databases, CPUs, and software were
ready in the 90's.
Association rule learning, DBSCAN
Clustering, and SVM are examples of
algorithms invented in the 90's, and
are often explained as common data
mining techniques.
Many moments from this point, we ‘A
see ajump in complexity ofthe = .
m nd innovation in
a to support the
. We see the same evoluti
i ~
hitpssiwwuchistoryfa.n! 129riar02t
We have difficulty to assign the term
intelligence to a single cell organism,
but for sure we call our own species
intelligent, or at least the brighter few
among us.
Is data mining Machine Learning? !s
Data Mining Al?
READ LESS
hitpssiwwuchistoryfa.n!
Astor of AL
2000
The use of GPUs
The first GPU (Graphics processing
unit)
(https://www.doc.ic.ac.uk/~jce317/histor
machine-learning.html¥the-
importance-of-gpus), which was
widely promoted as a ‘GPU’ on the
consumer market, was developed by
Nvidia in 1999. This was the GeForce
256 and was used for vector
calculations in video games,
shooting with guns in a 3D space!
This technology gave an enormous
boost to the graphic processing
performance and years later it was.
discovered by scientists and
1229nano History of AI
engineers
useful in sp
calculations.
et showed the value
cesses
ror is around 5%,
jore ae in 2016.
ie with GPUs, lke pretty much all
its successors as well.
GPUs are very valuable for machine
learning. They can have around 200
times more processors per chip than
ie From 2017, Nvidia started to
| develop GPUs with tensor cores,
8.
* ee aoe for machine learning.
Who would ever predicted, that
shooting in a 3D wees oer?
computational simil ep.
Learning, used in Al? Both are
examples of letting a computer
doing many relatively small matrix
multiplications, Which one do you
prefer?
READ LESS
i»
hitpssiwwuchistoryfa.n! 1329raro2 Astor of AL
some a practical approach. _
In his paper, he described what are
now standard methods at typical
"Machine Learning’ courses. He is
also the creator of one of the most
popular Machine-Learning
Algorithms, the Random Forest.
Not many know, Leo Brame also
was involved in founding the ideas
for another popular Machine
Learning Model, the XGBoost. The
idea to improve a weak model, along
the derivative of the loss function, by a
adding a correcting ‘booster’ on top,
was partly from Leo.
If you want to pin point a place in
history, where classical statistics
and modern machine learning
deviate, this paper could be that fs f
point, back in 2001. ah
READ Ess ey +=
- e 7 a =
es
hitpssiwwuchistoryfa.n! 1429raro2 Astor of AL
k, puter
tion, Ut interpreting
is far more difficult
it seems at frstiee
ita aoe. Watson
computer
(https://www.ibm.com/ibm/history/ibm1C
competed on Jeopardy! against the.
TV quiz shows two biggest all-time
champions and won. This was the
nore Al could win at a game
jich required language skills.
lenge was in the language:
— questions on this show are full
of subtlety, puns and wordplay. ~
Before this, computers were able to
answer straight forward questions,
but this combination of
understanding language in it's full ea
subtlety and to find the right answer
from several options in a knowledge
bank was too much asked, until IBM
Watson.
_ IBM Watson was a pot
. computer sof
‘ developed by IBI
h “en ‘i
hitpssiwwuchistoryfa.n! 1529raro2 Astor of AL
we might be part of a classic hype.
Itis difficult to exactly point where
the Big Data hype peak was at its
highest, we chose March 2014, when
Horton works raised 100M Dollar in
preparation to go public.
‘According to Hadoops co-founders, 3 = x
Doug Cutting and Mike Cafarella, the bee
genesis of Hadoop was the Google a
File System paper that was i
published in October 2003. This
paper spawned another one from
Google - "MapReduce: Simplified
Data Processing on Large Clusters’. *
Back in 2014, there were many
Hadoop distributions/versions.
Horton and Cloudera were
considered competitors, and are now
merged. IBM's Hadoop Distribution,
Big Insights, is gone. ="
_ 2014is also the year the Big Data
po st in Utrecht, a yearly
jere Hadoop vendors and
i _™
hitpssiwwuchistoryfa.n! 1629‘nar2024 Astor of AL
PB of data. Kids nowadays are
spoiled with Spark, the early Hadoop
adaptors had to write MapReduce()
in java code! How hardcore!
READ LESS
Operationalize Al aif
ML?
h oe
hitpsihewtistoryot 129‘nar2024 Astor of AL
jot as ea! !
MLOps are
methods to supp‘ smooth way to:
create, deploy, and maintain ML
ag aaupen r
eran Sn ie
e este to create, update, and —
ee Sntciner von. on Geman
Tools, libraries, and methods come
and go, but the challenge to
operationalize IT easy and robust will
always stay, and Machine Learning is
no exception. Thinking of pleasant
ways to operational and Alis
‘ %. »
' ao still a growing toy probably
ie te \r the next coming years.
READ LESS
2016 -
XGBoost and Kag
hitpssiwwuchistoryfa.n!raro2 Astor of AL
Kaggle, is the World Championship _ be
for Data Science, you compete with
other analysts on a clear task given a
dataset. In 2017, it reached 1 Million
users.
Kaggle is competitive Data Science,
and this is a different game than
solving real world problems. For
once, the dataset eee task is
given to you.
But for sure, XGBoost winning
competition after competition, made
people curious and apply it on real
world problems as well. XGBoost ae
became and still is a very popular
model for Machine Learning,
powerful and versatile.
It belongs to the family of Boosters,
where weaker models stacked on ‘
each other produce very sharpend
results, as each model has the ability
to mare the mistake of t! ey
revious.
i Eee ‘i
hitpssiwwuchistoryfa.n! 1929‘nar2024 Astor of AL
in
Strongest Go
layer with Al
é
Reenter we = say... ~
“Computers will never be able to...
[INSERT CHALLENGE]’. Well, for many
years Go was the example to cite as
a task Al will probably not be able to
do at Human expert level in the
ming 10 years. The folks from
_ Deepmind did it, d Neural
as a key component for
their Al, together with self learning
capabilities. > a.
Checkers, Chess, Go, Poker,
Starcraft... All were seen impossible
at some point in time for computers
to master, for some reason. ~
.
s
If a expert is among the strongest of
the world in a mind game, we call
that person Intelligent. In 1997
DeepBlue defeated Garry Ka
h Sa Y
hitpssiwwuchistoryfa.n! 20129rer024 Astor of AL
READ LESS:
2016
When Things go
wrong...
In this year, a twitter Al bot was
released online, and learned to talk
with humans. Just after 16 hours,
the bot needed to be shut-down. It
was too good at learning from
humans, and humans had far more
fun teaching it bad manners, than
having a serious conversation.
This showed us 2 things: Humans
can be mean. If Al learns from
humans, Al can be mean. Our History
hasn't always been pretty and fair,
and if we are not careful, Al will
simply be a human history
amplifier.
Cathy O'Neil observed this, and
unfortunately not only theoretical
Her book, weapons of Math
Destruction, was published in 2016
as well. She explains the dangers of
massively apply Machine-Learning in
our society, with shameful examples
of things already happened.
hitpssiwwuchistoryfa.n!raro2 Astor of AL
taken offline, it
woman.
Also, hackers shown self-driving cars
can be fooled with stickers on the
road, invisible for humans, making
self driving cars take sudden turns.
Another problem with Al, it becomes.
0 good mimicking human a
behaviour, it can generate realistic
human faces, and can even
impersonate humans like Presidents, Soe ™
and celebrities. 7
Is that Linkedin Request a real
person? Was the Photo perhaps a
machine generate? Al can Fool us!
The other way around is also true,
Humans can fool Al. By changing a
image invisible for a human, the
image of a cute cat can become a
truck inthe eyes of a Al model. So...
Humans can fool Al,Alcanfool =
{ rage it is fair after all?
i E See ‘i
hitpssiwwuchistoryfa.n! 2229‘nar2024 Astor of AL
uplayme —
a"
_ some Jazz? Se
SI nn as ne o~
book about the dangers of
technology, and people massively
paid to put a personal assistant in
their home. We certainly are a weird
species! :
te One of first type of Virtual Assistants
teas ter Child (released in
-~ i sai text only and could play
games, check the weather and look
things up. The conv
capability was limited.
The first modern virtual assistant
was Siri, which was released in 201 i
It was able to set the time, assist at
text messages and dial somebody.
Later on it became able to perform
more complex tasks, like giving
directions or advice on rant.
h oe
hitpsihewtistoryotriar02t Aisin of
released. Virtual assistants became
quite common to consumers,
especially on mobile phones and
home devices.
READ LESS
November 2019
Hey! Want to buy a
real robot?
The child's imagination of what a
Robot was, became reality when in
2019 Boston Dynamics made its first
commercially available Robot, spot!
Different than a Robot arm doing
repetitive task in a factory, Spot can
self balance, avoid or climb over
obstacles, and has some sense of
her surroundings.
READ LESS
2019
Automated Machine
hitpssiwwuchistoryfa.n!
2129‘nar2024 Astor of AL
Ina way, Automated Machine
arnin« computer
Promenade ape
jomenon thi data, like
ergs
‘intuition’, through data. -
"TPOT is well known Open-Source
Automated Machine Learning library.
IBM has named hers ‘AutoAl
(https://www.ibm.com/cloud/watson-
___ studio/autoal). Wiad
. ee list ing...
“—_ = ted Machine learning can
return very sharp predictive models,
given a fixed d:
not (yet) do, is automatically
combine relevant data from
elsewhere, create a richer context so
to speak.
READ LESS
~ 2
h 2- A AMillinn
hitpsiwwchistoryfa.n! 25129‘ar2024 Astor of AL
Q: How many eyes does ag
have?
AA giraffe has two eyes
Q: How many legs does a frog have?
AA frog has four legs
Q: Are there any animals with three
legs?
—-
A: No, there are no animals with three
legs.
What you saw, was a conversation
with the GPT-3 model
(https://lacker.io/ai/2020/07/06/giving-
gpt-3-a-turing-test. html). It can A
generate text based on a single S
sentence, do Q&A like above,
translate and transform, actually
quite a lot! It is Version 3, the latest
and greatest, larger and stronger
than all its predecessors. Version 3 4
is called ‘Scary Good!’ bysome.
GPT-2is quite OK, and possible to ‘
run of ly fast computer. For
you need a computer th:
h ne ‘ih
hitpssiwwuchistoryfa.n! 26129‘nar2024 Astor of AL
The Future?
&
e start relationships with Al
4
é Soe like in the movie ill we start a
i : in the Terminator?
a
Or, maybe both?
Will we have
Intelligence - computers that actually,
think, learn, and act themselves - in
our lifetime?
Fortunately, Al and Machine Learning
have already shown to be useful
without being self-aware (whatever
that may be?). And if you are afraid
“ of wars, killing, and des
our world: hum:
pretty good at
h a
hitpssiwwuchistoryfa.n! 21129raro2 Astor of AL
“ee
reno vess
Did we miss your hero? Want to ask us something? Feel free to reach us; _
Robin van Tilburg (https://www.linkedin.com/i in-van-tilburg-4674213/) and
Willem Hendriks (https://www.linkedin.com/it mhendriks/).
We used Pexels for many images and the backgro ), and want to thank and
mention the content creators Olia Danilevich, Guilherme Almeida, Suzy Hazelwood, _
freestocks.org, Jason Dolin, cottonbro, and Pressmaster, for creating;
Guys behind Notebook (https://www.pexels.com/photo/two-men-looking-at-a-
laptop-4974920/)
‘Smiling Lade (https://www.pexels.com/photo/womanr-sitting-and-smiling- ae!
1858175/) ~
Robot (https://www.pexels.com/photo/blue-and-yellow-robot-toy-4013157/)
Container (https://www.pexels.com/photo/steel-container-on-container-dock-
122164/)
Cliff (https://www.pexels.com/photo/geological-formation-2855113/)
Rainbow Hands (https://www.pexels.com/photo/persons-hands-wi
colors-3693901/) .
* Clouds (video) (https://www, com/video/cumulus-cloud
sky-3129769/)
an eee GD
hitpssiwwuchistoryfa.n! 20129‘nar2024 A History of A
/) from Sava Lazic, '
ground-vi
video: 01 m
=" =
hipeswinnhisory ofa