You are on page 1of 29
rar2024 AHistory of AL A History of Al We chose key moments in the history of AI and Machine Learning. The first was "Deep Learning" that launched a decade ago that made it accessible and easy to use. The others were "Visual Machine Learning” and "Machine Learning" to create new methods that could have a better understanding of the neural network. It is exciting that this is the case. It is a new and innovative approach that will eventually be used on the public-health market. Note: This intro text is generated with the GPT-2 Model by OpenAl, based on the first sentence: ‘We chose key moments in the history of AI and Machine Learning.’ hitpssiwwuchistoryfa.n! sr9 raro2 Astor of AL 1800 ~ 1850 Linear Regression - First Machine Learning? It is difficult to exactly pinpoint where in history the first Linear Regression ia.org/wiki/Linear_regression) and who was For sure, names such as Legendre (https://en.wikipedia.org/wiki/Adrien- Marie_Legendre), Gauss ://en.wikipedia.org/wiki/Carl_Friedrich_Gauss), ‘uf |.wikipedia.org/wiki/Francis_Galton), htpssiwwuchistoryfa.n! 209 riar02t A tistory of anid) Pearson (https://en.wikipedia.org/wiki/Karl_Pearson) are involved and contributed in making Linear Regression Analysis famous, arguably the first Machine Learning Model. ‘An examination of publications of Sir Francis Galton and Karl Pearson revealed that Galton's work on inherited characteristics of sweet peas led to the initial conceptualization of linear regression. Galtor's first regression line was presented at a lecture in 1877. Only under Pearson's later treatment did 'r’ come to stand for the correlation coefficient (Pearson 1896). In 1896, Pearson published his first rigorous treatment of correlation and regression in the Philosophical Transactions of the Royal Society of London. In this paper, Pearson credited Bravais (1846) with ascertaining the initial mathematical formulae for correlation. Pearson noted that Bravais happened upon the product-moment (that is, the “moment” or mean of a set of products) method for calculating the correlation coefficient but failed to prove that this provided the best fit to the data Is it really Machine Learning, when a computer/algorithm learns to fit a line through points on a flat surface? Or maybe if its dimension is 3, 10, or 100? hitpssiwouchistoryfa.n! 329 riar02t A tis of Many times we call systems intelligent when itis able to provide us the solution, in a large space of possible answers (search). The right move in a game, the right object on a photo, the right answer to some question formed in natural language Fitting the best line through a set of points, is perhaps the most clean and simple version of this, and maybe therefore the reason many Machine Learning courses start with Linear Regression. READ LESS hitpssiwwuchistoryfain! 1941 First Electronic Statistical Computer The first prototype of an electronic device to perform statistics calculation was built by John Atanasoff (https://en.wikipedia.org/wiki/John_Vince Statistical computing was mostly done in statistical labs, using mechanical tabulators with punched cards. 429 rar2024 1950 Can machines think? | propose to consider the question, ‘Can machines think? This should begin with definitions of the meaning hitpssiwwuchistoryfain! Atistory of AL This first electronic device had a lot of similarities with modern computers. It was electronic and had a memory unit, a central processor, and binary arithmetic. The same as your smart-phone! Atanasoff built this machine to solve linear equations. An often told story is that Atanasoff set off on along drive across lowa to think about this problem sometime during the winter of 1937-1938. Several hundred miles later, at a roadside bar in Illinois, the basic elements for a machine to solve systems of linear equations were conceived. Did you know that Linear Regression comes down to solving a linear equation? This machine is the first Linear Regression Solver Machine. READ LESS 929 Astor of AL tA h cha subject to guess if he in a machine, or a real human. Ever doubted if the person on the other side is a human or machine? Try teach it something new, because learning something new with little . information is something humans ~ still excel at, at least, for now... # READ LESS “a 1959 ea hitpssiwwuchistoryfain! ‘nar2024 Astor of AL ld of stu t gives the ability to = being ic arithmetic and math to jut so far no human was le to lay down simple rules that S ne a computer see the ee ae. acat log. The peter learned s herself. To less extend the same goes fre complex games, like Chess, Go, and Poker, where the impossible task of transferring some masters intuition to rules into a computer rises. see ion is something we cannot and the same time the very hy \g we want a computer be able to to. Humans can do so much, can explain so Leer ot being capable of far more than you can explain is Polyani paradox. Arthur's definition is still valid today, ae many times the computer's learning ability amazes us. Games like Poker, Go, and complex computer games all are no more a challenge. Fortunately there is en h oe hitpssiwwuchistoryfa.n! 728 Astor of AL Can this be considered the firs (commercial) Machine-Learning software? The point where statistics becomes. machine learning is not exact, but for sure ML and A| is built on statistical concepts, hence we want to mention the first commercial sot packages. Before the 70's! Before the internet! SAS was developed at North Carolina State University from 1966 until 1976, when SAS Institute was incorporated. SAS was further in the * 1980s and 1990s with the addition of new statistical procedures, additional components and the introduction of JMP. SPSS released its first version in 1968 as the Statistical Package for the Social Sciences (SPSS) after being developed by Norman H. Nik and C. Hadlai Hull. “ee = hitpssiwwuchistoryfa.n! 229 raro2 Astor of AL Back Prépagating! Artificial Neural Networks can finally learn! eural Networks have a long history, ack to before 1950. Mostly os : {- oe jenting, and creating the mathematical foundation to make them work and learn. 1986 was the year David Rumelhart, Geoffrey Hinton, and Ronald Williams published their paperon Back propagation. Back propagation is the most popular way to make artificial neural networks learn, and will probably be for the next coming concepts for net med, the | pa . hitpssiwwuchistoryfa.n! 929 raro2 hitpssiwwuchistoryfa.n! Astor of AL similar learning al paper mentioned above is. pethaps once networks to scognize p de char with an extreme Soren fern Neural Networks (often ous andkey pointin. = Neural Networks (CNN), " called Deep Learning), contain much more layers, and come in many variations; VGG, Inception, ResNet, ResNeXt, DenseNet, the list is endless! They are all examples of st 2010 architectures. All have 2 things in common, they are trained he ick propagations, and in the ‘many layers you can find CNN layers. LSTM, a special kind of neural network which had the notion of a memory, was discovered about 10 years later by Sepp Hochreiter and Jiirgen Schmidhuber. Notice how all important foundations for modern Deep Learning were done before 2000. Yann LeCun and Yoshua Benjio, and Geoffrey Hinton Award in 2018, é as recognition f raro2 Astor of AL computer explore and min data itself, usually by crawling il subsets in databases. What will she find?! The step from a classical methods like linear regression, to data mining, can be seen as the first dimension 4 ast a explosion, the space thecomputer “ needed to explore and find — oe parameters for got bigger, and our a hardware and software could _ support the needed resources. Databases, CPUs, and software were ready in the 90's. Association rule learning, DBSCAN Clustering, and SVM are examples of algorithms invented in the 90's, and are often explained as common data mining techniques. Many moments from this point, we ‘A see ajump in complexity ofthe = . m nd innovation in a to support the . We see the same evoluti i ~ hitpssiwwuchistoryfa.n! 129 riar02t We have difficulty to assign the term intelligence to a single cell organism, but for sure we call our own species intelligent, or at least the brighter few among us. Is data mining Machine Learning? !s Data Mining Al? READ LESS hitpssiwwuchistoryfa.n! Astor of AL 2000 The use of GPUs The first GPU (Graphics processing unit) (https://www.doc.ic.ac.uk/~jce317/histor machine-learning.html¥the- importance-of-gpus), which was widely promoted as a ‘GPU’ on the consumer market, was developed by Nvidia in 1999. This was the GeForce 256 and was used for vector calculations in video games, shooting with guns in a 3D space! This technology gave an enormous boost to the graphic processing performance and years later it was. discovered by scientists and 1229 nano History of AI engineers useful in sp calculations. et showed the value cesses ror is around 5%, jore ae in 2016. ie with GPUs, lke pretty much all its successors as well. GPUs are very valuable for machine learning. They can have around 200 times more processors per chip than ie From 2017, Nvidia started to | develop GPUs with tensor cores, 8. * ee aoe for machine learning. Who would ever predicted, that shooting in a 3D wees oer? computational simil ep. Learning, used in Al? Both are examples of letting a computer doing many relatively small matrix multiplications, Which one do you prefer? READ LESS i» hitpssiwwuchistoryfa.n! 1329 raro2 Astor of AL some a practical approach. _ In his paper, he described what are now standard methods at typical "Machine Learning’ courses. He is also the creator of one of the most popular Machine-Learning Algorithms, the Random Forest. Not many know, Leo Brame also was involved in founding the ideas for another popular Machine Learning Model, the XGBoost. The idea to improve a weak model, along the derivative of the loss function, by a adding a correcting ‘booster’ on top, was partly from Leo. If you want to pin point a place in history, where classical statistics and modern machine learning deviate, this paper could be that fs f point, back in 2001. ah READ Ess ey += - e 7 a = es hitpssiwwuchistoryfa.n! 1429 raro2 Astor of AL k, puter tion, Ut interpreting is far more difficult it seems at frstiee ita aoe. Watson computer (https://www.ibm.com/ibm/history/ibm1C competed on Jeopardy! against the. TV quiz shows two biggest all-time champions and won. This was the nore Al could win at a game jich required language skills. lenge was in the language: — questions on this show are full of subtlety, puns and wordplay. ~ Before this, computers were able to answer straight forward questions, but this combination of understanding language in it's full ea subtlety and to find the right answer from several options in a knowledge bank was too much asked, until IBM Watson. _ IBM Watson was a pot . computer sof ‘ developed by IBI h “en ‘i hitpssiwwuchistoryfa.n! 1529 raro2 Astor of AL we might be part of a classic hype. Itis difficult to exactly point where the Big Data hype peak was at its highest, we chose March 2014, when Horton works raised 100M Dollar in preparation to go public. ‘According to Hadoops co-founders, 3 = x Doug Cutting and Mike Cafarella, the bee genesis of Hadoop was the Google a File System paper that was i published in October 2003. This paper spawned another one from Google - "MapReduce: Simplified Data Processing on Large Clusters’. * Back in 2014, there were many Hadoop distributions/versions. Horton and Cloudera were considered competitors, and are now merged. IBM's Hadoop Distribution, Big Insights, is gone. =" _ 2014is also the year the Big Data po st in Utrecht, a yearly jere Hadoop vendors and i _™ hitpssiwwuchistoryfa.n! 1629 ‘nar2024 Astor of AL PB of data. Kids nowadays are spoiled with Spark, the early Hadoop adaptors had to write MapReduce() in java code! How hardcore! READ LESS Operationalize Al aif ML? h oe hitpsihewtistoryot 129 ‘nar2024 Astor of AL jot as ea! ! MLOps are methods to supp‘ smooth way to: create, deploy, and maintain ML ag aaupen r eran Sn ie e este to create, update, and — ee Sntciner von. on Geman Tools, libraries, and methods come and go, but the challenge to operationalize IT easy and robust will always stay, and Machine Learning is no exception. Thinking of pleasant ways to operational and Alis ‘ %. » ' ao still a growing toy probably ie te \r the next coming years. READ LESS 2016 - XGBoost and Kag hitpssiwwuchistoryfa.n! raro2 Astor of AL Kaggle, is the World Championship _ be for Data Science, you compete with other analysts on a clear task given a dataset. In 2017, it reached 1 Million users. Kaggle is competitive Data Science, and this is a different game than solving real world problems. For once, the dataset eee task is given to you. But for sure, XGBoost winning competition after competition, made people curious and apply it on real world problems as well. XGBoost ae became and still is a very popular model for Machine Learning, powerful and versatile. It belongs to the family of Boosters, where weaker models stacked on ‘ each other produce very sharpend results, as each model has the ability to mare the mistake of t! ey revious. i Eee ‘i hitpssiwwuchistoryfa.n! 1929 ‘nar2024 Astor of AL in Strongest Go layer with Al é Reenter we = say... ~ “Computers will never be able to... [INSERT CHALLENGE]’. Well, for many years Go was the example to cite as a task Al will probably not be able to do at Human expert level in the ming 10 years. The folks from _ Deepmind did it, d Neural as a key component for their Al, together with self learning capabilities. > a. Checkers, Chess, Go, Poker, Starcraft... All were seen impossible at some point in time for computers to master, for some reason. ~ . s If a expert is among the strongest of the world in a mind game, we call that person Intelligent. In 1997 DeepBlue defeated Garry Ka h Sa Y hitpssiwwuchistoryfa.n! 20129 rer024 Astor of AL READ LESS: 2016 When Things go wrong... In this year, a twitter Al bot was released online, and learned to talk with humans. Just after 16 hours, the bot needed to be shut-down. It was too good at learning from humans, and humans had far more fun teaching it bad manners, than having a serious conversation. This showed us 2 things: Humans can be mean. If Al learns from humans, Al can be mean. Our History hasn't always been pretty and fair, and if we are not careful, Al will simply be a human history amplifier. Cathy O'Neil observed this, and unfortunately not only theoretical Her book, weapons of Math Destruction, was published in 2016 as well. She explains the dangers of massively apply Machine-Learning in our society, with shameful examples of things already happened. hitpssiwwuchistoryfa.n! raro2 Astor of AL taken offline, it woman. Also, hackers shown self-driving cars can be fooled with stickers on the road, invisible for humans, making self driving cars take sudden turns. Another problem with Al, it becomes. 0 good mimicking human a behaviour, it can generate realistic human faces, and can even impersonate humans like Presidents, Soe ™ and celebrities. 7 Is that Linkedin Request a real person? Was the Photo perhaps a machine generate? Al can Fool us! The other way around is also true, Humans can fool Al. By changing a image invisible for a human, the image of a cute cat can become a truck inthe eyes of a Al model. So... Humans can fool Al,Alcanfool = { rage it is fair after all? i E See ‘i hitpssiwwuchistoryfa.n! 2229 ‘nar2024 Astor of AL uplayme — a" _ some Jazz? Se SI nn as ne o~ book about the dangers of technology, and people massively paid to put a personal assistant in their home. We certainly are a weird species! : te One of first type of Virtual Assistants teas ter Child (released in -~ i sai text only and could play games, check the weather and look things up. The conv capability was limited. The first modern virtual assistant was Siri, which was released in 201 i It was able to set the time, assist at text messages and dial somebody. Later on it became able to perform more complex tasks, like giving directions or advice on rant. h oe hitpsihewtistoryot riar02t Aisin of released. Virtual assistants became quite common to consumers, especially on mobile phones and home devices. READ LESS November 2019 Hey! Want to buy a real robot? The child's imagination of what a Robot was, became reality when in 2019 Boston Dynamics made its first commercially available Robot, spot! Different than a Robot arm doing repetitive task in a factory, Spot can self balance, avoid or climb over obstacles, and has some sense of her surroundings. READ LESS 2019 Automated Machine hitpssiwwuchistoryfa.n! 2129 ‘nar2024 Astor of AL Ina way, Automated Machine arnin« computer Promenade ape jomenon thi data, like ergs ‘intuition’, through data. - "TPOT is well known Open-Source Automated Machine Learning library. IBM has named hers ‘AutoAl (https://www.ibm.com/cloud/watson- ___ studio/autoal). Wiad . ee list ing... “—_ = ted Machine learning can return very sharp predictive models, given a fixed d: not (yet) do, is automatically combine relevant data from elsewhere, create a richer context so to speak. READ LESS ~ 2 h 2- A AMillinn hitpsiwwchistoryfa.n! 25129 ‘ar2024 Astor of AL Q: How many eyes does ag have? AA giraffe has two eyes Q: How many legs does a frog have? AA frog has four legs Q: Are there any animals with three legs? —- A: No, there are no animals with three legs. What you saw, was a conversation with the GPT-3 model (https://lacker.io/ai/2020/07/06/giving- gpt-3-a-turing-test. html). It can A generate text based on a single S sentence, do Q&A like above, translate and transform, actually quite a lot! It is Version 3, the latest and greatest, larger and stronger than all its predecessors. Version 3 4 is called ‘Scary Good!’ bysome. GPT-2is quite OK, and possible to ‘ run of ly fast computer. For you need a computer th: h ne ‘ih hitpssiwwuchistoryfa.n! 26129 ‘nar2024 Astor of AL The Future? & e start relationships with Al 4 é Soe like in the movie ill we start a i : in the Terminator? a Or, maybe both? Will we have Intelligence - computers that actually, think, learn, and act themselves - in our lifetime? Fortunately, Al and Machine Learning have already shown to be useful without being self-aware (whatever that may be?). And if you are afraid “ of wars, killing, and des our world: hum: pretty good at h a hitpssiwwuchistoryfa.n! 21129 raro2 Astor of AL “ee reno vess Did we miss your hero? Want to ask us something? Feel free to reach us; _ Robin van Tilburg (https://www.linkedin.com/i in-van-tilburg-4674213/) and Willem Hendriks (https://www.linkedin.com/it mhendriks/). We used Pexels for many images and the backgro ), and want to thank and mention the content creators Olia Danilevich, Guilherme Almeida, Suzy Hazelwood, _ freestocks.org, Jason Dolin, cottonbro, and Pressmaster, for creating; Guys behind Notebook (https://www.pexels.com/photo/two-men-looking-at-a- laptop-4974920/) ‘Smiling Lade (https://www.pexels.com/photo/womanr-sitting-and-smiling- ae! 1858175/) ~ Robot (https://www.pexels.com/photo/blue-and-yellow-robot-toy-4013157/) Container (https://www.pexels.com/photo/steel-container-on-container-dock- 122164/) Cliff (https://www.pexels.com/photo/geological-formation-2855113/) Rainbow Hands (https://www.pexels.com/photo/persons-hands-wi colors-3693901/) . * Clouds (video) (https://www, com/video/cumulus-cloud sky-3129769/) an eee GD hitpssiwwuchistoryfa.n! 20129 ‘nar2024 A History of A /) from Sava Lazic, ' ground-vi video: 01 m =" = hipeswinnhisory ofa

You might also like