Professional Documents
Culture Documents
Artificial Neural Networks
Artificial Neural Networks
Neural networks are the backbone of AI. They are used in a variety
of applications, including image recognition, natural language
processing, and even self-driving cars. In short, if you’re interested
in AI and want to understand how it works, you need to understand
neural networks.
Let´s Begin!
7. Transformers
Now you may ask — what happens in the step 1 — How does neural
network actually train data? We put our data (usually called
something like “training set”) into our model. It performs some
magic that we will get to later (it includes applying some functions
on the data) and it produces a result. Result is usually in form of
numbers (let´s say one number, for simplicity). The result is then
compared with the target.
The result is usually called output, the data that we use are input.
CHAPTER 2: NEURONS
What happens in one neuron when they get our data? Suppose we
put the data about people´s weight and height into a program.
Neuron then first creates from the pair (height, weight) (which are
numbers, so the programming language we use can read them well)
another number multiplied by something called weight and added
to something called bias. (DISCLAIMER: The weight in neural
network is not the weight we mentioned as weight of people.)
What is weight and bias? And why do we multiply and add them?
f( Σ(Xi * Wi ) + bias )
https://deepai.org/machine-learning-glossary-and-terms/weight-
artificial-neural-network
CHAPTER 4: ACTIVATING FUNCTION
Now, some math comes to the scene. What we have seen with the
weights and biases is called linear function (just a function that
takes x and creates w*x + b.)
After that, the neuron has to do one more thing — apply activation
function on the result of applying linear function.
https://medium.com/technologymadeeasy/for-dummies-the-
introduction-to-neural-networks-we-all-need-c50f6012d5eb
You can imagine gradient descent like a process where you go for a
hike in mountains and in every place, you choose to move in the
steepest direction. “Gradient” is a vector (an arrow) pointing to the
steepest direction you can walk if you want to get to the highest
point. Of course, it is different in every place where you are located.
The reason for having more layers is to improve the accuracy of the
final result. The different layers are called hidden layers and they
represent the intermediary nodes.
It is not true that more layers mean better training. Having too
many hidden layers can cause overfitting, which is a problem that
we won´t devote time to in this text.
CHAPTER 7: BACKPROPAGATION
[1] https://victorzhou.com/blog/intro-to-neural-networks/)
[2] https://www.bbntimes.com/science/artificial-intelligence-
vs-machine-learning-vs-artificial-neural-networks-vs-deep-
learning
[3] https://cs.stackexchange.com/questions/53521/what-is-
difference-between-multilayer-perceptron-and-multilayer-
neural-network?
newreg=948ddd49cab64af09eb5d1c5ad6ed3ce
Useful resources
https://towardsdatascience.com/deep-learning-with-python-
neural-networks-complete-tutorial-6b53c0b06af0
3Blue1Brown (Youtube)
https://www.freecodecamp.org/news/neural-networks-for-
dummies-a-quick-intro-to-this-fascinating-field-
795b1705104a/
https://victorzhou.com/blog/intro-to-neural-networks/
https://www.youtube.com/watch?v=ov_RkIJptwE
https://playground.tensorflow.org/
#activation=tanh&batchSize=10&dataset=circle®Dataset=
reg-
plane&learningRate=0.03®ularizationRate=0&noise=0&n
etworkShape=4,2&seed=0.23935&showTestData=false&discr
etize=false&percTrainData=50&x=true&y=true&xTimesY=fal
se&xSquared=false&ySquared=false&cosX=false&sinX=false
&cosY=false&sinY=false&collectStats=false&problem=classifi
cation&initZero=false&hideText=false
http://colah.github.io/posts/2014-03-NN-Manifolds-
Topology/