Professional Documents
Culture Documents
Assignment PTRN RCGN
Assignment PTRN RCGN
Ans 1
P(X≤−1)=P(Y≥2)P(X≤−1)=P(Y≥2)
We can compare their values using standard normal distributions:
XμX=+1σX2=4ZX=X−14–√2ZX+1=XYμY=−1σY2=?ZY=Y−
(−1)σYY=σYZY−1XYμX=+1μY=−1σX2=4σY2=?ZX=X−14ZY=Y−
(−1)σY2ZX+1=XY=σYZY−1
⟹P(2ZX+1≤−1)=P(σYZY−1≥2)⟹P(2ZX+1≤−1)=P(σYZY−1≥2)
⟹P(ZX≤−1)=P(ZY≥3σY)⟹P(ZX≤−1)=P(ZY≥3σY)
⟹−(−1)=3σY⟹−(−1)=3σY
⟹σY=3
Ans 2
Couldn’t find the answer
Ans 3
op die (0.3)
no op
U(0)
U(3)
Ans 4
: You should take the $2 insurance and bet on Belle. The expected utility is U($2.05) (see figure
1).
$2.50
win (0.7)
lose (0.3)
$2.05 1.00
Belle
win (0.1) $7
Take Jeb
$2.05 $1.60 lose (0.9)
insurance
$1
$2.05 Don’t take
insurance
win (0.7) $1
Belle
$0.40 $0.40 lose (0.3)
Jeb
$-1
win (0.1)
$0.1 $10
lose (0.9)
Assignment 3
Ans 1
Despite the fact that we are dealing with two-dimensional shape and its
description, our world is three-dimensional and the same objects, if seen
from different angles (or changing position/orientation in space), may form
very different 2D projections.
The ideal case would be to have a universal shape descriptor capable of
overcoming these changes -- to design projection-invariant descriptors.
Consider an object with planar faces and imagine how many very different
2D shapes may result from a given face if the position and 3D orientation of
this simple object changes with respect to an observer. In some special
cases, like circles which transform to ellipses, or planar polygons,
projectively invariant features (invariants} can be found.
Region identification
Region identification is necessary for region description. One of the many
methods for region identification is to label each region (or each boundary)
with a unique (integer) number; such identification is
called labeling or coloring, also connected component labeling.
Goal of segmentation was to achieve complete segmentation, now, the
regions must be labeled.
Label collision is a very common occurrence -- examples of image shapes
experiencing this are U-shaped objects, mirrored E objects, etc.
The equivalence table is a list of all label pairs present in an image; all
equivalent labels are replaced by a unique label in the second step.
The algorithm is basically the same in 4-connectivity and 8-connectivity, the
only difference being in the neighborhood mask shape.
Ans 2
Start Search direction: look to the left first and check neighbors in clockwise
direction
40 3 1
7
5
6
Ans 3
Ans 1
Ans 2
Linear Classifiers. Logistic regression. ...
Support vector machines. Least squares support vector machines.
Quadratic classifiers.
Kernel estimation. k-nearest neighbor.
Decision trees. Random forests.
Neural networks.
Learning vector quantization.
Ans 3
1. Linear Regression
If you want to start machine learning, Linear regression is the
best place to start. Linear Regression is a regression model,
meaning, it’ll take features and predict a continuous output, eg :
stock price,salary etc. Linear regression as the name says, finds
a linear curve solution to every problem.
Basic Theory :
Loss function :
Advantages :
Fast training.
Hyperparameters :
Assumptions for LR :
LR vs Decision Tree :
LR vs SVM :
SVM supports both linear and non-linear
solutions using kernel trick.
LR vs KNN :
LR vs Neural Networks :
2. Logistic Regression
Just like linear regression, Logistic regression is the right
algorithm to start with classification algorithms. Eventhough,
the name ‘Regression’ comes up, it is not a regression model,
but a classification model. It uses a logistic function to frame
binary output model. The output of the logistic regression will
be a probability (0≤x≤1), and can be used to predict the binary 0
or 1 as the output ( if x<0.5, output= 0, else output=1).
Basic Theory :
Loss function :
Advantages :
Disadvantages :
Hyperparameters :
Assumptions of LR :
3. K-nearest neighbors
K-nearest neighbors is a non-parametric method used for
classification and regression. It is one of the most easy ML
technique used. It is a lazy learning model, with local
approximation.
Basic Theory :
Loss function :
Advantages :
Disadvantages :
Hyperparameters :
Assumptions :
KNN vs SVM :
4. Decision Tree
Decision tree is a tree based algorithm used to solve regression
and classification problems. An inverted tree is framed which is
branched off from a homogeneous probability distributed root
node, to highly heterogeneous leaf nodes, for deriving the
output. Regression trees are used for dependent variable with
continuous values and classification trees are used for
dependent variable with discrete values.
Basic Theory :
Advantages :
Disadvantages :
Chances for overfitting the model if we keep on
building the tree to achieve high purity. decision
tree pruning can be used to solve this issue.
Prone to outliers.
Hyperparameters :
Ans 4
Assignment 5
Ans 1
Let’s get a closer look at what this means and how correlation
can be useful. There are three types of correlations:
There are multiple ways to deal with this problem. The easiest
way is to delete or eliminate one of the perfectly correlated
features. Another way is to use a dimension reduction algorithm
such as Principle Component Analysis (PCA).
Ans 2
The algorithm use the concepts of variance matrix,
covariance matrix, eigenvector and eigenvalues pairs
to perform PCA, providing a set of eigenvectors and its
respectively eigenvalues as a result.
Ans 3
If you take up a certain path then there is always a possibility
that you might have to compromise with some other parameter.
Machine Learning models are no different, considering the case
of k-Nearest Neighbor there has always been a problem which
has a huge impact over classifiers that rely on pairwise distance
and that problem is nothing but the “Curse of Dimensionality”.
By the end of this article you will be able to create your own k-
Nearest Neighbor Model and observe the impact of increasing
the dimension to fit a data set.
Ans 4
The bigger the dots are means that the confidence score is
higher for those points.