You are on page 1of 81

Please read this disclaimer before proceeding:

This document is confidential and intended solely for the educational purpose of
RMK Group of Educational Institutions. If you have received this document
through email in error, please notify the system manager. This document
contains proprietary information and is intended only to the respective group /
learning community as intended. If you are not the addressee you should not
disseminate, distribute or copy through e-mail. Please notify the sender
immediately by e-mail if you have received this document by mistake and delete
this document from your system. If you are not the intended recipient you are
notified that disclosing, copying, distributing or taking any action in reliance on
the contents of this information is strictly prohibited.
22IT401 - ARTIFICIAL
INTELLIGENCE AND
MACHINE LEARNING

Department: Information Technology


Batch/Year: 2022 – 26 / II Year
Created by:
Dr. T. Mahalingam
Dr. S. Selvakanmani

Date: 01.02.2024
Table of Contents
SLIDE
S.NO. CONTENTS NO.

1 CONTENTS 5

2 COURSE OBJECTIVES 7

3 PRE REQUISITES (COURSE NAMES WITH CODE) 8

4 SYLLABUS (WITH SUBJECT CODE, NAME, LTPC DETAILS) 9

5 COURSE OUTCOMES (6) 13

6 CO- PO/PSO MAPPING 14

7 LECTURE PLAN –UNIT 3 15

8 ACTIVITY BASED LEARNING –UNIT 3 17

9 LECTURE NOTES – UNIT 3

10 LEARNING FROM EXAMPLES: FORMS OF LEARNING 20

11 SUPERVISED LEARNING, LEARNING DECISION TREES 21

12 EVALUATING AND CHOOSING THE BEST HYPOTHESIS 30

THE THEORY OF LEARNING, REGRESSION AND


13 36
CLASSIFICATION WITH LINEAR MODELS

14 ARTIFICIAL NEURAL NETWORKS 47

15 APPLICATIONS: HUMAN COMPUTER INTERACTION (HCI), 52


KNOWLEDGE MANAGEMENT TECHNOLOGIES

16 AI FOR CUSTOMER RELATIONSHIP MANAGEMENT 55

17 EXPERT SYSTEMS 57
Table of Contents

SLIDE
S.NO. CONTENTS
NO.

18 DATA MINING, TEXT MINING, AND WEB MINING 62

19 OTHER CURRENT TOPICS LIKE NLP, DEEP LEARNING, ETC… 64

20 ASSIGNMENT 3 - UNIT 3 66

21 PART A Q & A 68

22 PART B Q s 72

23 SUPPORTIVE ONLINE CERTIFICATION COURSES 73

REAL TIME APPLICATIONS IN DAY TO DAY LIFE AND TO


24 74
INDUSTRY

25 CONTENTS BEYOND THE SYLLABUS 77

26 ASSESSMENT SCHEDULE 78

27 PRESCRIBED TEXT BOOKS & REFERENCE BOOKS 79

28 MINI PROJECT SUGGESTIONS 80


2. COURSE OBJECTIVES

Understand the concept of agents, problem solving and


searching strategies.

Familiarize with Knowledge reasoning and representation


based AI systems and approaches.

Apply the aspect of Probabilistic approach to AI.

Understanding of concepts of machine learning approaches.

Recognize the concepts of Machine Learning and its


deterministic tools
3. PRE REQUISITES

PRE-REQUISITE CHART
Artificial INTELLIGENCE
Intelligence
LEARNING

22MA401- Probability
Machine Learning

and Statistics

22CS303- Design and


ARTIFICIAL
andMACHINE

Analysis of Algorithms

22CS102- Problem
22IT401–
AND

Solving using C++


20IT602
4. 22IT401 ARTIFICIAL INTELLIGENCE AND MACHINE
LEARNING

OBJECTIVES
• Understand the concept of Artificial Intelligence
• Familiarize with Knowledge based AI systems and approaches
• Apply the aspect of Probabilistic approach to AI
• Identify the Neural Networks and NLP in designing AI models
• Recognize the concepts of Machine Learning and its deterministic tools
UNIT 1 PROBLEM SOLVING AND SEARCH STARTEGIES
Introduction: What Is Ai, The Foundations Of Artificial Intelligence, The
History Of Artificial Intelligence, The State Of The Art. Intelligent Agents: Agents And
Environments, Good Behaviour: The Concept Of Rationality, The Nature Of
Environments, And The Structure Of Agents. Solving Problems By Searching: Problem-
Solving Agents, Uninformed Search Strategies, Informed (Heuristic) Search Strategies,
Heuristic Functions. Beyond Classical Search: Local Search Algorithms and Optimization
Problems, Searching With Nondeterministic Actions And Partial Observations, Online
Search Agents And Unknown Environments. Constraint Satisfaction Problems:
Definition, Constraint Propagation, Backtracking Search, Local Search, The Structure Of
Problems.
List of Exercise/Experiments
1. Implementation of uninformed search algorithm (BFS and DFS).
2. Implementation of Informed Search algorithm (A* and Hill
Climbing Algorithm)

UNIT 2 KNOWLEDGE REPRESENTATION AND REASONING


Logical Agents: Knowledge-Based Agents, Propositional Logic,
Propositional Theorem Proving, Effective Propositional Model Checking, Agents Based
on Propositional Logic. FirstOrder Logic: Syntax and Semantics, Knowledge Engineering
in FOL, Inference in First-Order Logic, Unification and Lifting, Forward Chaining,
Backward Chaining, Planning: Definition, Algorithms, Planning Graphs, Hierarchical
Planning, Multi-agent Planning. Knowledge Representation: Ontological Engineering,
Categories and Objects, Events, Mental Events and Mental Objects, Reasoning Systems
for Categories, Reasoning with Default Information, The Internet Shopping World.
List of Exercise/Experiments
1. Implementation of forward and backward chaining.
2. Implementation of unification algorithms.

9
4. 22IT401 ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

UNIT 3 LEARNING
Learning from Examples: Forms of Learning, Supervised Learning, Learning
Decision Trees, Evaluating and Choosing the Best Hypothesis, The Theory of Learning,
Regression and Classification with Linear Models, Artificial Neural Networks. Applications:
Human computer interaction (HCI), Knowledge management technologies, AI for customer
relationship management, Expert systems, Data mining, text mining, and Web mining,
Other current topics.
List of Exercise/Experiments
1. Numpy Operations
2. NumPy arrays
3. NumPy Indexing and Selection
4. NumPy Exercise:
(i) Write code to create a 4x3 matrix with values ranging from 2 to 13.
(ii) Write code to replace the odd numbers by -1 in the following array.
(iii) Perform the following operations on an array of mobile phones prices 6999,
7500, 11999, 27899, 14999, 9999.
a) Create a 1d-array of mobile phones prices
b) Convert this array to float type
c) Append a new mobile having price of 13999 Rs. to this array
d) Reverse this array of mobile phones prices
e) Apply GST of 18% on mobile phones prices and update this array.
f) Sort the array in descending order of price
g) What is the average mobile phone price.

10
4. 22IT401 ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

UNIT 4 FUNDAMENTALS OF MACHINE LEARNING


Motivation for Machine Learning, Applications, Machine Learning,
Learning associations, Classification, Regression, The Origin of machine learning, Uses
and abuses of machine learning, Success cases, How do machines learn, Abstraction
and knowledge representation, Generalization, Factors to be considered, Assessing the
success of learning, Metrics for evaluation of classification method, Steps to apply
machine learning to data, Machine learning process, Input data and ML algorithm,
Classification of machine learning algorithms, General ML architecture, Group of
algorithms, Reinforcement learning, Supervised learning, Unsupervised learning, Semi-
Supervised learning, Algorithms, Ensemble learning, Matching data to an appropriate
algorithm.
List of Exercise/Experiments
1. Build linear regression models to predict housing prices using python ,
using data set available Google colabs.
2. Stock Ensemble-based Neural Network for Stock Market Prediction using
Historical Stock Data and Sentiment Analysis.
UNIT 5 MACHINE LEARNING AND TYPES
Supervised Learning, Regression, Linear regression, Multiple linear
regression, A multiple regression analysis, The analysis of variance for multiple
regression, Examples for multiple regression, Overfitting, Detecting overfit models:
Cross validation, Cross validation: The ideal procedure, Parameter estimation, Logistic
regression, Decision trees: Background, Decision trees, Decision trees for credit card
promotion, An algorithm for building decision trees, Attribute selection measure:
Information gain, Entropy, Decision Tree: Weekend example, Occam’s Razor,
Converting a tree to rules, Unsupervised learning, Semi Supervised learning,
Clustering, K – means clustering, Automated discovery, Reinforcement learning, Multi-
Armed Bandit algorithms, Influence diagrams, Risk modelling, Sensitivity analysis,
Casual learning.

11
4. 22IT401 ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

List of Exercise/Experiments
Use Cases
Case Study 1: Churn Analysis and Prediction (Survival Modelling)
Cox-proportional models
Churn Prediction
Case Study 2: Credit card Fraud Analysis
Imbalanced Data
Neural Network
Case study 3: Sentiment Analysis or Topic Mining from New York Times
Similarity measures (Cosine Similarity, Chi-Square, N Grams)
Part-of-Speech Tagging
Stemming and Chunking
Case Study 4: Sales Funnel Analysis
A/B testing
Campaign effectiveness, Web page layout effectiveness
Scoring and Ranking
Case Study 5: Recommendation Systems and Collaborative filtering
User based
Item Based
Singular value decomposition–based recommenders
Case Study 6: Customer Segmentation and Value
Segmentation Strategies
Lifetime Value
Case Study 7: Portfolio Risk Conformance
Risk Profiling
Portfolio Optimization
Case Study 8: Uber Alternative Routing
Graph Construction
Route Optimization

12
5.COURSE OUTCOME

Cognitive/
Expected
Affective
Course Level of
Course Outcome Statement Level of the
Code Attainmen
Course
t
Outcome
Course Outcome Statements in Cognitive Domain
Explain the problem solving Understand
C211.1 70%
and search strategies. K2
Demonstrate the techniques for
Apply
C211.2 knowledge representation and 70%
K3
reasoning.
Interpret various forms of Apply
C211.3 learning, artificial neural K3 70%
networks and its applications.

Experiment various machine Analyse


C211.4 70%
learning algorithms. K4

Employ AI and machine


Apply
C211.5 learning algorithms to solve 70%
K3
real world problems.

13
6.CO-PO/PSO MAPPING

Correlation Matrix of the Course Outcomes to


Programme Outcomes and Programme Specific
Outcomes Including Course Enrichment Activities

Programme Outcomes (POs), Programme Specific Outcomes (PSOs)


P P P
Course PO PO PO PO PO PO PO PSO PS PS
PO3 PO4 O O O
Outcome 1 2 5 9 10 11 12 1 O2 O3
6 7 8
s (COs)
K3/ A A A
K3 K4 K5 K5 A3 A3 A3 A2 K3 K3 K3
K5 2 3 3
C211
K2 2 1 1 1 3 3
.1
C211
K3 3 2 1 1 3 3
.2
C211
K3 3 2 1 1 3 2 3
.3
C211
K4 3 3 2 2 3 3
.4
C211
K3 3 2 1 1 3 2 3
.5

14
LECTURE PLAN – UNIT III
UNIT I INTRODUCTION

Sl.
No
NO ACTUAL
PROPOSED
OF LECTURE MODE OF
LECTURE PERTAININ TAXONOMY
TOPIC PER DELIVER
G CO(s) LEVEL
IOD Y
S PERIOD PERIOD

LEARNING FROM
1 EXAMPLES: FORMS OF 1 CO3 K2
LEARNING BB, ICT

SUPERVISED
LEARNING, LEARNING 1 CO3 K2 BB, ICT
DECISION TREES,
2

EVALUATING AND
3 CHOOSING THE BEST 1 CO3 K2 BB, ICT
HYPOTHESIS

THE THEORY OF
LEARNING,
REGRESSION AND 1 CO3 K3 BB, ICT
4 CLASSIFICATION WITH
LINEAR MODELS

ARTIFICIAL NEURAL
5 NETWORKS 1 CO3 K3 BB, ICT

APPLICATIONS: HUMAN
6 COMPUTER
INTERACTION (HCI),
KNOWLEDGE 1 CO3 K2 BB, ICT
MANAGEMENT
TECHNOLOGIES,

AI FOR CUSTOMER
RELATIONSHIP 1 CO3 K2 BB, ICT
7 MANAGEMENT,

8 EXPERT SYSTEMS 1 CO3 K2 BB, ICT

DATA MINING, TEXT


9 MINING, AND WEB 1 CO3 K2 BB, ICT
MINING
LECTURE PLAN – UNIT III

ASSESSMENT COMPONENTS MODE OF DELEIVERY


AC 1. Unit Test MD 1. Oral presentation
AC 2. Assignment MD 2. Tutorial
AC 3. Course Seminar MD 3. Seminar
AC 4. Course Quiz MD 4 Hands On
AC 5. Case Study MD 5. Videos
AC 6. Record Work MD 6. Field Visit
AC 7. Lab / Mini Project
AC 8. Lab Model Exam
AC 9. Project Review

15
Activity Based Learning - Crossword Puzzle
Activity Based Learning – Online Games

•ArtBot is a game for players of all ages which


teaches the basics of Artificial Intelligence and
Machine Learning.
•Your quest is to find and retrieve stolen art
objects.
•You train your AI helper to recognise and locate
the objects hidden in a maze of dungeons, and see
how supervised and reinforcement learning works.
•The game introduces players to core principles
and concepts of Artificial Intelligence and Machine
Learning.
•By setting the game in the context of cultural
heritage our aim was to address the application of
AI systems in multiple different areas, beyond
computing and programming, such as archaeology,
art, and transportation.
UNIT III

LEARNING

13
Lecture Notes - UNIT 3 LEARNING
Machine Learning (ML) is an automated learning with little or no human intervention. It
involves programming computers so that they learn from the available inputs. The main
purpose of machine learning is to explore and construct algorithms that can learn from the
previous data and make predictions on new input data.

The input to a learning algorithm is training data, representing experience, and


the output is any expertise, which usually takes the form of another algorithm that can
perform a task. The input data to a machine learning system can be numerical, textual,
audio, visual, or multimedia. The corresponding output data of the system can be a
floating-point number, for instance, the velocity of a rocket, an integer representing a
category or a class, for example, a pigeon or a sunflower from image recognition.

FORMS OF LEARNING

Concepts of Learning

Learning is the process of converting experience into expertise or knowledge.

Learning can be broadly classified into three categories, as mentioned below, based on
the nature of the learning data and interaction between the learner and the environment.

• Supervised Learning

• Unsupervised Learning

• Semi-supervised Learning

Similarly, there are four categories of machine learning algorithms as shown below −

• Supervised learning algorithm

• Unsupervised learning algorithm

• Semi-supervised learning algorithm

• Reinforcement learning algorithm

• However, the most commonly used ones are supervised and unsupervised learning.
Lecture Notes
Supervised Learning
Supervised learning is commonly used in real world applications, such as face and
speech recognition, products or movie recommendations, and sales forecasting.
Supervised learning can be further classified into two types -
Regression and Classification.

Regression trains on and predicts a continuous-valued response, for example


predicting real estate prices.

Classification attempts to find the appropriate class label, such as analyzing


positive/negative sentiment, male and female persons, benign and malignant tumors,
secure and unsecure loans etc.

In supervised learning, learning data comes with description, labels, targets or


desired outputs and the objective is to find a general rule that maps inputs to
outputs. This kind of learning data is called labeled data. The learned rule is then
used to label new data with unknown outputs.

Supervised learning involves building a machine learning model that is based


on labeled samples. For example, if we build a system to estimate the price of a
plot of land or a house based on various features, such as size, location, and so on,
we first need to create a database and label it. We need to teach the algorithm what
features correspond to what prices. Based on this data, the algorithm will learn how
to calculate the price of real estate using the values of the input features.

Supervised learning deals with learning a function from available training data. Here,
a learning algorithm analyzes the training data and produces a derived function that
can be used for mapping new examples. There are many supervised learning
algorithms such as Logistic Regression, Neural networks, Support Vector Machines
(SVMs), and Naive Bayes classifiers.

Common examples of supervised learning include classifying e-mails into spam and
not-spam categories, labeling webpages based on their content, and voice
recognition.
Lecture Notes

Unsupervised Learning
Unsupervised learning is used to detect anomalies, outliers, such as fraud or
defective equipment, or to group customers with similar behaviors for a sales
campaign. It is the opposite of supervised learning. There is no labeled data here.

When learning data contains only some indications without any description or
labels, it is up to the coder or to the algorithm to find the structure of the
underlying data, to discover hidden patterns, or to determine how to describe the
data. This kind of learning data is called unlabeled data.

Suppose that we have a number of data points, and we want to classify them into
several groups. We may not exactly know what the criteria of classification would
be. So, an unsupervised learning algorithm tries to classify the given dataset into
a certain number of groups in an optimum way.

Unsupervised learning algorithms are extremely powerful tools for analyzing data
and for identifying patterns and trends. They are most commonly used for
clustering similar input into logical groups. Unsupervised learning algorithms
include Kmeans, Random Forests, Hierarchical clustering and so on.

Semi-supervised Learning
If some learning samples are labeled, but some other are not labeled, then it is
semi-supervised learning. It makes use of a large amount of unlabeled data for
training and a small amount of labeled data for testing. Semi-supervised
learning is applied in cases where it is expensive to acquire a fully labeled dataset
while more practical to label a small subset. For example, it often requires skilled
experts to label certain remote sensing images, and lots of field experiments to
locate oil at a particular location, while acquiring unlabeled data is relatively easy.
Lecture Notes

Reinforcement Learning
Here learning data gives feedback so that the system adjusts to dynamic
conditions in order to achieve a certain objective. The system evaluates its
performance based on the feedback responses and reacts accordingly. The best
known instances include self-driving cars and chess master algorithm AlphaGo.

Purpose of Machine Learning


Machine learning can be seen as a branch of AI or Artificial Intelligence, since, the
ability to change experience into expertise or to detect patterns in complex data is
a mark of human or animal intelligence.

As a field of science, machine learning shares common concepts with other


disciplines such as statistics, information theory, game theory, and optimization.

As a subfield of information technology, its objective is to program machines so


that they will learn.

However, it is to be seen that, the purpose of machine learning is not building an


automated duplication of intelligent behavior, but using the power of computers to
complement and supplement human intelligence. For example, machine learning
programs can scan and process huge databases detecting patterns that are
beyond the scope of human perception.
Lecture Notes - Decision Trees

Supervised Classification: Decision Tree Learning


Example: learn concept PlayTennis (i.e., decide whether our friend
will play tennis or not in a given day)
Dataset for Play Tennis
Lecture Notes- Decision Trees
Introduction

Decision Tree learning is used to approximate discrete valued target functions, in


which the learned function is approximated by Decision Tree. To imagine, think of
decision tree as if or else rules where each if-else condition leads to certain
answer at the end. You might have seen many online games which asks several
question and lead to something that you would have thought at the end. A classic
famous example where decision tree is used is known as Play Tennis.

If the outlook is sunny and humidity is normal, then yes, you may play tennis.

Where should you use decision tree?

At any scenario where learning data has following traits:

• The learning data has attribute value pair like in the example shown above: Wind
as an attribute has two possible values – strong or weak

• Target function has discreet output. Here, the target function is – should you play
tennis? And the output to this discreet output – Yes and No

• The training data might be missing or have error


Lecture Notes- Decision Trees

ID3 Algorithm
Although there are various decision tree learning algorithms, we will explore the
Iterative Dichotomiser 3 or commonly known as ID3. ID3 was invented by Ross
Quinlan.

Before we deep down further, we will discuss some key concepts:

Entropy

Entropy is a measure of randomness. In other words, its a measure of


unpredictability. We will take a moment here to give entropy in case of binary
event(like the coin toss, where output can be either of the two events, head or
tail) a mathematical face:

Entropy = -(probability(a) * log2(probability(a))) – (probability(b) *


log2(probability(b)))

where probability(a) is probability of getting head and probability(b) is


probability of getting tail.

Of course this formulae can be generalised for n discreet outcome as follow:

Entropy = -p(1)*log2(p(1)) -p(2)*log2(p(2))-

p(3)*log2(p(3))………………………..p(n)*log(2p(n))

Entropy is an important concept. You can find more descriptive


explanation

Information Gain

In the decision tree shown here:


Lecture Notes- Decision Trees

We decided to break the first decision on the basis of outlook. We could have our
first decision based on humidity or wind but we chose outlook. Why?
Because making our decision on the basis of outlook reduced our randomness in the
outcome(which is whether to play or not), more than what it would have been
reduced in case of humidity or wind.
Let’s understand with the example here. Please refer to the play tennis dataset that
is pasted above.
We have data for 14 days. We have only two outcomes :
Either we played tennis or we didn’t play.
In the given 14 days, we played tennis on 9 occasions and we did not play on 5
occasions.
Probability of playing tennis:
Number of favourable events : 9
Number of total events : 14
Probability = (Number of favourable events) / (Number of total events)
= 9/14
= 0.642
Now, we will see probability of not playing tennis.
Probability of not playing tennis:
Number of favourable events : 5
Number of total events : 14

Probability = (Number of favourable events) / (Number of total events)


=5/14
=0.357
Lecture Notes- Decision Trees
And now entropy of outcome,

Entropy at source= -(probability(a) * log2(probability(a))) – (probability(b) *


log2(probability(b)))

= -(Probability of playing tennis) * log2(Probability of playing tennis) –


(Probability of not playing tennis) * log2(Probability of not playing tennis)

= -0.652 * log2(0.652) – 0.357*log2(0.357)

=0.940

So, entropy of whole system before we make our firest question is 0.940

Now, we have four features to make decision and they are:

1. Outlook

2. Temperature

3. Windy

4. Humidity

Let’s see what happens to entropy when we make our first decision on the basis
of Outlook.

Outlook
If we make a decison tree divison at this level 0 based on outlook, we have three
branches possible; either it will be Sunny or Overcast or it will be Raining.

1. Sunny : In the given data, 5 days were sunny. Among those 5 days, tennis was
played on 2 days and tenis was not played on 3 days. What is the entropy here?

Probablity of playing tennis = 2/5 = 0.4

Probablity of not playing tennis = 3/5 = 0.6

Entropy when sunny = -0.4 * log2(0.4) – 0.6 * log2(0.6)

= 0.97
2.Overcast: In the given data, 4 days were overcast and tennis was played on all the four
days. Let

Probablity of playing tennis = 4/4 = 1

Probablity of not playing tennis = 0/4 = 0

Entropy when overcast = 0.0

3.Rain: In the given data, 5 days were rainy. Among those 5 days, tennis was played on 3
days and tenis was not played on 2 days. What is the entropy here?
Probablity of not playing tennis = 2/5 = 0.4

Probablity of playing tennis = 3/5 = 0.6

Entropy when rainy = -0.4 * log2(0.4) – 0.6 * log2(0.6)

= 0.97

Entropy among the three branches:


Entropy among three branches = ((number of sunny days)/(total days) * (entropy when
sunny)) + ((number of overcast days)/(total days) * (entropy when overcast))
+ ((number of rainy days)/(total days) * (entropy when rainy))

= ((5/14) * 0.97) + ((4/14) * 0) + ((5/14) * 0.97)

= 0.69
What is the reduction is randomness due to choosing outlook as a decsion maker?

Reduction in randomness = entropy source – entropy of branches

= 0.940 – 0.69

= 0.246
This reduction in randomness is called Information Gain. Similar calculation can be
done for other features.

Temperature
Information Gain = 0.029

Windy
Information Gain = 0.048

Humidity
Information Gain = 0.152
We can see that decrease in randomness, or information gain is most for Outlook. So,
we choose first decision maker as Outlook.
Lecture Notes- Evaluating and Choosing the
Best Hypothesis
Bayesian Belief Network in artificial intelligence

Bayesian belief network is key computer technology for dealing with


probabilistic events and to solve a problem which has uncertainty. We can
define a Bayesian network as:

"A Bayesian network is a probabilistic graphical model which represents a


set of variables and their conditional dependencies using a directed acyclic
graph."

It is also called a Bayes network, belief network, decision network,


or Bayesian model.

Bayesian networks are probabilistic, because these networks are built from
a probability distribution, and also use probability theory for prediction
and anomaly detection.

Real world applications are probabilistic in nature, and to represent the


relationship between multiple events, we need a Bayesian network. It can
also be used in various tasks including prediction, anomaly detection,
diagnostics, automated insight, reasoning, time series prediction,
and decision making under uncertainty.

Bayesian Network can be used for building models from data and experts
opinions, and it consists of two parts:
1. Directed Acyclic Graph

2. Table of conditional probabilities.

The generalized form of Bayesian network that represents and solve


decision problems under uncertain knowledge is known as an Influence
diagram.
• Each node corresponds to the random variables, and a variable can
be continuous or discrete.

• Arc or directed arrows represent the causal relationship or


conditional probabilities between random variables. These directed
links or arrows connect the pair of nodes in the graph.
These links represent that one node directly influence the other node,
and if there is no directed link that means that nodes are independent
with each other
• In the above diagram, A, B, C, and D are random
variables represented by the nodes of the network
graph.
• If we are considering node B, which is connected with
node A by a directed arrow, then node A is called the
parent of Node B.
• Node C is independent of node A.

Bayesian network is based on Joint probability distribution and


conditional probability. So let's first understand the joint probability
distribution:
Joint probability distribution:
If we have variables x1, x2, x3,....., xn, then the probabilities of a
different combination of x1, x2, x3.. xn, are known as Joint
probability distribution.
P[x1, x2, x3,....., xn], it can be written as the following way in terms of
the joint probability distribution.
= P[x1| x2, x3,....., xn]P[x2, x3,....., xn]
= P[x1| x2, x3,....., xn]P[x2|x3,....., xn]....P[xn-1|xn]P[xn].
In general for each variable Xi, we can write the equation as:
P(Xi|Xi-1,........., X1) = P(Xi |Parents(Xi ))
Explanation of Bayesian network:
Let's understand the Bayesian network through an example by
creating a directed acyclic graph:
Example: Harry installed a new burglar alarm at his home to detect
burglary. The alarm reliably responds at detecting a burglary but also
responds for minor earthquakes. Harry has two neighbors David and
Sophia, who have taken a responsibility to inform Harry at work
when they hear the alarm. David always calls Harry when he hears
the alarm, but sometimes he got confused with the phone ringing
and calls at that time too. On the other hand, Sophia likes to listen to
high music, so sometimes she misses to hear the alarm. Here we
would like to compute the probability of Burglary Alarm.
Problem:
Calculate the probability that alarm has sounded, but there is
neither a burglary, nor an earthquake occurred, and David and
Sophia both called the Harry.
List of all events occurring in this network:
• Burglary (B)
• Earthquake(E)
• Alarm(A)
• David Calls(D)
• Sophia calls(S)
We can write the events of problem statement in the form of
probability: P[D, S, A, B, E], can rewrite the above probability
statement using joint probability distribution:
P[D, S, A, B, E]= P[D | S, A, B, E]. P[S, A, B, E]
=P[D | S, A, B, E]. P[S | A, B, E]. P[A, B, E]
= P [D| A]. P [ S| A, B, E]. P[ A, B, E]
= P[D | A]. P[ S | A]. P[A| B, E]. P[B, E]
= P[D | A ]. P[S | A]. P[A| B, E]. P[B |E]. P[E]
Let's take the observed probability for the Burglary and earthquake
component:

P(B= True) = 0.002, which is the probability of burglary.


P(B= False)= 0.998, which is the probability of no burglary.
P(E= True)= 0.001, which is the probability of a minor earthquake
P(E= False)= 0.999, Which is the probability that an earthquake not
occurred.
We can provide the conditional probabilities as per the below tables:
Conditional probability table for Alarm A:
The Conditional probability of Alarm A depends on Burglar and
earthquake:

B E P(A= True) P(A= False)

True True 0.94 0.06

True False 0.95 0.04

False True 0.31 0.69

False False 0.001 0.999

Conditional probability table for David Calls:


The Conditional probability of David that he will call depends on the
probability of Alarm.
A P(D= True) P(D= False)

True 0.91 0.09

False 0.05 0.95

Conditional probability table for Sophia Calls:


The Conditional probability of Sophia that she calls is depending on its Parent Node "Alarm."
From the formula of joint distribution, we can write the problem statement in the form of
probability distribution:

A P(S= True) P(S= False)

True 0.75 0.25

False 0.02 0.98


P(S, D, A, ¬B, ¬E) = P (S|A) *P (D|A)*P (A|¬B ^ ¬E) *P (¬B)
*P (¬E).
= 0.75* 0.91* 0.001* 0.998*0.999
= 0.00068045.
Hence, a Bayesian network can answer any query about the
domain by using Joint distribution.
The semantics of Bayesian Network:
There are two ways to understand the semantics of the Bayesian
network, which is given below:
1.To understand the network as the representation of the Joint
probability distribution.
It is helpful to understand how to construct the network.
2.To understand the network as an encoding of a collection of
conditional independence statements.
It is helpful in designing inference procedure.
The Theory of Learning, Regression and
Classification with Linear Models,

ML | Types of Learning – Supervised Learning

Let us discuss what is learning for a machine is as shown below media as follows:

A machine is said to be learning from past Experiences(data feed-in) with


respect to some class of tasks if its Performance in a given Task improves with
the Experience. For example, assume that a machine has to predict whether a
customer will buy a specific product let’s say “Antivirus” this year or not. The
machine will do it by looking at the previous knowledge/past experiences i.e
the data of products that the customer had bought every year and if he buys
Antivirus every year, then there is a high probability that the customer is going to
buy an antivirus this year as well. This is how machine learning works at the basic
conceptual level.
Supervised learning is when the model is getting trained on a labelled dataset.
A labelled dataset is one that has both input and output parameters. In this type of
learning both training and validation, datasets are labelled as shown in the figures
below.

Both the above figures have labelled data set as follows:

• Figure A: It is a dataset of a shopping store that is useful in predicting whether a


customer will purchase a particular product under consideration or not based on
his/ her gender, age, and salary.

Input: Gender, Age, Salary


Output: Purchased i.e. 0 or 1; 1 means yes the customer will purchase and 0
means that the customer won’t purchase it.

• Figure B: It is a Meteorological dataset that serves the purpose of predicting


wind speed based on different parameters.

Input: Dew Point, Temperature, Pressure, Relative Humidity, Wind Direction


Output: Wind Speed

Training the system: While training the model, data is usually split in the ratio of
80:20 i.e. 80% as training data and the rest as testing data. In training data, we
feed input as well as output for 80% of data. The model learns from training data
only. We use different machine learning algorithms(which we will discuss in detail in
the next articles) to build our model. Learning means that the model will build some
logic of its own.
Once the model is ready then it is good to be tested. At the time of testing, the input is
fed from the remaining 20% of data that the model has never seen before, the model
will predict some value and we will compare it with the actual output and calculate the
accuracy.

A.Classification: It is a Supervised Learning task where output is having defined


labels(discrete value). For example in above Figure A, Output – Purchased has defined
labels i.e. 0 or 1; 1 means the customer will purchase, and 0 means that the customer
won’t purchase. The goal here is to predict discrete values belonging to a particular
class and evaluate them on the basis of accuracy.
It can be either binary or multi-class classification. In binary classification, the model
predicts either 0 or 1; yes or no but in the case of multi-class classification, the model
predicts more than one class. Example: Gmail classifies mails in more than one class
like social, promotions, updates, and forums.

B. Regression: It is a Supervised Learning task where output is having continuous


value.

For example in above Figure B, Output – Wind Speed is not having any discrete value
but is continuous in a particular range. The goal here is to predict a value as much
closer to the actual output value as our model can and then evaluation is done by
calculating the error value. The smaller the error the greater the accuracy of our
regression model.
Example of Supervised Learning Algorithms:

• Linear Regression

• Logistic Regression

• Nearest Neighbor

• Gaussian Naive Bayes

• Decision Trees

• Support Vector Machine (SVM)

• Random Forest

ML | Types of Learning – Supervised Learning

• Supervised learning is a type of machine learning in which the algorithm is trained


on a labeled dataset, which means that the output (or target) variable is already
known. The goal of supervised learning is to learn a function that can accurately
predict the output variable based on the input variables. Supervised learning is
further divided into two main categories as follows:

1. Classification: In classification, the output variable is a categorical variable, and the


goal is to predict the class or category to which a new data point belongs. Examples
of classification problems include image classification, spam detection, and
sentiment analysis.

2. Regression: In regression, the output variable is a continuous variable, and the goal
is to predict the value of the output variable based on the input variables. Examples
of regression problems include predicting stock prices, weather forecasting, and
sales forecasting.
• Supervised learning algorithms are widely used in various
fields, such as natural language processing, computer vision,
medical diagnosis, speech recognition, and many others.
Some of the popular supervised learning algorithms include:
linear regression, logistic regression, decision trees, random
forest, k-nearest neighbors (KNN), support vector machine
(SVM), and neural

networks.

• It’s worth noting that supervised learning is useful when we


have a labeled data, and it’s not always the case. In some
scenarios, the data is not labeled or it’s too expensive to
label, then unsupervised learning, semi-supervised learning,
or self-supervised learning could be a better approach.
Regression and Classification | Supervised
Machine Learning

What is Regression and Classification in Machine Learning?

Data scientists use many different kinds of machine learning algorithms to


discover patterns in big data that lead to actionable insights. At a high level, these
different algorithms can be classified into two groups based on the way they
“learn” about data to make predictions: supervised and unsupervised learning.

Supervised Machine Learning: The majority of practical machine learning uses


supervised learning. Supervised learning is where you have input variables (x) and
an output variable (Y) and you use an algorithm to learn the mapping function from
the input to the output Y = f(X) . The goal is to approximate the mapping function
so well that when you have new input data (x) that you can predict the output
variables (Y) for that data.
Techniques of Supervised Machine Learning algorithms include linear and logistic
regression, multi-class classification, Decision Trees and support vector
machines. Supervised learning requires that the data used to train the algorithm is
already labelled with correct answers. For example, a classification algorithm will
learn to identify animals after being trained on a dataset of images that are properly
labelled with the species of the animal and some identifying characteristics.

Supervised learning problems can be further grouped


into Regression and Classification problems. Both problems have a goal of the
construction of a succinct model that can predict the value of the dependent
attribute from the attribute variables. The difference between the two tasks is the
fact that the dependent attribute is numerical for regression and categorical for
classification.
Regression
A regression problem is when the output variable is a real or continuous value, such
as “salary” or “weight”. Many different models can be used, the simplest is the linear
regression. It tries to fit data with the best hyper-plane which goes through the
points.

Types of Regression Models:

For Examples:
Which of the following is a regression task?

• Predicting age of a person

• Predicting nationality of a person

• Predicting whether stock price of a company will increase tomorrow

• Predicting whether a document is related to sighting of UFOs?

Solution : Predicting age of a person (because it is a real value, predicting


nationality is categorical, whether stock price will increase is discrete-yes/no
answer, predicting whether a document is related to UFO is again discrete- a
yes/no answer).
Let’s take an example of linear regression. We have a Housing data set and we
want to predict the price of the house. Following is the python code for it.
Classification
A classification problem is when the output variable is a category, such as “red” or
“blue” or “disease” and “no disease”. A classification model attempts to draw some
conclusion from observed values. Given one or more inputs a classification model will
try to predict the value of one or more outcomes.
For example, when filtering emails “spam” or “not spam”, when looking at transaction
data, “fraudulent”, or “authorized”. In short Classification either predicts categorical
class labels or classifies data (construct a model) based on the training set and the
values (class labels) in classifying attributes and uses it in classifying new data. There
are a number of classification models. Classification models include logistic regression,
decision tree, random forest, gradient-boosted tree, multilayer perceptron, one-vs-rest,
and Naive Bayes.

For example :
Which of the following is/are classification problem(s)?

• Predicting the gender of a person by his/her handwriting style

• Predicting house price based on area

• Predicting whether monsoon will be normal next year

• Predict the number of copies a music album will be sold next month

Solution : Predicting the gender of a person Predicting whether monsoon will be


normal next year. The other two are regression.
As we discussed classification with some examples. Now there is an example of
classification in which we are performing classification on the iris dataset
using RandomForestClassifier in python. You can download the dataset from Here
References:

• https://machinelearningmastery.com/logistic-regression-for-machine-learning/
• https://machinelearningmastery.com/linear-regression-for-machine-learning/
# Python code to illustrate
# classification using data set
#Importing the required library
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.ensemble import RandomForestClassifier
from sklearn.preprocessing import LabelEncoder
from sklearn.metrics import confusion_matrix
from sklearn.metrics import accuracy_score
from sklearn.metrics import classification_report
#Importing the dataset

dataset = pd.read_csv(
'https://archive.ics.uci.edu/ml/machine-learning-'+
'databases/iris/iris.data',sep= ',', header= None)
data = dataset.iloc[:, :]
#checking for null values
print("Sum of NULL values in each column. ")
print(data.isnull().sum())
#separating the predicting column from the whole dataset
X = data.iloc[:, :-1].values
y = dataset.iloc[:, 4].values
#Encoding the predicting variable
labelencoder_y = LabelEncoder()
y = labelencoder_y.fit_transform(y)
#Splitting the data into test and train dataset
X_train, X_test, y_train, y_test = train_test_split(
X, y, test_size = 0.3, random_state = 0)
#Using the random forest classifier for the prediction
classifier=RandomForestClassifier()
classifier=classifier.fit(X_train,y_train)
predicted=classifier.predict(X_test)
#printing the results
print ('Confusion Matrix :')
print(confusion_matrix(y_test, predicted))
print ('Accuracy Score :',accuracy_score(y_test, predicted))
print ('Report : ')
print (classification_report(y_test, predicted))
Artificial Neural Network
What is Artificial Neural Network?

The term "Artificial Neural Network" is derived from Biological neural


networks that develop the structure of a human brain. Similar to the human brain
that has neurons interconnected to one another, artificial neural networks also have
neurons that are interconnected to one another in various layers of the networks.
These neurons are known as nodes.

The given figure illustrates the typical diagram of Biological Neural Network.

The typical Artificial Neural Network looks something like the given figure.
Artificial Neural Network
Dendrites from Biological Neural Network represent inputs in Artificial Neural
Networks, cell nucleus represents Nodes, synapse represents Weights, and Axon
represents Output.

Relationship between Biological neural network and artificial neural network:

Biological Neural Network Artificial Neural Network

Dendrites Inputs

Cell nucleus Nodes

Synapse Weights

Axon Output
An Artificial Neural Network in the field of Artificial intelligence where it
attempts to mimic the network of neurons makes up a human brain so that
computers will have an option to understand things and make decisions in a
human-like manner. The artificial neural network is designed by programming
computers to behave simply like interconnected brain cells.

There are around 1000 billion neurons in the human brain. Each neuron has an
association point somewhere in the range of 1,000 and 100,000. In the human
brain, data is stored in such a manner as to be distributed, and we can extract
more than one piece of this data when necessary from our memory parallelly. We
can say that the human brain is made up of incredibly amazing parallel
processors.

We can understand the artificial neural network with an example, consider an


example of a digital logic gate that takes an input and gives an output. "OR" gate,
which takes two inputs. If one or both the inputs are "On," then we get "On" in
output. If both the inputs are "Off," then we get "Off" in output. Here the output
depends upon input. Our brain does not perform the same task. The outputs to
inputs relationship keep changing because of the neurons in our brain, which are
"learning."
Artificial Neural Network
The architecture of an artificial neural network:

Lets us look at various types of layers available in an artificial neural network.

Artificial Neural Network primarily consists of three layers:

Input Layer:

As the name suggests, it accepts inputs in several different formats provided by


the programmer.

Hidden Layer:

The hidden layer presents in-between input and output layers. It performs all the
calculations to find hidden features and patterns.

Output Layer:

The input goes through a series of transformations using the hidden layer, which
finally results in output that is conveyed using this layer.

The artificial neural network takes input and computes the weighted sum of the
inputs and includes a bias. This computation is represented in the form of a
transfer function.
Artificial Neural Network
It determines weighted total is passed as an input to an activation function to produce the
output. Activation functions choose whether a node should fire or not. Only those who are
fired make it to the output layer. There are distinctive activation functions available that can
be applied upon the sort of task we are performing.

Advantages of Artificial Neural Network (ANN)

1. Parallel processing capability

2. Storing data on the entire network

3. Capability to work with incomplete knowledge

4. Having a memory distribution

5. Having fault tolerance

Disadvantages of Artificial Neural Network:

1. Assurance of proper network structure:

2. Unecognized behavior of the network

3. Hardware dependence

4. Difficulty of showing the issue to the network

5. The duration of the network is unknown

How do artificial neural networks work?


Artificial Neural Network can be best represented as a weighted directed graph, where
the artificial neurons form the nodes. The association between the neurons outputs
and neuron inputs can be viewed as the directed edges with weights. The Artificial
Neural Network receives the input signal from the external source in the form of a
pattern and image in the form of a vector. These inputs are then mathematically
assigned by the notations x(n) for every n number of inputs.
Artificial Neural Network

Afterward, each of the input is multiplied by its corresponding weights ( these


weights are the details utilized by the artificial neural networks to solve a specific
problem ). In general terms, these weights normally represent the strength of the
interconnection between neurons inside the artificial neural network. All the
weighted inputs are summarized inside the computing unit.

If the weighted sum is equal to zero, then bias is added to make the output non-zero
or something else to scale up to the system's response. Bias has the same input, and
weight equals to 1. Here the total of weighted inputs can be in the range of 0 to
positive infinity. Here, to keep the response in the limits of the desired value, a
certain maximum value is benchmarked, and the total of weighted inputs is passed
through the activation function.

The activation function refers to the set of transfer functions used to achieve the
desired output. There is a different kind of the activation function, but primarily
either linear or non-linear sets of functions. Some of the commonly used sets of
activation functions are the Binary, linear, and Tan hyperbolic sigmoidal activation
functions. Let us take a look at each of them in details:

Binary:

In binary activation function, the output is either a one or a 0. Here, to accomplish this,
there is a threshold value set up. If the net weighted input of neurons is more than 1,
then the final output of the activation function is returned as one or else the output is
returned as 0.
Artificial Neural Network
Feed-Forward ANN:

A feed-forward network is a basic neural network comprising of an input layer, an


output layer, and at least one layer of a neuron. Through assessment of its output
by reviewing its input, the intensity of the network can be noticed based on group
behavior of the associated neurons, and the output is decided. The primary
advantage of this network is that it figures out how to evaluate and recognize input
patterns.

Applications :

• Human computer interaction (HCI)

• Knowledge management technologies

• AI for customer relationship management

• Expert systems,

• Data mining, text mining,

• Web mining.

Human computer interaction (HCI)

Human Computer Interface (HCI) was previously known as the man-machine


studies or man-machine interaction. It deals with the design, execution and
assessment of computer systems and related phenomenon that are for human use.

HCI can be used in all disciplines wherever there is a possibility of computer


installation. Some of the areas where HCI can be implemented with distinctive
importance are mentioned below −

• Computer Science − For application design and engineering.

• Psychology − For application of theories and analytical purpose.

• Sociology − For interaction between technology and organization.

• Industrial Design − For interactive products like mobile phones, microwave oven,
etc.
Human computer interaction (HCI)
What Is Human-Computer Interaction (HCI)?

HCI is the study of how people interact with computers, especially as it relates to
technology design. User-centered design, UI, and UX are combined with HCI to provide
intuitive technology and products.

HCI specialists consider how to develop and deploy computer systems that satisfy human
users. The majority of this research focuses on enhancing human-computer interaction by
enhancing how people utilize and comprehend an interface.

Components of Human-Computer Interaction

HCI is primarily composed of four essential elements:

The User

An individual or a group of individuals who work together on a project is referred to as


the user component. HCI researches the needs, objectives, and interaction styles of
users.

The Goal-Oriented Task

When using a computer, a user always has a purpose or aim in mind. To achieve this, the
computer presents a digital representation of things.

The Interface

An essential HCI element that can improve the quality of user interaction is the interface.
Many interface-related factors need to be taken into account, including the type of
interaction, screen resolution, display size, and even color contrast.

The Context

HCI is not only about providing better communication between users and computers but
also about factoring in the context and environment in which the system is accessed.

Examples of HCI

Let's examine some well-known HCI examples that have sped up the field's development.

1. IoT: pre-touch sensing and paper ID

2. Eye tracking: gaze detection

3. Speed recognition

4. Cloud Computing
Knowledge management technologies

New-age information solutions require the latest technology and one of the fast-
paced developments in customer service. With easy solutions that can be
powered by AI (Artificial Intelligence) tied up with knowledge management.

Right now, it is vital to give customers exactly what they want and quickly, so
adopting this technology will provide a positive CX and an excellent brand image.

Artificial Intelligence is the theory and development of computer systems able to


perform tasks typically requiring human Intelligence, such as translating between
languages, visual perception, decision making, and speech recognition.

Knowledge management is the process of creating, sharing, using, and managing


the knowledge and information of an organization.

What is knowledge management?

Knowledge management is the process of collecting, organizing, managing, and


disseminating knowledge across the organization.

As knowledge grows in organizations, the chances of data silos increase. With a


knowledge management process, the goal is to help users access relevant
information at the right time.

Organizations can democratize knowledge with a systematic knowledge


management strategy to improve employee efficiency while reducing operational
costs.

What does knowledge management help with?

Knowledge management lays the foundation for operations and communication


inside and outside an organization. It helps in assigning roles and responsibilities
while striking a balance in the delegation of authority — analytics through
knowledge management software help in ascertaining the loopholes and helps in
mending the same with organization-wide compliance. A workflow is set in motion
for employees to establish a standard code of conduct for operations.
Knowledge management technologies

1. Assigning responsibilities
The knowledge management process helps in assigning responsibilities to the
concerned parties. The designation of the employees is updated, and their
responsibilities are filled in.
2. Content retargeting
Retargeting content helps reduce the labor of the content and product team. A
piece of knowledge base, once created, can be adapted for any media and any
channel. This provides seamless circulation eliminating multiple works on the
same topic. Admins can bracket the period for knowledge to be scheduled or
published.
3. Analytics and feedback
An analytics dashboard made comparing set standards and achieved targets easy
and feasible. The effective knowledge management platform users are free to
rate and comment on it. This helps judge the practical credibility of users,
creators, and knowledge management solutions.

4. Cloud repository
The knowledge base created can be online, offline, cloud, or on-premise, making
accessibility extremely flexible, customizable, and secure. 24×7 accessibility to
authorized parties helps to reduce the time gap between the arrival of the
problem and the delivery of an apt solution.
5. Customizable IT knowledge base
A knowledge management system represents brand tonality to its users. These
users can be both employees and self-service users. A knowledge website or
application can be fully customized per the client’s needs to ensure easy usage
and a friendly UI. Multiple templates are available.
6. Omni-channel service
Over-whelming customers can backfire but supporting them by being present in
the exact manner they want you to be is nothing short of a miracle.
Online knowledge management has information and solutions in guides, charts,
images, videos, and chats.
7. Implement SOP
Adherence to Standard Operating Procedure is not killing creativity but kindling it
towards productive results. It helps establish a chain of command and syncs the
activities of employees on multiple fronts.
8. Better decision making
A good knowledge management tool has multiple options, from archiving
knowledge to acquiring facts, data, leads, feedback, communication, and
rating. DIY knowledge creation helps in user-generated content, thus being more
relevant as it is passed through expert magnification.
AI for customer relationship
management
Combining generative AI with CRM allows companies to automate business
processes better, develop more personalized communications, and provide
customers with the most helpful answers to questions.

At its core, CRM AI helps businesses better organize customer information and
access that information more easily. This includes contact details, demographic
data, communication history, purchase history, and other pertinent data used to
build sales opportunities and better serve customers.

With this intelligence, companies can create finely-tuned segmentation of leads


and contacts to nurture better relationships with prospects and customers.

Additionally, AI can help businesses respond faster to customer inquiries and


provide more informed answers to requests.

On the sales side, CRM with embedded AI gives users things like:

• Predictive lead scoring

• More accurate sales forecasts

• Recommendations

• Natural language search

• Data entry automation

• Automatically generated follow-up emails

Intelligent case routing may be a coming AI feature on the customer service side.

The goal of artificial intelligence in CRM is to let AI handle the analysis and make
intelligent recommendations about a customer or prospect based on all the data
about that person the system has collected.

With AI, a salesperson can open a contact record and ask the system for
suggestions on how to best connect with that person without spending time
sifting through company news and Twitter or LinkedIn profiles.
Examples of CRM With AI
Not surprisingly, some of the biggest names in CRM eagerly tout their artificial
intelligence functionality.

Companies like Salesforce, HubSpot, SugarCRM, Zoho, and Microsoft have all integrated
some form of AI into their platforms.

In 2023, these vendors are stepping up their games.

HubSpot
HubSpot has released an alpha version of ChatSpot, which combines GPT-4, HubSpot
CRM, DALLE·2, and Google Docs.

In this video, HubSpot co-founder Dharmesh Shah demonstrates AI use cases for sales,
reporting, and marketing.

Salesforce
Salesforce’s original AI platform was called Einstein. Salesforce billed Einstein as “a layer
of artificial intelligence that delivers predictions and recommendations based on your
unique business processes and customer data.”

On March 7, 2023, Salesforce announced Einstein GPT, which “delivers AI-


created content across every sales, marketing, commerce, and IT interaction.”

Einstein GPT can be integrated into many facets of the Salesforce platform, including
sales, service, marketing, commerce, and Tableau (analytics).

An impressive marketing use of Einstein GPT is its ability to design an event or campaign-
specific landing page, including text, image, and web form.

Microsoft
Also, in March 2023, Microsoft introduced Microsoft Dynamics 365 Copilot for CRM and
ERP.

Generative AI functionality will be available in Dynamics 365 Sales and Viva Sales,
Dynamics 365 Customer Service, Dynamics 365 Customer Insights, and Dynamics 365
Marketing.
What is an Expert System?
The expert systems are the computer applications developed to solve complex problems in
a particular domain, at the level of extra-ordinary human intelligence and expertise.

Characteristics of Expert Systems

• High performance

• Understandable

• Reliable

• Highly responsive

Capabilities of Expert Systems


The expert systems are capable of −

• Advising

• Instructing and assisting human in decision making

• Demonstrating

• Deriving a solution

• Diagnosing

• Explaining

• Interpreting input

• Predicting results

• Justifying the conclusion

• Suggesting alternative options to a problem

They are incapable of −


• Substituting human decision makers

• Possessing human capabilities

• Producing accurate output for inadequate knowledge base

• Refining their own knowledge


Components of Expert Systems
The components of ES include −
• Knowledge Base
• Inference Engine
• User Interface
Let us see them one by one briefly −

Knowledge Base
It contains domain-specific and high-quality knowledge.
Knowledge is required to exhibit intelligence. The success of any ES majorly depends
upon the collection of highly accurate and precise knowledge.
What is Knowledge?
The data is collection of facts. The information is organized as data and facts about the
task domain. Data, information, and past experience combined together are
termed as knowledge.
Components of Knowledge Base

The knowledge base of an ES is a store of both, factual and heuristic


knowledge.

• Factual Knowledge − It is the information widely accepted by the


Knowledge Engineers and scholars in the task domain.

• Heuristic Knowledge − It is about practice, accurate judgement,


one’s ability of evaluation, and guessing.

Knowledge representation

It is the method used to organize and formalize the knowledge in the


knowledge base. It is in the form of IF-THEN-ELSE rules.

Knowledge Acquisition

The success of any expert system majorly depends on the quality,


completeness, and accuracy of the information stored in the knowledge
base.

The knowledge base is formed by readings from various experts,


scholars, and the Knowledge Engineers. The knowledge engineer is a
person with the qualities of empathy, quick learning, and case analyzing
skills.

He acquires information from subject expert by recording, interviewing,


and observing him at work, etc. He then categorizes and organizes the
information in a meaningful way, in the form of IF-THEN-ELSE rules, to
be used by interference machine. The knowledge engineer also
monitors the development of the ES.
Inference Engine
Use of efficient procedures and rules by the Inference Engine is
essential in deducting a correct, flawless solution.
In case of knowledge-based ES, the Inference EEngine acquires and
manipulates the knowledge from the knowledge base to arrive at a
particular solution.
In case of rule based ES, it −
• Applies rules repeatedly to the facts, which are obtained from earlier
rule application.
• Adds new knowledge into the knowledge base if required.
• Resolves rules conflict when multiple rules are applicable to a
particular case.
To recommend a solution, the Inference Engine uses the following
strategies −
• Forward Chaining
• Backward Chaining

User Interface
User interface provides interaction between user of the ES and the ES
itself. It is generally Natural Language Processing so as to be used by
the user who is well-versed in the task domain. The user of the ES need
not be necessarily an expert in Artificial Intelligence.
It explains how the ES has arrived at a particular recommendation. The
explanation may appear in the following forms −
• Natural language displayed on screen.
• Verbal narrations in natural language.
• Listing of rule numbers displayed on the screen.
The user interface makes it easy to trace the credibility of the
deductions.
Expert Systems Limitations
No technology can offer easy and complete solution. Large systems are
costly, require significant development time, and computer resources.
ESs have their limitations which include −
• Limitations of the technology
• Difficult knowledge acquisition
• ES are difficult to maintain
• High development costs
Applications of Expert System
The following table shows where ES can be applied.

Application Description

Design Domain Camera lens design, automobile design.

Diagnosis Systems to deduce cause of disease


Medical Domain from observed data, conduction medical
operations on humans.

Comparing data continuously with observed


Monitoring Systems system or with prescribed behavior such as
leakage monitoring in long petroleum pipeline.

Controlling a physical process based on


Process Control Systems
monitoring.

Knowledge Domain Finding out faults in vehicles, computers.

Detection of possible fraud, suspicious


Finance/Commerce transactions, stock market trading, Airline
scheduling, cargo scheduling
Data Mining, Text Mining and Web Mining
Data Mining is a process of finding patterns and extracting useful
information from the pool of large data sets by transforming the data
with a bunch of business rules.
With the help of Data Mining procedures, Raw datasets are converted
into valuable datasets, which developers can further use to analyze
and determine the patterns.

Data Mining is an effective procedure for any organization as it helps


improve the marketing strategies and helps them target the customer
base based on the data.
With the help of structured data, it also allows you to study different
aspects of data and then get more innovative ideas to increase
productivity and sales.

The Data Mining process breaks down into the following steps –

1. Collect, Extract, Transform and Load the data into the data
warehouse
2. Store and manage the data in the database or on the cloud.
3. Provide access to data to the business analyst, management
teams, and Information Technology professionals.

Text Mining

Text Mining is a subset of Data Mining, and it involves the processing


of data from various text documents.
It is the process of transforming unstructured text into a structured
format and interpreting these data to identify patterns. In Text Mining,
various deep learning algorithms are used to evaluate the text and
generate useful information effectively.

The basic idea behind Text Mining is to find patterns in large datasets
that can be used for various purposes.

Text Mining requires both Sophistcated linguistic and statistical


techniques to analyze the unstructured text format data and provide
valuable insights. Text mining consists of a wide variety of methods
and technologies such as:
Keyword-based Technologies: Keyword-based technologies depend
on selecting keywords that input data contains and are then filtered as
a series of character strings.
Statistics Technologies: Statistical technology refers to the system
that is completely based on Machine learning. It uses certain text to
model the data and, in turn, uses the same model to manage and
categorize text.
Linguistic-based Technologies: Lingustinc based system uses a
Natual language processing system. The NLP models read the input text
and understand the structure of the text, grammar, logic, and context of
the text.

Web Mining : Web Mining is a process of extracting various useful


information readily available on the Internet (or World Wide Web). Web
Mining is a subset of Data Mining. It helps to analyze user activities on
different web pages and track them over a period of time to understand
customers’ behavior and surfing patterns. Web Mining is broadly
categorized into three main subcategories –

There are three main types of Web Data, as shown in the above image.
Let’s discuss in brief these Web Data types.
Web Content Data: The widespread form of data in Web Content are
HTML, web pages, images, etc. All these various data types constitute
Web Content data. The main layout for the Internet/Web content is
HTML, with a slight difference depending upon the use of the browser,
but the basic layout structure is the same everywhere.
Web Structure Data: On a typical web page, the contents are arranged
within HTML tags. The pages are hyperlinked, allowing users to navigate
back and forth to find relevant information. So basically, relationship/links
describing the connection between webpages is web structure data.
Web Usage Data: The main Data is generated by the Web Server and
Application Server on a typical web page. Web/Application server collects
the log data, including information about the users like their geographical
location, time, the content they interacted with, etc.
Natural Language Processing (NLP) - Introduction

Natural Language Processing (NLP) is a dynamic field of artificial intelligence that


focuses on the interaction between computers and human languages.
The primary goal of NLP is to enable machines to understand, interpret, and
generate human-like text.

Definition
NLP is a subfield of artificial intelligence that deals with the interaction between
computers and human languages. It encompasses a wide range of tasks, from basic
language understanding to advanced language generation.

Key Concepts
•Tokenization:Breaking down text into smaller units, such as words or phrases.
•Part-of-Speech Tagging:Assigning grammatical categories to words.
•Named Entity Recognition (NER):Identifying and classifying entities (e.g., names,
locations) in text.

2. Key Techniques in Natural Language Processing


Machine Translation - Translating text from one language to another, a
fundamental application of NLP.
Information Extraction - Identifying and extracting structured information from
unstructured text, often involving NER.
Sentiment Analysis - Determining the sentiment expressed in text, whether it's
positive, negative, or neutral.
Text Summarization - Condensing large volumes of text into concise summaries
while retaining essential information.

3. NLP Workflow
Text Preprocessing
Tokenization:Breaking down text into individual units.
Stopword Removal:Eliminating common words without significant meaning.
Stemming/Lemmatization:Reducing words to their base form.

Feature Extraction
Word Embeddings:Capturing semantic relationships between words.
TF-IDF (Term Frequency-Inverse Document Frequency): Assigning weights to words
based on their importance in a document.
Model Training and Evaluation
Supervised Learning: Training models on labeled datasets for tasks
like sentiment analysis or named entity recognition.
Unsupervised Learning: Extracting patterns and relationships from
unlabeled data, as seen in clustering or topic modeling.

4. Applications of Natural Language Processing

Virtual Assistants
NLP powers virtual assistants like Siri or Alexa, enabling them to
understand and respond to user queries.

Information Retrieval
Enhancing search engines to provide more relevant and accurate
results.

Healthcare Applications
Analyzing medical texts for information extraction, diagnosis, and
research.

5. Challenges and Considerations

Ambiguity in Language - The ambiguity of language poses


challenges in accurately interpreting meaning, especially in context.

Lack of Standardization - Differences in language use and


expressions can complicate NLP tasks, particularly across diverse
domains.

Conclusion
Natural Language Processing plays a pivotal role in bridging the gap
between human languages and computing systems. From machine
translation to sentiment analysis, NLP empowers various applications
that enhance our interaction with technology. While facing challenges
related to the complexity and diversity of language, continuous
advancements in NLP contribute to the development of more
intelligent and language-aware systems
ASSIGNMENT – UNIT III

Very Easy: (CO3, K1)


1.Explain the difference between regression and classification in supervised
learning. Provide an example scenario for each.
2.Describe the input layer in an artificial neural network.

Easy: (CO3, K2)


1.Provide a detailed explanation of the Play Tennis dataset mentioned in the
lecture notes. What are its attributes, and how is it used in decision tree
learning?
2.Explain the ID3 algorithm for decision tree learning. Also Give an example
of a classification problem.

Moderate: (CO3, K3)


1. Use decision tree to classify the students in a class based on their
academic performance.

2. Provide outline of the ID3 algorithm used for inducing decision tree from
the training tuples.

Hard: (CO3, K4)


1. Supervised Learning algorithms can be categorized into classification and
regression.

a) If your task is to build a model for COVID - 10 daily cases prediction, which
learning category would you use? Justify your answer.

b) Explain the differences and similiarities between classification and


regression algorithms.

1.Provide examples of applications for expert systems in different domains.


2.Critically analyze the application of decision trees in real-world scenarios,
considering their strengths and weaknesses

58
Very hard: (CO3, K4)

•On the network shown in Figure above, consider the event E =


"battery age is less than three years". Let A = P(E | "battery
dead"=true AND "no oil"=true) and B = P(E | "battery dead"=true
AND "no oil"=false). Which of the following three cases can
possibly be true: A > B, A = B, or A < B? Why?
•On the network shown in Figure 1, consider again the event E =
"battery age is less than three years". Let A = P(E | "battery
dead"=true AND "no oil"=true) and B = P(E | "battery dead"=true
AND "starter broken"=true). Which of the following three cases can
possibly be true: A > B, A = B, or A < B? Why?
•The "alternator broken" event and the "fanbelt broken" event are
both causes of the "no charging event." Let A = P("alternator
broken"=true | "no charging"=true) and B = P("alternator
broken"=true | "no charging"=true AND "fanbelt broken"=true).
Which of the following three cases do you expect to be true: A >
B, A = B, or A < B? Why?
•On the network shown in Figure 1, the "no gas" event is a cause
of the "car won't start" event. Let A = P("no gas"=true) and B =
P("no gas"=true | "car won't start"=true). Which of the following
three cases do you expect to be true: A > B, A = B, or A < B?
Why?
PART A- UNIT-3
1. What Is Learning?
Learning is the process of converting experience into expertise or knowledge.
Learning can be broadly classified into three categories, as mentioned below,
based on the nature of the learning data and interaction between the learner
and the environment.
*Supervised Learning
*Unsupervised Learning
*Semi-supervised Learning

2. What is Supervised Learning?


Supervised learning is commonly used in real world applications, such as
face and speech recognition, products or movie recommendations, and sales
forecasting. Supervised learning can be further classified in to two types -
Regression and Classification.

3. What is un-supervised learning?


Unsupervised learning is used to detect anomalies, outliers, such as
fraud or defective equipment, or to group customers with similar behaviours
for a sales campaign. It is the opposite of supervised learning. There is no
labelled data here.

4. What is reinforcement Learning?


Here learning data gives feedback so that the system adjusts to dynamic
conditions in order to achieve a certain objective. The system evaluates its
performance based on the feedback responses and reacts accordingly. The
best known instances include self-driving cars and chess master algorithm
AlphaGo.

5. What is Bayes Theory?


Bayes' theorem is a fundamental concept in probability theory and
statistics. It describes the probability of an event, based on prior knowledge
of conditions that might be related to the event. In machine learning, Bayes'
theorem is often used in Bayesian inference, particularly in Bayesian
classification algorithms like Naive Bayes.
PART A- UNIT-3
6. What is Bayesian Network?
A Bayesian network, also known as a belief network or causal probabilistic
network, is a probabilistic graphical model that represents a set of variables and
their conditional dependencies via a directed acyclic graph (DAG). It is used to
model uncertain relationships between variables and is commonly used in decision-
making, diagnosis, and prediction tasks.

7. What is CART?
CART is a machine learning algorithm used for both classification and
regression tasks. It works by recursively partitioning the input space into regions
and fitting simple models (e.g., decision trees) to each partition. CART builds binary
trees and splits the data into two subsets based on a threshold value for a given
feature at each node.

8. Define ANN?
An artificial neural network (ANN) is a computational model inspired by the
structure and function of biological neural networks. It consists of interconnected
nodes (neurons) organized in layers, including an input layer, one or more hidden
layers, and an output layer. ANN is used for tasks such as pattern recognition,
classification, regression, and optimization.

9. What is feed forward Technique?


In the context of artificial neural networks, feedforward refers to the flow of
data from input nodes through intermediate nodes (hidden layers) to output nodes
without any feedback loops. It is the basic architecture of most neural networks,
where information moves in only one direction, from input to output, without any
recurrent connections.

10. What is Expert system?


An expert system is a computer program that emulates the decision-making
ability of a human expert in a specific domain. It uses a knowledge base, which
contains expert knowledge encoded in the form of rules or heuristics, and an
inference engine, which reasons over the knowledge base to provide advice or
solutions to problems within the domain.

11. How CRM works in Al?


A Customer Relationship Management(CRM) system helps manage customer data.
AI-powered CRM systems leverage artificial intelligence techniques such as machine
learning and natural language processing to analyze customer data, automate tasks,
and provide personalized experiences. These systems can segment customers,
predict customer behaviour, automate email responses, and offer recommendations
to sales and marketing teams.
PART A- UNIT-3
12. Write down some application used in machine learning.
*Fraud detection in finance and cybersecurity
*Medical diagnosis and healthcare management
*Autonomous vehicles for object detection and navigation
*Speech recognition and language translation
*Financial trading and algorithmic trading strategies

13. What are the types of supervised learning?


*Classification (defined labels)
*Regression (no defined labels)

14. Define classification.


It is a Supervised Learning task where output is having defined labels
(discrete value). For example in above Figure A, Output – Purchased has defined
labels i.e. 0 or 1; 1 means the customer will purchase, and 0 means that the
customer won’t purchase. The goal here is to predict discrete values belonging to
a particular class and evaluate them on the basis of accuracy.

15. Define Regression.


It is a Supervised Learning task where output is having continuous value.
Machine Learning Regression is a technique for investigating the relationship
between independent variables or features and a dependent variable or outcome.
It's used as a method for predictive modelling in machine learning, in which an
algorithm is used to predict continuous outcomes.

16. Mention the types of regression models.

17. What is Artificial Neural Network?


The term "Artificial Neural Network" is derived from Biological neural
networks that develop the structure of a human brain. Similar to the human brain
that has neurons interconnected to one another; artificial neural networks also
have neurons that are interconnected to one another in various layers of the
networks. These neurons are known as nodes.
PART A- UNIT-3
18. What is the relationship between Biological neural network and
artificial neural network?

19. What is output Layer?


•The output layer is the final layer in the neural network where desired predictions
are obtained. There is one output layer in a neural network that produces the
desired final prediction. It has its own set of weights and biases that are applied
before the final output is derived.

20. Write the advantages of Artificial Neural Network (ANN)?


*Parallel processing capability
*Storing data on the entire network
*Capability to work with incomplete knowledge
*Having a memory distribution
*Having fault tolerance

21. Write the disadvantages of Artificial Neural Network (ANN).


Assurance of proper network structure:
Unrecognized behaviour of the network
Hardware dependence
Difficulty of showing the issue to the network
The duration of the network is unknown

22. What is feed forward ANN?


A feed-forward network is a basic neural network comprising of an input layer,
an output layer, and at least one layer of a neuron. Through assessment of its
output by reviewing its input, the intensity of the network can be noticed based on
group behaviour of the associated neurons, and the output is decided.

23.What is human computer interaction(HCI)?


HCI is the study of how people interact with computers, especially as it relates
to technology design. User-centered design, UI, and UX (User Experience) are
combined with HCI to provide intuitive technology and products.

24.What is knowledge management?


Knowledge management is the process of collecting, organizing, managing, and
disseminating knowledge across the organization.
PART B- UNIT-3
1. Explain different types of learning with example?

2. Explain Bayesian network with suitable example?

3. Define CART and Explain how it in details?

4. Explain Expert system in detail with neat diagram?

5. Define ANN, Explain the layers of ANN with example?

6. Write a short note on supervised, unsupervised and reinforcement


algorithm?

7. Explain any algorithm that evaluating and choosing best hypothesis?

8. Discuss in detail decision tree algorithm with example

9. What is neural network? Explain ANN and its application.

10. Write a short note on machine learning and its types? Explain how ML
works on CRM
SUPPORTIVE ONLINE COURSES – UNIT III

https://onlinecourses.nptel.ac.in/noc21_cs42/preview
An Introduction to Artificial Intelligence
By Prof. Mausam | IIT Delhi

https://www.coursera.org/learn/computational-thinking-problem-
solving

https://www.coursera.org/learn/artificial-intelligence-education-
for-teachers

https://www.coursera.org/specializations/ai-healthcare

https://www.coursera.org/learn/predictive-modeling-machine-
learning
https://www.drdobbs.com/parallel/the-practical-application-of-
prolog/184405220

61
REAL TIME APPLICATION- UNIT III
Neural Networks find extensive applications in areas where traditional computers
don’t fare too well. Like, for problem statements where instead of programmed
outputs, you’d like the system to learn, adapt, and change the results in sync with
the data you’re throwing at it. Neural networks also find
rigorous applications whenever we talk about dealing with noisy or incomplete data.
And honestly, most of the data present out there is indeed noisy.

With their brain-like ability to learn and adapt, Neural Networks form the entire
basis and have applications in Artificial Intelligence, and consequently, Machine
Learning algorithms. Before we get to how Neural Networks power Artificial
Intelligence, let’s first talk a bit about what exactly is Artificial Intelligence.

For the longest time possible, the word “intelligence” was just associated with the
human brain. But then, something happened! Scientists found a way of training
computers by following the methodology our brain uses. Thus came Artificial
Intelligence, which can essentially be defined as intelligence originating from
machines. To put it even more simply, Machine Learning is simply providing
machines with the ability to “think”, “learn”, and “adapt”.

With so much said and done, it’s imperative to understand what exactly are the use
cases of AI, and how Neural Networks help the cause. Let’s dive into
the applications of Neural Networks across various domains – from Social
Media and Online Shopping, to Personal Finance, and finally, to the smart assistant
on your phone.

You should remember that this list is in no way exhaustive, as the applications
of neural networks are widespread. Basically, anything that makes the machines
learn is deploying one or the other type of neural network.

62
Social Media
The ever-increasing data deluge surrounding social media gives the creators of these
platforms the unique opportunity to dabble with the unlimited data they have. No
wonder you get to see a new feature every fortnight. It’s only fair to say that all of this
would’ve been like a distant dream without Neural Networks to save the day.

Neural Networks and their learning algorithms find extensive applications in the world of
social media. Let’s see how:

Facebook
As soon as you upload any photo to Facebook, the service automatically highlights faces
and prompts friends to tag. How does it instantly identify which of your friends is in the
photo?
The answer is simple – Artificial Intelligence. In a video highlighting Facebook’s Artificial
Intelligence research, they discuss the applications of Neural Networks to power their
facial recognition software. Facebook is investing heavily in this area, not only within the
organization, but also through the acquisitions of facial-recognition startups
like Face.com (acquired in 2012 for a rumored $60M), Masquerade (acquired in 2016 for
an undisclosed sum), and Faciometrics (acquired in 2016 for an undisclosed sum).
In June 2016, Facebook announced a new Artificial Intelligence initiative that uses
various deep neural networks such as DeepText – an artificial intelligence engine
that can understand the textual content of thousands of posts per second, with
near-human accuracy.
Instagram
Instagram, acquired by Facebook back in 2012, uses deep learning by making use
of a connection of recurrent neural networks to identify the contextual meaning of
an emoji – which has been steadily replacing slangs (for instance, a laughing
emoji could replace “rofl”).
By algorithmically identifying the sentiments behind emojis, Instagram creates
and auto-suggests emojis and emoji related hashtags. This may seem like a
minor application of AI, but being able to interpret and analyze this emoji-to-text
translation at a larger scale sets the basis for further analysis on how people use
Instagram
Online Shopping
Do you find yourself in situations where you’re set to buy something, but you end
up buying a lot more than planned, thanks to some super-awesome
recommendations?
Yeah, blame neural networks for that. By making use of neural network and its
learnings, the e-commerce giants are creating Artificial Intelligence systems that
know you better than yourself. Let’s see how:
Search
Your Amazon searches (“earphones”, “pizza stone”, “laptop charger”, etc) return a
list of the most relevant products related to your search, without wasting much
time. In a description of its product search technology, Amazon states that
its algorithms learn automatically to combine multiple relevant features. It uses
past patterns and adapts to what is important for the customer in question.
And what makes the algorithms “learn”? You guessed it right – Neural Networks!
Recommendations
Amazon shows you recommendations using its “customers who viewed this item
also viewed”, “customers who bought this item also bought”, and also via curated
recommendations on your homepage, on the bottom of the item pages, and
through emails. Amazon makes use of Artificial Neural Networks to train its
algorithms to learn the pattern and behaviour of its users. This, in turn, helps
Amazon provide even better and customized recommendations.
CONTENT BEYOND SYLLABUS – UNIT III

1. Explain about MYCIN

MYCIN was an early backward chaining expert system that used


artificial intelligence to identify bacteria causing severe infections and
to recommend antibiotics, with the dosage adjusted for patient’s body
weight.
•MYCIN was developed over five or six years in the early 1970s at
Stanford University.
•It was written in Lisp
•MYCIN operated using fairly simple inference engine, and a
knowledge base of approximately 600 rules.
MYCIN is a computer program designed to provide attending
physicians with advice comparable to that which they would otherwise
get from a consulting physician specializing in bacteremia and
meningitis infections. To use MYCIN, the attending physician must sit in
front of a computer terminal that is connected to a DEC-20 (one of
Digital Equipment Corporation’s mainframe computers) where the
MYCIN program is stored. When the MYCIN program is evoked, it
initiates a dialogue. The physician types answers in response to various
questions. Eventually MYCIN provides a diagnosis and a detailed drug
therapy recommendation. The detailed structure of MYCIN is given
below:

65
ASSESSMENT SCHEDULE

Tentative schedule for the Assessment During 2023-2024 EVEN semester

Name of the
S.NO Start Date End Date Portion
Assessment

1 UNIT TEST 1 2.2.24 9.2.24 UNIT 1

2 IAT 1 12.2.24 17.2.24 UNIT 1 & 2

3 UNIT TEST 2 11.3.24 16.3.24 UNIT 3

4 IAT 2 1.4.24 6.4.24 UNIT 3 & 4

5 MODEL 20.4.24 30.4.24 ALL 5 UNITS

78
PRESCRIBED TEXT BOOKS AND REFERENCE BOOKS

1. Introduction to Artificial Intelligence and Machine Learning (IBM ICE


Publications).

2. Stuart Russell, Peter Norvig, “Artificial Intelligence: A Modern


Approach”, Third Edition, Pearson Education I Prentice Hall of India,
2010.

3. Elaine Rich and Kevin Knight, “Artificial Intelligence”, Third Edition,


Tata McGraw-Hill, 2010.

REFERENCES:

1. Patrick H. Winston. "Artificial Intelligence", Third edition, Pearson


Edition, 2006.

2. Dan W.Patterson, “Introduction to Artificial Intelligence and Expert


Systems”, PHI, 2006.

3. Nils J. Nilsson, “Artificial Intelligence: A new Synthesis”, Harcourt


Asia Pvt. Ltd., 2000.

67
Mini Projects
Air and water quality index and environment monitoring

A software that suggests drugs and formulations for a


disease/pharmacological property based on the Ayurvedic
classical books/Repositories.

Self-identifying the mental health status and get guidance for


support.

Student dropout analysis for school education

Fake Social Media Profile detection and reporting

Online integrated platform for projects taken up by the students


of various universities/colleges

Reporting and Monitoring Tool for Women's Safety in


Colleges/Universities
Thank you

Disclaimer:

This document is confidential and intended solely for the educational purpose of
RMK Group of Educational Institutions. If you have received this document through
email in error, please notify the system manager. This document contains proprietary
information and is intended only to the respective group / learning community as
intended. If you are not the addressee you should not disseminate, distribute or
copy through e-mail. Please notify the sender immediately by e-mail if you have
received this document by mistake and delete this document from your system. If
you are not the intended recipient you are notified that disclosing, copying,
distributing or taking any action in reliance on the contents of this information is
strictly prohibited.

69

You might also like