Professional Documents
Culture Documents
Group Members:
Guide : Ms. Nitha K P Sruthi K S
Teslin Rose P V
Vini Sasidharan
Vishnu V U
Outline
Introduction Modules
Objectives Architecture
Methodology Status
➟ Air pollution is one of the great killers of our
time. Polluted air was responsible for 6.4
million deaths worldwide: 2.8 million from
Predicting
techniques ➟ The first step to an accurate air quality
forecast is an excellent weather forecast.
Data process.
Humidity
Feature Wind
Selection Pressure
Rainfall
SO2
NO2
O3
PM10
PM2.5
➟ Random Forest
Predictive
Models ➟ Deep Learning (Multilayer perceptron)
➟ CatBoost
The proposed best air pollution prediction model.
Random Forest
➟ Random forests or random decision forests are an ensemble learning method for classification,
regression.
➟ Random forest is a bagging technique. The trees in random forests are run in parallel. There is no
interaction between these trees while building the trees. It operates by constructing a multitude
of decision trees at training time and outputting the mean prediction (regression) of the
individual trees.
Error and Feature Importance
Deep Learning
➟ AQI is set as the dependent variable,and the rest of the features are set as the independent
variables.
➟ The data is converted into a matrix form,and is split into two independent samples, training
and testing, 70% and 30% respectively.
➟ The test and train samples are then normalised, using Z Normalization.
➟ The tuned model has two hidden layer, with ten and five neurons. The activation function
used in Rectified Linear Unit (relu).
Deep Learning
➟ The input layer has one neuron each for the independent variables,and the output layer has
one neuron for response.
➟ As we are using regression and the output variable is numeric, mse is used as the loss
function ,and rmsprop is used as the optimizer function and mae is used as the metric
function , to compile the model.
➟ To fit the model,the training data is used ,which is run over a 100 epochs,then to evaluate the
model ,the testing data is used. The test data is used for getting the prediction.
Neural Network Model
Prediction and Loss
eXtreme Gradient Boosting
➔ XGBoost is an optimized distributed gradient boosting library designed to be
highly efficient, flexible and portable.
➔ You can also extend your Shiny apps with CSS themes, htmlwidgets, and
JavaScript actions.
User Interface
Status
90% 80%
Completed
➟ Data Pre-processing
60%
➟ Random Forest
➟ Deep Learning(MLP)
XG
Ca
Us
er
tB
ooB
oo
Int
st
st
er
fa
ce
Thank you!
Any questions?