Professional Documents
Culture Documents
Aachen
Exercise 5: MVA training
B B
S S
• Gradient Boosting
• XGBoost: eXtreme Gradient Boosting
• XGBoost algorithm has recently been dominating applied machine learning. Why?
• Parallel Computing: when you run xgboost, by default, it would use all the cores of
your machine
• Regularization: a technique used to avoid overfitting
• Flexibility: supports user defined evaluation metrics
• Availability: Currently, it is available for programming languages such as R, Python,
Java, etc.
• We will use it to train electron MVA ID
Next slides are technical and will be explained
as we take you through the tutorial
Any questions?
Start the exercise by following instructions here
https://github.com/guitargeek/ElectronMVATutorial
Once you have ran the ntuplizing step, open the root tree and have a look what’s in there
At the end,
check
variable importance
More things to try
• Change hyper-parameters and see if performance improves. You can
change learning rate.
• Consult this: https://xgboost.readthedocs.io/en/latest/parameter.html