Professional Documents
Culture Documents
• Example
• C(3,3)*0.6*0.6*0.6+C(3,2)*0.6*0.6*0.4=0.648
• Random forest
• AdaBoost
• Gradient Boosting
• Draw n’ out of n data instances (n’ < n), usually with replacement
• Bootstrap aggregating
Majoring
voting/averaging Final model
• Random forest = bagging + randomized feature set
)
Set !" the weight of each instance based on
ensemble prediction
2
alpha
"
1 − -.. 0
−2
" "23
-.. = ∑0 10 ⋅ 4" 5" 60 , 80 0.00 0.25 0.50 0.75 1.00
err
" "
= is the normalization term such that ∑0 10 sums to one
• Prediction as a weighted sum of the base learners
"!# = %& '& (# + %* '* (# + ⋯ + %, ', (#
• Similar to AdaBoost
• Let + = *$ + *'
• + #0 = *$ #0 + *' #0 + ⋯
• #
Loss function: ! = ∑& '& − ) *&
$
• Algorithm
$
• Initially, ) 1+ = 2# 1+
• Gradient of ! to ) *+
• Update by gd until termination
• ,-
,. */
= '+ − ) *+ −1 = ) *+ − '+ condition is met:
34# 3
5!
) 1+ = ) 1+ −
5) *+
3 3
=) *+ + '+ − ) *+
= 2# *+ + ⋯ + 23 *+ + 234# *+
• For regression with square loss
• If ! = %& + ⋯ + %) , we
Output
• Ensemble to improve the base learners
• Random forest