Slide 34
Slide 34 text
Gradient Boosted Trees
● Trains models sequentially and emphasis on minimizing loss function
Y = ax + b + e, special focus is on ‘e’ i.e., error term
● Learner iteratively fits new model so error term should be minimized
● It goes like this
Y = M(x) + error
Error = G(x) + error1
Error1 = P(x) + error2
Now on combining, Y = M(x) + G(x) + P(x) + error2
And model get appropriate weight for each learner then,
Y = aplha * M(x) + beta * G(x) + gamma * P(x) + error2