sikethatsarightemail
105 views
14 slides
Feb 02, 2024
Slide 1 of 14
1
2
3
4
5
6
7
8
9
10
11
12
13
14
About This Presentation
Says about ensemble method
Size: 5.55 MB
Language: en
Added: Feb 02, 2024
Slides: 14 pages
Slide Content
EASWARI ENGINEERING COLLEGE (Autonomous) Ramapuram , Chennai – 600 089 BACHELOR OF ENGINEERING in COMPUTER SCIENCE AND ENGINEERING 191CSC701T – Data Science Group 1: Raghul V – 3106201040105 Prakash S - 310620104098 Pramadeish SM – 310620104099 Prithvi S - 310620104102 Joel Thomas Joe – 310620104064 Nithish S - 310620104094 Ensemble Methods
Agenda Introduction CATEGORIES OF ENSEMBLE METHODS Main Types of Ensemble Methods How These Types Work Advantages and Disadvantages of using Ensemble Methods 2
INTRODUCTION Ensemble methods 3 Ensemble learning helps improve machine learning results by combining several models. This approach allows the production of better predictive performance compared to a single model Ensemble methods are meta-algorithms that combine several machine learning techniques into one predictive model in order to decrease variance (bagging) , bias (boosting) , or improve predictions (stacking) . Why do we use Ensemble methods ?
Categories of Ensemble Methods Ensemble methods 4 Sequential ensemble techniques generate base learners in a sequence, e.g., Adaptive Boosting (AdaBoost). The sequential generation of base learners promotes the dependence between the base learners. The performance of the model is then improved by assigning higher weights to previously misrepresented learners. Parallel ensemble techniques , base learners are generated in a parallel format, e.g., random forest. Methods utilize the parallel generation of base learners to encourage independence between the base learners. The independence of base learners significantly reduces the error due to the application of averages.
Ensemble methods 5 Bagging Bagging, the short form for bootstrap aggregating, is mainly applied in classification and regression. It increases the accuracy of models through decision trees, which reduces variance to a large extent. The reduction of variance increases accuracy, eliminating overfitting, which is a challenge to many predictive models. Boosting Boosting is an ensemble technique that learns from previous predictor mistakes to make better predictions in the future. Technique combines several weak base learners to form one strong learner, thus significantly improving the predictability of models. Stacking Stacking, another ensemble method, is often referred to as stacked generalization This technique works by allowing a training algorithm to ensemble several other similar learning algorithm predictions. Main Types of Ensemble Methods
Ensemble methods 6 Consider dataset D. It has many rows and columns Consider models or base learners M(M1,M2,…,Mn) for dataset D For each model we provide dataset D’M,D’’M, Etc. Suppose we have n records we select sample of n records and provide a particular record to model 1 Similarly for next model we use row sampling with replacement. For example in model M1 if there is data (A,B) ,then for model M2(B,C) where B is repetitive After training is done we give new test data to predict. Now we consider this method in binary classifier model How Bagging Works?
Ensemble methods 7 Suppose we give new test data and made them to pass The models gives their values as 1 or 0 as we consider binary classifier In the given dataset by voting classifier the majority (1) is taken as O/P How Bagging Works?
Ensemble methods 8 Consider a dataset with records Consider models(M1,M2,..,Mn) or base learners Some data are passed to base learners or model once it is trained. After training we will pass records to base learners or model and see how particular model is performed. How Boosting Works? Dataset
Ensemble methods 9 The records are allowed to pass to model M1 and red colored 2 records are incorrectly classified, the next model will be created sequentially and only 2 records will be passed to next model M2 If M2 gives some wrong records then the error will be passed continuously to M3 This will go until we specify some strong learners. This boosting technique will make weak learners to strong learners. How Boosting Works? Dataset
Ensemble methods 10 It is use heterogeneous method (strong learner + weak learner) where other methods use Homogenous method (strong learner or weak learner) Meta model How stacking works in meta model? Let have 100 records to train data 80 % trained on these data will be used for Prediction on 20% data Here we use: Logistic regression SVM Neural Networks In this we take this group How Stacking Works?
Ensemble methods 11 We can take k fold approach in 75 % typically trained data, we can create k buckets. We can always create meta model on 1 bucket out of k bucket or k-1 bucket How Stacking Works?
Ensemble methods 12 Adv and DisAdv of using Ensemble Methods Advantages of Ensemble Methods Disadvantages of Ensemble Methods Improved Predictive Performance Increased Complexity Reduction of Overfitting Computationally Intensive Robustness to Noisy Data Longer Training Times Handles Different Data Types Difficulty in Interpretation Versatility in Model Selection Decreased Transparency Increased Generalization Possibility of Overfitting Flexibility in Model Combination Reduced Intuitiveness
References presentation title 13 https://corporatefinanceinstitute.com/resources/data-science/ensemble-methods/ https://machinelearningmastery.com/tour-of-ensemble-learning-algorithms/ https://en.wikipedia.org/wiki/Ensemble_learning https://www.analyticsvidhya.com/blog/2023/01/ensemble-learning-methods-bagging-boosting-and-stacking/