Supervised and Unsupervised Learning .pptx

KerenEvangelineI 72 views 49 slides Sep 25, 2024
Slide 1
Slide 1 of 49
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25
Slide 26
26
Slide 27
27
Slide 28
28
Slide 29
29
Slide 30
30
Slide 31
31
Slide 32
32
Slide 33
33
Slide 34
34
Slide 35
35
Slide 36
36
Slide 37
37
Slide 38
38
Slide 39
39
Slide 40
40
Slide 41
41
Slide 42
42
Slide 43
43
Slide 44
44
Slide 45
45
Slide 46
46
Slide 47
47
Slide 48
48
Slide 49
49

About This Presentation

Supervised and Unsupervised Learning Algorithm


Slide Content

SUPERVISED AND UNSUPERVISED LEARNING TECHNIQUES -Ms. Keren Evangeline. I / AP(O.G) SRM Valliammai Engineering College Department of Medical Electronics Faculty Development Programme

LIST OF COMMON UNSUPERVISED ML MODELS K – Means Clustering Principal component analysis (Dimensionality Reduction) Hierarchial clustering

ADVANTAGES OF UNSUPERVISED ML MODELS

DISADVANTAGES OF UNSUPERVISED ML MODELS

DECISION TREE

[Size < 2000?] / \ Yes (<=2000) No (>2000) / \ [Bedrooms < 3?] [Lot Size < 5000 sq.ft ?] / \ / \ Yes (<=3) No (>3) Yes (<5000) No (>5000) / \ / \ [Year < 1990?] [Price: $400K] [Price: $450K] [Price:$550K] / \ Yes No / \ [Price: $300K] [Price: $350K] Continuous Variable Decision Tree

RANDOM FOREST

Introduction A random Forest Algorithm is a supervised machine learning Algorithm consisting decision trees.  The general method of random decision forests was first proposed by Ho in 1995. After that, It was developed by Leo Breiman in 2001. RANDOM FOREST 33

Decision Tree Random forest 34

Decision Tree – Important Terms Random forest 35

How Decision Tree Works It follows a tree-like model of decisions and their possible consequences.  The algorithm works by recursively splitting the data into subsets based on the most significant feature at each node of the tree. Random forest 36

How Decision Tree Works Random forest 37

Ensemble Learning      Ensemble learning creates a stronger model by aggregating the predictions of multiple weak models. Random Forest is an example of ensemble learning where each model is a decision tree. The idea behind it is- the wisdom of the crowd. The majority vote aggregation can have better accuracy than the individual models. Random  forest 38

Dataset Preparation     Bootstrap Aggregating (Bagging)    Creating a different training subset randomly from the original training dataset with replacement is called Bagging. With replacement refers to the bootstrap sample having duplicate elements. Reduces variance, helps to avoid overfitting.    Out Of Bag (OOB)    The out-of-bag dataset represents the remaining elements which were not in the bootstrap dataset. Used for cross validation or to evaluate the performance.     Random Subspace Method(Feature Bagging)    While building the tree, for splitting, a randomly selected subset of the features are used. So, the trees are more different having low correlation.  Random forest 39

Bagging at training time​ Random forest 40

Bagging at inference time Random forest 41

Random Subspace Method at training time Random forest 42

Random Subspace Method at inference time Random forest 43

Definition The Random Forest algorithm is an ensemble learning method consisting of many decision trees that are built by using bagging and feature bagging which helps to create an uncorrelated forest of trees whose combined prediction is more accurate than that of a single tree. For classification tasks, final prediction is done by taking majority votes and for regression tasks, average of all the individual trees.  Random forest 44

Random Forest Model Random forest 45

Why Use Random Forest Random forests are an effective tool in prediction. Forests give results competitive with boosting and adaptive bagging, yet do not progressively change the training set. Random inputs and random features produce good results in classification- less so in regression. For larger data sets, we can gain accuracy by combining random features with boosting. Random forest 46

Advantages Versatile uses Easy-to-understand hyperparameters Classifier doesn't overfit with enough trees Disadvantages Increased accuracy requires more trees More trees slow down model Can’t describe relationships within data Advantages and Disadvantages

Random  Forest Applications Detects reliable debtors and potential fraudsters in finance Verifies medicine components and patient data in healthcare Gauges whether customers will like products in e-commerce Random forest 48

Thank you Any Question? Random forest 49
Tags