Introduction to probability machine learning by Kevin Murphy

KrishnaYoga2 102 views 25 slides Sep 21, 2024
Slide 1
Slide 1 of 25
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25

About This Presentation

introduction about probability theory in machine learning fields


Slide Content

Chapter 1 - Introduction Prepared by: Shier Nee, SAW Based on: Probabilistic Machine Learning by Kevin Murphy

2 What is Machine Learning? By Shakir Mohamed, research at Deep Mind.

3 AI Techniques

4 Supervised Learning - Classification  

5 Supervised Learning - Classification  

6 Empirical Risk Minimization Supervised Learning - Classification

7 Uncertainty  using conditional probability distribution Constraints:     Softmax function )   ))   Supervised Learning - Classification Supervised Learning - Classification Supervised Learning - Classification

8 Maximum Likelihood Estimation Supervised Learning - Classification Minimizing Negative Log Likelihood

Supervised Learning - Regression Output = Real-value  Quadratic loss

10 Uncertainty  Assume output distribution = Gaussian Supervised Learning - Regression

Supervised Learning - Regression Features Engineering Feature Extraction Automatically 1 Feature

Generalization

Unsupervised Learning

Unsupervised Learning - Clustering Goal: Partition the input into regions that contain “similar” points.

Unsupervised Learning - Clustering Linear Model Non-linear Model

Unsupervised Learning – Self-supervised Learning From https://medium.com/analytics-vidhya/self-supervised-representation-learning-in-computer-vision-part-2-8254aaee937c

Unsupervised Learning – Evaluation It is difficult to evaluate since no ground truth. Use learned unsupervised representation as features to a downstream supervised learning method If unsupervised model learn useful features, the prediction in supervised model will increase. Evaluate unsupervised model by reducing the number of labelled sample to get good performance.

Reinforcement Learning

Reinforcement Learning A system or agent has to learn how to interact with its environment. This can be encoded by means of a policy a = (x) , which specifies which action to take in response to each possible input x (derived from the environment state).

Common Small Image Datasets MNIST CIFAR EMNIST Fashion - MNIST

Common Large Image Datasets ImageNet This dataset spans 1000 object classes 1,281,167 training images, 50,000 validation images and 100,000 test images https://www.image-net.org/download.php

Natural Language Processing IMDB movie review Natural Language Processing (NLP) Classification

Natural Language Processing Natural Language Processing (NLP) Translation Canadian parliament (English-French pairs) the European Union ( Europarl ). Document summarization, Question answering

Discrete Input Data One hot encoding Feature Interaction 3 colors (say red, green and blue) one-hot(red) = [1; 0; 0], one-hot(green) = [0; 1; 0], one-hot(blue) = [0; 0; 1].

Text Data Bag of words TF-IDF: T erm Frequency — Inverse Document Frequency dropping punctuation, converting all words to lower case; dropping common but uninformative words – “and”, “the” ( stop word removal_ Replace word with their base form – “running”, “runs”  “run” (word stemming) DF i is the number of documents with term i https://towardsdatascience.com/tf-idf-for-document-ranking-from-scratch-in-python-on-real-world-dataset-796d339a4089
Tags