SlideShare una empresa de Scribd logo
1 de 18
Statistical Learning
Statistical Approaches
•   Statistical Learning
•   Naïve Bayes
•   Instance-based Learning
•   Neural Networks
Example: Candy Bags
• Candy comes in two flavors: cherry () and lime ()
• Candy is wrapped, can’t tell which flavor until opened
• There are 5 kinds of bags of candy:
   –   H1= all cherry (P(c|h1) = 1, P(l|h1) = 0)
   –   H2= 75% cherry, 25% lime
   –   H3= 50% cherry, 50% lime
   –   H4= 25% cherry, 75% lime
   –   H5= 100% lime
• Given a new bag of candy, predict H
• Observations: D1, D2 , D3, …
Bayesian Learning
• Calculate the probability of each hypothesis, given the data,
  and make prediction weighted by this probability (i.e. use all
  the hypothesis, not just the single best)
                        P ( d|hi )P (hi )
           P(hi | d)          P ( d)
                                            P(d | hi )P(hi )
• Now, if we want to predict some unknown quantity X



               P( X | d)             P( X |hi )P(hi | d)
                                 i
Bayesian Learning cont.
• Calculating P(h|d)


               P(hi | d)    P(d | hi )P(hi )
                            likelihood prior
• Assume the observations are i.i.d.—independent
  and identically distributed


               P(d | hi )        P(dj | hi )
                             j
Example:

• Hypothesis Prior over h1, …, h5 is
  {0.1,0.2,0.4,0.2,0.1}
• Data:



• Q1: After seeing d1, what is P(hi|d1)?
• Q2: After seeing d1, what is P(d2= |d1)?
Making Statistical Inferences
• Bayesian –
   – predictions made using all hypothesis, weighted by their probabilities

• MAP – maximum a posteriori
   – uses the single most probable hypothesis to make prediction
   – often much easier than Bayesian; as we get more and more data,
     closer to Bayesian optimal

• ML – maximum likelihood
   – assume uniform prior over H
   – when
Naïve Bayes (20.2)
Naïve Bayes
• aka Idiot Bayes
• particularly simple BN
• makes overly strong independence
  assumptions
• but works surprisingly well in practice…
Bayesian Diagnosis
• suppose we want to make a diagnosis D and there are n
  possible mutually exclusive diagnosis d1, …, dn
• suppose there are m boolean symptoms, E1, …, Em


                                    P(di )P(e1 ,..., e m | di )
        P(di | e1 ,..., em )
                                         P(e1 ,..., e m )

            how do we make diagnosis?


        we need:
                         P ( di )        P(e1 ,..., em | di )
Naïve Bayes Assumption
• Assume each piece of evidence (symptom) is
  independent give the diagnosis
• then

                                           m
               P(e1 ,..., em | di )             P(ek | di )
                                          k 1



  what is the structure of the corresponding BN?
Naïve Bayes Example

• possible diagnosis: Allergy, Cold and OK
• possible symptoms: Sneeze, Cough and Fever

                      Well           Cold       Allergy

       P(d)           0.9            0.05        0.05

    P(sneeze|d)       0.1            0.9         0.9

    P(cough|d)        0.1            0.8         0.7

     P(fever|d)       0.01           0.7         0.4

     my symptoms are: sneeze & cough, what is
     the diagnosis?
Learning the Probabilities
• aka parameter estimation
• we need
  – P(di) – prior
  – P(ek|di) – conditional probability
• use training data to estimate
Maximum Likelihood Estimate (MLE)
• use frequencies in training set to estimate:


                                       ni
                          p(di )
                                       N

                                            nik
                       p(e k | di )
                                            ni
  where nx is shorthand for the counts of events in
  training set
what is:         P(Allergy)?
Example:            P(Sneeze| Allergy)?
                    P(Cough| Allergy)?

   D       Sneeze          Cough                   Fever
 Allergy    yes                no                   no
  Well      yes                no                   no
 Allergy    yes                no                  yes
 Allergy    yes                no                   no
  Cold      yes                yes                 yes
 Allergy    yes                no                   no
  Well      no                 no                   no
  Well      no                 no                   no
 Allergy    no                 no                   no
 Allergy    yes                no                   no
Laplace Estimate (smoothing)
• use smoothing to eliminate zeros:


                                    ni 1       many other
                       p(di )                  smoothing
                                    N n        schemes…


                                       nik 1
                    p(e k | di )
                                       ni 2
  where n is number of possible values for d
  and e is assumed to have 2 possible values
Comments
• Generally works well despite blanket
  assumption of independence
• Experiments show competitive with decision
  trees on some well known test sets (UCI)
• handles noisy data
Learning more complex Bayesian
             networks
• Two subproblems:
• learning structure: combinatorial search over
  space of networks
• learning parameters values: easy if all of the
  variables are observed in the training set;
  harder if there are ‘hidden variables’

Más contenido relacionado

La actualidad más candente

Statistical learning
Statistical learningStatistical learning
Statistical learning
Slideshare
 
Stuart russell and peter norvig artificial intelligence - a modern approach...
Stuart russell and peter norvig   artificial intelligence - a modern approach...Stuart russell and peter norvig   artificial intelligence - a modern approach...
Stuart russell and peter norvig artificial intelligence - a modern approach...
Lê Anh Đạt
 
Knowledge representation in AI
Knowledge representation in AIKnowledge representation in AI
Knowledge representation in AI
Vishal Singh
 
Issues in knowledge representation
Issues in knowledge representationIssues in knowledge representation
Issues in knowledge representation
Sravanthi Emani
 
Propositional And First-Order Logic
Propositional And First-Order LogicPropositional And First-Order Logic
Propositional And First-Order Logic
ankush_kumar
 

La actualidad más candente (20)

Bayes Classification
Bayes ClassificationBayes Classification
Bayes Classification
 
Natural Language Processing: Parsing
Natural Language Processing: ParsingNatural Language Processing: Parsing
Natural Language Processing: Parsing
 
Frames
FramesFrames
Frames
 
Compiler Design Lecture Notes
Compiler Design Lecture NotesCompiler Design Lecture Notes
Compiler Design Lecture Notes
 
AI: Planning and AI
AI: Planning and AIAI: Planning and AI
AI: Planning and AI
 
Semantic nets in artificial intelligence
Semantic nets in artificial intelligenceSemantic nets in artificial intelligence
Semantic nets in artificial intelligence
 
Specification-of-tokens
Specification-of-tokensSpecification-of-tokens
Specification-of-tokens
 
POST’s CORRESPONDENCE PROBLEM
POST’s CORRESPONDENCE PROBLEMPOST’s CORRESPONDENCE PROBLEM
POST’s CORRESPONDENCE PROBLEM
 
Reasoning in AI
Reasoning in AIReasoning in AI
Reasoning in AI
 
Statistical learning
Statistical learningStatistical learning
Statistical learning
 
Decision Tree Learning
Decision Tree LearningDecision Tree Learning
Decision Tree Learning
 
Artificial Intelligence Notes Unit 2
Artificial Intelligence Notes Unit 2Artificial Intelligence Notes Unit 2
Artificial Intelligence Notes Unit 2
 
Stuart russell and peter norvig artificial intelligence - a modern approach...
Stuart russell and peter norvig   artificial intelligence - a modern approach...Stuart russell and peter norvig   artificial intelligence - a modern approach...
Stuart russell and peter norvig artificial intelligence - a modern approach...
 
Unit IV UNCERTAINITY AND STATISTICAL REASONING in AI K.Sundar,AP/CSE,VEC
Unit IV UNCERTAINITY AND STATISTICAL REASONING in AI K.Sundar,AP/CSE,VECUnit IV UNCERTAINITY AND STATISTICAL REASONING in AI K.Sundar,AP/CSE,VEC
Unit IV UNCERTAINITY AND STATISTICAL REASONING in AI K.Sundar,AP/CSE,VEC
 
Introduction to Image Compression
Introduction to Image CompressionIntroduction to Image Compression
Introduction to Image Compression
 
Knowledge representation in AI
Knowledge representation in AIKnowledge representation in AI
Knowledge representation in AI
 
P, NP, NP-Complete, and NP-Hard
P, NP, NP-Complete, and NP-HardP, NP, NP-Complete, and NP-Hard
P, NP, NP-Complete, and NP-Hard
 
Problems, Problem spaces and Search
Problems, Problem spaces and SearchProblems, Problem spaces and Search
Problems, Problem spaces and Search
 
Issues in knowledge representation
Issues in knowledge representationIssues in knowledge representation
Issues in knowledge representation
 
Propositional And First-Order Logic
Propositional And First-Order LogicPropositional And First-Order Logic
Propositional And First-Order Logic
 

Similar a Statistical learning (20)

Introduction to Bayesian Methods
Introduction to Bayesian MethodsIntroduction to Bayesian Methods
Introduction to Bayesian Methods
 
Bayes 6
Bayes 6Bayes 6
Bayes 6
 
Induction
InductionInduction
Induction
 
NaiveBayes.ppt
NaiveBayes.pptNaiveBayes.ppt
NaiveBayes.ppt
 
NaiveBayes.ppt
NaiveBayes.pptNaiveBayes.ppt
NaiveBayes.ppt
 
NaiveBayes.ppt
NaiveBayes.pptNaiveBayes.ppt
NaiveBayes.ppt
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
Naïve bayes
Naïve bayesNaïve bayes
Naïve bayes
 
2주차
2주차2주차
2주차
 
Bayesian data analysis1
Bayesian data analysis1Bayesian data analysis1
Bayesian data analysis1
 
Tutorial 2 (mle + language models)
Tutorial 2 (mle + language models)Tutorial 2 (mle + language models)
Tutorial 2 (mle + language models)
 
A Gentle Introduction to Bayesian Nonparametrics
A Gentle Introduction to Bayesian NonparametricsA Gentle Introduction to Bayesian Nonparametrics
A Gentle Introduction to Bayesian Nonparametrics
 
M18 learning
M18 learningM18 learning
M18 learning
 
Uncertainty
UncertaintyUncertainty
Uncertainty
 
naive bayes example.pdf
naive bayes example.pdfnaive bayes example.pdf
naive bayes example.pdf
 

Más de Slideshare

Crystal report generation in visual studio 2010
Crystal report generation in visual studio 2010Crystal report generation in visual studio 2010
Crystal report generation in visual studio 2010
Slideshare
 
Report generation
Report generationReport generation
Report generation
Slideshare
 
Security in Relational model
Security in Relational modelSecurity in Relational model
Security in Relational model
Slideshare
 
Entity Relationship Model
Entity Relationship ModelEntity Relationship Model
Entity Relationship Model
Slideshare
 
Major issues in data mining
Major issues in data miningMajor issues in data mining
Major issues in data mining
Slideshare
 
Data preprocessing
Data preprocessingData preprocessing
Data preprocessing
Slideshare
 
What is in you
What is in youWhat is in you
What is in you
Slideshare
 
Propositional logic & inference
Propositional logic & inferencePropositional logic & inference
Propositional logic & inference
Slideshare
 
Logical reasoning 21.1.13
Logical reasoning 21.1.13Logical reasoning 21.1.13
Logical reasoning 21.1.13
Slideshare
 
Resolution(decision)
Resolution(decision)Resolution(decision)
Resolution(decision)
Slideshare
 
Reinforcement learning 7313
Reinforcement learning 7313Reinforcement learning 7313
Reinforcement learning 7313
Slideshare
 
Neural networks
Neural networksNeural networks
Neural networks
Slideshare
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
Slideshare
 
Neural networks
Neural networksNeural networks
Neural networks
Slideshare
 
Logical reasoning
Logical reasoning Logical reasoning
Logical reasoning
Slideshare
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
Slideshare
 
Input & output devices
Input & output devicesInput & output devices
Input & output devices
Slideshare
 

Más de Slideshare (20)

Crystal report generation in visual studio 2010
Crystal report generation in visual studio 2010Crystal report generation in visual studio 2010
Crystal report generation in visual studio 2010
 
Report generation
Report generationReport generation
Report generation
 
Trigger
TriggerTrigger
Trigger
 
Security in Relational model
Security in Relational modelSecurity in Relational model
Security in Relational model
 
Entity Relationship Model
Entity Relationship ModelEntity Relationship Model
Entity Relationship Model
 
OLAP
OLAPOLAP
OLAP
 
Major issues in data mining
Major issues in data miningMajor issues in data mining
Major issues in data mining
 
Data preprocessing
Data preprocessingData preprocessing
Data preprocessing
 
What is in you
What is in youWhat is in you
What is in you
 
Propositional logic & inference
Propositional logic & inferencePropositional logic & inference
Propositional logic & inference
 
Logical reasoning 21.1.13
Logical reasoning 21.1.13Logical reasoning 21.1.13
Logical reasoning 21.1.13
 
Logic agent
Logic agentLogic agent
Logic agent
 
Resolution(decision)
Resolution(decision)Resolution(decision)
Resolution(decision)
 
Reinforcement learning 7313
Reinforcement learning 7313Reinforcement learning 7313
Reinforcement learning 7313
 
Neural networks
Neural networksNeural networks
Neural networks
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
 
Neural networks
Neural networksNeural networks
Neural networks
 
Logical reasoning
Logical reasoning Logical reasoning
Logical reasoning
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
 
Input & output devices
Input & output devicesInput & output devices
Input & output devices
 

Último

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
heathfieldcps1
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 

Último (20)

Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-IIFood Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 

Statistical learning

  • 2. Statistical Approaches • Statistical Learning • Naïve Bayes • Instance-based Learning • Neural Networks
  • 3. Example: Candy Bags • Candy comes in two flavors: cherry () and lime () • Candy is wrapped, can’t tell which flavor until opened • There are 5 kinds of bags of candy: – H1= all cherry (P(c|h1) = 1, P(l|h1) = 0) – H2= 75% cherry, 25% lime – H3= 50% cherry, 50% lime – H4= 25% cherry, 75% lime – H5= 100% lime • Given a new bag of candy, predict H • Observations: D1, D2 , D3, …
  • 4. Bayesian Learning • Calculate the probability of each hypothesis, given the data, and make prediction weighted by this probability (i.e. use all the hypothesis, not just the single best) P ( d|hi )P (hi ) P(hi | d) P ( d) P(d | hi )P(hi ) • Now, if we want to predict some unknown quantity X P( X | d) P( X |hi )P(hi | d) i
  • 5. Bayesian Learning cont. • Calculating P(h|d) P(hi | d) P(d | hi )P(hi ) likelihood prior • Assume the observations are i.i.d.—independent and identically distributed P(d | hi ) P(dj | hi ) j
  • 6. Example: • Hypothesis Prior over h1, …, h5 is {0.1,0.2,0.4,0.2,0.1} • Data: • Q1: After seeing d1, what is P(hi|d1)? • Q2: After seeing d1, what is P(d2= |d1)?
  • 7. Making Statistical Inferences • Bayesian – – predictions made using all hypothesis, weighted by their probabilities • MAP – maximum a posteriori – uses the single most probable hypothesis to make prediction – often much easier than Bayesian; as we get more and more data, closer to Bayesian optimal • ML – maximum likelihood – assume uniform prior over H – when
  • 9. Naïve Bayes • aka Idiot Bayes • particularly simple BN • makes overly strong independence assumptions • but works surprisingly well in practice…
  • 10. Bayesian Diagnosis • suppose we want to make a diagnosis D and there are n possible mutually exclusive diagnosis d1, …, dn • suppose there are m boolean symptoms, E1, …, Em P(di )P(e1 ,..., e m | di ) P(di | e1 ,..., em ) P(e1 ,..., e m ) how do we make diagnosis? we need: P ( di ) P(e1 ,..., em | di )
  • 11. Naïve Bayes Assumption • Assume each piece of evidence (symptom) is independent give the diagnosis • then m P(e1 ,..., em | di ) P(ek | di ) k 1 what is the structure of the corresponding BN?
  • 12. Naïve Bayes Example • possible diagnosis: Allergy, Cold and OK • possible symptoms: Sneeze, Cough and Fever Well Cold Allergy P(d) 0.9 0.05 0.05 P(sneeze|d) 0.1 0.9 0.9 P(cough|d) 0.1 0.8 0.7 P(fever|d) 0.01 0.7 0.4 my symptoms are: sneeze & cough, what is the diagnosis?
  • 13. Learning the Probabilities • aka parameter estimation • we need – P(di) – prior – P(ek|di) – conditional probability • use training data to estimate
  • 14. Maximum Likelihood Estimate (MLE) • use frequencies in training set to estimate: ni p(di ) N nik p(e k | di ) ni where nx is shorthand for the counts of events in training set
  • 15. what is: P(Allergy)? Example: P(Sneeze| Allergy)? P(Cough| Allergy)? D Sneeze Cough Fever Allergy yes no no Well yes no no Allergy yes no yes Allergy yes no no Cold yes yes yes Allergy yes no no Well no no no Well no no no Allergy no no no Allergy yes no no
  • 16. Laplace Estimate (smoothing) • use smoothing to eliminate zeros: ni 1 many other p(di ) smoothing N n schemes… nik 1 p(e k | di ) ni 2 where n is number of possible values for d and e is assumed to have 2 possible values
  • 17. Comments • Generally works well despite blanket assumption of independence • Experiments show competitive with decision trees on some well known test sets (UCI) • handles noisy data
  • 18. Learning more complex Bayesian networks • Two subproblems: • learning structure: combinatorial search over space of networks • learning parameters values: easy if all of the variables are observed in the training set; harder if there are ‘hidden variables’