




已阅读5页,还剩26页未读, 继续免费阅读
版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
1,Artificial Intelligence: Bayesian Networks,2,Graphical Models,If no assumption of independence is made, then an exponential number of parameters must be estimated for sound probabilistic inference. No realistic amount of training data is sufficient to estimate so many parameters. If a blanket assumption of conditional independence is made, efficient training and inference is possible, but such a strong assumption is rarely warranted. Graphical models use directed or undirected graphs over a set of random variables to explicitly specify variable dependencies and allow for less restrictive independence assumptions while limiting the number of parameters that must be estimated. Bayesian Networks: Directed acyclic graphs that indicate causal structure. Markov Networks: Undirected graphs that capture general dependencies.,3,Bayesian Networks,Directed Acyclic Graph (DAG) Nodes are random variables Edges indicate causal influences,4,Conditional Probability Tables,Each node has a conditional probability table (CPT) that gives the probability of each of its values given every possible combination of values for its parents (conditioning case). Roots (sources) of the DAG that have no parents are given prior probabilities.,Burglary,Earthquake,Alarm,JohnCalls,MaryCalls,5,CPT Comments,Probability of false not given since rows must add to 1. Example requires 10 parameters rather than 251 = 31 for specifying the full joint distribution. Number of parameters in the CPT for a node is exponential in the number of parents (fan-in).,6,Joint Distributions for Bayes Nets,A Bayesian Network implicitly defines a joint distribution.,Example,Therefore an inefficient approach to inference is: 1) Compute the joint distribution using this equation. 2) Compute any desired conditional probability using the joint distribution.,7,Nave Bayes as a Bayes Net,Nave Bayes is a simple Bayes Net,Y,X1,X2,Xn,Priors P(Y) and conditionals P(Xi|Y) for Nave Bayes provide CPTs for the network.,8,Independencies in Bayes Nets,If removing a subset of nodes S from the network renders nodes Xi and Xj disconnected, then Xi and Xj are independent given S, i.e. P(Xi | Xj, S) = P(Xi | S) However, this is too strict a criteria for conditional independence since two nodes will still be considered independent if their simply exists some variable that depends on both. For example, Burglary and Earthquake should be considered independent since they both cause Alarm.,9,Independencies in Bayes Nets,If removing a subset of nodes S from the network renders nodes Xi and Xj disconnected, then Xi and Xj are independent given S, i.e. P(Xi | Xj, S) = P(Xi | S) However, this is too strict a criteria for conditional independence since two nodes will still be considered independent if their simply exists some variable that depends on both. For example, Burglary and Earthquake should be considered independent since they both cause Alarm.,P(Xi | Xj, S) = P(Xi | S) , is equivalent to P(Xi , Xj | S) = P(Xi | S) P(Xj | S) How to prove?,10,Independencies in Bayes Nets,If removing a subset of nodes S from the network renders nodes Xi and Xj disconnected, then Xi and Xj are independent given S, i.e. P(Xi | Xj, S) = P(Xi | S) However, this is too strict a criteria for conditional independence since two nodes will still be considered independent if their simply exists some variable that depends on both. For example, Burglary and Earthquake should be considered independent since they both cause Alarm.,11,Independencies in Bayes Nets (cont.),Unless we know something about a common effect of two “independent causes” or a descendent of a common effect, then they can be considered independent. For example, if we know nothing else, Earthquake and Burglary are independent. However, if we have information about a common effect (or descendent thereof) then the two “independent” causes become probabilistically linked since evidence for one cause can “explain away” the other. For example, if we know the alarm went off that someone called about the alarm, then it makes earthquake and burglary dependent since evidence for earthquake decreases belief in burglary. and vice versa.,12,Bayes Net Inference,Given known values for some evidence variables, determine the posterior probability of some query variables. Example: Given that John calls, what is the probability that there is a Burglary?,Burglary,Earthquake,Alarm,JohnCalls,MaryCalls,?,John calls 90% of the time there is an Alarm and the Alarm detects 94% of Burglaries so people generally think it should be fairly high. However, this ignores the prior probability of John calling.,13,Bayes Net Inference,Example: Given that John calls, what is the probability that there is a Burglary?,Burglary,Earthquake,Alarm,JohnCalls,MaryCalls,?,John also calls 5% of the time when there is no Alarm. So over 1,000 days we expect 1 Burglary and John will probably call. However, he will also call with a false report 50 times on average. So the call is about 50 times more likely a false report: P(Burglary | JohnCalls) 0.02,14,Bayes Net Inference,Example: Given that John calls, what is the probability that there is a Burglary?,Burglary,Earthquake,Alarm,JohnCalls,MaryCalls,?,Actual probability of Burglary is 0.016 since the alarm is not perfect (an Earthquake could have set it off or it could have gone off on its own). On the other side, even if there was not an alarm and John called incorrectly, there could have been an undetected Burglary anyway, but this is unlikely.,15,Types of Inference,16,Sample Inferences,Diagnostic (evidential, abductive): From effect to cause. P(Burglary | JohnCalls) = 0.016 P(Burglary | JohnCalls MaryCalls) = 0.29 P(Alarm | JohnCalls MaryCalls) = 0.76 P(Earthquake | JohnCalls MaryCalls) = 0.18 Causal (predictive): From cause to effect P(JohnCalls | Burglary) = 0.86 P(MaryCalls | Burglary) = 0.67 Intercausal (explaining away): Between causes of a common effect. P(Burglary | Alarm) = 0.376 P(Burglary | Alarm Earthquake) = 0.003 Mixed: Two or more of the above combined (diagnostic and causal) P(Alarm | JohnCalls Earthquake) = 0.03 (diagnostic and intercausal) P(Burglary | JohnCalls Earthquake) = 0.017,17,Sample Inferences,Diagnostic (evidential, abductive): From effect to cause. P(Burglary | JohnCalls) = 0.016 P(Burglary | JohnCalls MaryCalls) = 0.29 P(Alarm | JohnCalls MaryCalls) = 0.76 P(Earthquake | JohnCalls MaryCalls) = 0.18 Causal (predictive): From cause to effect P(JohnCalls | Burglary) = 0.86 P(MaryCalls | Burglary) = 0.67 Intercausal (explaining away): Between causes of a common effect. P(Burglary | Alarm) = 0.376 P(Burglary | Alarm Earthquake) = 0.003 Mixed: Two or more of the above combined (diagnostic and causal) P(Alarm | JohnCalls Earthquake) = 0.03 (diagnostic and intercausal) P(Burglary | JohnCalls Earthquake) = 0.017,Assignment: Calculate these results!,18,Probabilistic Inference in Humans,People are notoriously bad at doing correct probabilistic reasoning in certain cases. One problem is they tend to ignore the influence of the prior probability of a situation.,19,Monty Hall Problem,1,2,3,One Line Demo: /crypto/Monty/monty.html,20,Multiply Connected Networks,Networks with undirected loops, more than one directed path between some pair of nodes.,In general, inference in such networks is NP-hard. Some methods construct a polytree(s) from given network and perform inference on transformed graph.,21,Node Clustering,Eliminate all loops by merging nodes to create meganodes that have the cross-product of values of the merged nodes.,Number of values for merged node is exponential in the number of nodes merged. Still reasonably tractable for many network topologies requiring relatively little merging to eliminate loops.,22,Bayes Nets Applications,Medical diagnosis Pathfinder system outperforms leading experts in diagnosis of lymph-node disease. Microsoft applications Problem diagnosis: printer problems Recognizing user intents for HCI Text categorization and spam filtering Student modeling for intelligent tutoring systems.,23,Statistical Revolution,Across AI there has been a movement from logic-based approaches to approaches based on probability and statistics. Statistical natural language processing Statistical computer vision Statistical robot navigation Statistical learning Most approaches are feature-based and “propositional” and do not handle complex relational descriptions with multiple entities like those typically requiring predicate logic.,Structured (Multi-Relational) Data,In many domains, data consists of an unbounded number of entities with an arbitrary number of properties and relations between them. Social networks Biochemical compounds Web sites,25,Biochemical Data,Predicting mutagenicity Srinivasan et. al, 1995,Web-KB Dataset Slattery & Craven, 1998,Faculty,Grad Student,Research Project,Other,Collective Classification,Traditional learning methods assume that objects to be classified are independent (the first “i” in the i.i.d. assumption) In structured data, the class of an entity can be influenced by the classes of related entities. Need to assign classes to all objects simultaneously to produce the most probable globally-consistent interpretation.,Logical AI Paradigm,Represents knowledge and data in a binary symbolic logic such as FOPC. + Rich representation that handles arbitrary sets of objects, with properties, relations, quantifiers, etc. Unable to handle uncertain knowledge and probabilistic reasoning.,Pro
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 护士护理业务查房:提升技能与优化护理实践
- 四川应用技术职业学院《开发基础》2023-2024学年第二学期期末试卷
- 江苏省宜兴市丁蜀区重点名校2024-2025学年初三下学期3月10日周中测数学试题试卷含解析
- 云南师范大学《产品U界面设计》2023-2024学年第二学期期末试卷
- 浙江省龙游第二高级中学2025届高三一轮第五次阶段性过关物理试题试卷含解析
- 浙江宇翔职业技术学院《金融综合实训》2023-2024学年第二学期期末试卷
- 山西林业职业技术学院《冶金资源工程》2023-2024学年第二学期期末试卷
- 云南医药健康职业学院《寄生虫学及检验》2023-2024学年第二学期期末试卷
- 中国科学技术大学《运动技能学习与控制》2023-2024学年第二学期期末试卷
- 内江师范学院《工程软件应用》2023-2024学年第二学期期末试卷
- 【MOOC】新科学家英语:演讲与写作-哈尔滨工业大学 中国大学慕课MOOC答案
- 2024火力发电厂运煤设备抑尘技术规范第4部分:输送及转运设备抑尘
- 第一届山东省职业能力大赛济南市选拔赛制造团队挑战赛项目技术工作文件(含样题)
- 高中历史选择性必修第3册试卷
- 老干工作业务培训
- GB/T 44744-2024粮食储藏低温储粮技术规程
- 加工制作合同(储存罐)
- 2023-2024学年广东省深圳市宝安区八年级(下)期末英语试卷
- 双碳全景系列培训第一章碳达峰、碳中和
- DB11T 594.2-2014 地下管线非开挖铺设工程施工及验收技术规程第2部分 顶管施工
- DB11∕T 1832.17-2021 建筑工程施工工艺规程 第17部分:电气动力安装工程
评论
0/150
提交评论