机器学习概论机器学习概论 (5).pdf
![资源得分’ title=](/images/score_1.gif)
![资源得分’ title=](/images/score_1.gif)
![资源得分’ title=](/images/score_1.gif)
![资源得分’ title=](/images/score_1.gif)
![资源得分’ title=](/images/score_05.gif)
《机器学习概论机器学习概论 (5).pdf》由会员分享,可在线阅读,更多相关《机器学习概论机器学习概论 (5).pdf(62页珍藏版)》请在淘文阁 - 分享文档赚钱的网站上搜索。
1、Welcome to Introduction to Machine Learning!2010.3.51*Images come from InternetCoffee TimeACMA.M.TuringAward 2018Announced on Mar.27,20193Yann LeCunGeoffrey E HintonYoshua Bengiohttps:/amturing.acm.orgA.M Turing Awards 20184“For conceptual and engineering breakthroughs that have made deep neural net
2、works a critical component of computing.”Bengio:Professor at the University of Montreal and Scientific Director at Mila,Quebecs Artificial Intelligence InstituteHinton:VP and Engineering Fellow of Google,Chief Scientific Adviser of The Vector Institute,and University Professor Emeritus at the Univer
3、sity of TorontoLeCun:Professor at New York University and VP and Chief AI Scientist at Facebook.https:/amturing.acm.orgComments to their contribution5Working independently and togetherHinton,LeCun and Bengio developed conceptual foundations for the field,Identified surprising phenomena through exper
4、imentsContributed engineering advances that demonstrated the practical advantages of deep neural networksIn recent years,deep learning methods have been responsible for astonishing breakthroughs in computer vision,speech recognition,natural language processing,and robotics among other applicationsht
5、tps:/amturing.acm.orgSelect Technical Select Technical Accomplishments:Accomplishments:Geoffrey HintonGeoffrey Hinton6Backpropagation:1986,“Learning Internal Representations by Error Propagation,”David Rumelhart and Ronald Williams,Hinton BP algo.allowed NNs to discover their own internal representa
6、tions of datamaking it possible to use NNs to solve problems that had previously been thought to be beyond their reach.The BP algo.is standard in most neural networks today.Boltzmann Machines:1983,Terrence Sejnowski and Hinton One of the first NNs capable of learning internal representations in neur
7、ons that were not part of the input or output.Improvements to convolutional neural networks:2012,with his students,Alex Krizhevsky and Ilya Sutskever,and Hinton Improved CNN using rectified linear neurons and dropout regularization.In ImageNet competition,Hinton and his students(AlexNet)almost halve
8、d the error ratefor object recognition and reshaped the computer vision field.Select Technical AccomplishmentsSelect Technical Accomplishments:YoshuaYoshua BengioBengio7Probabilistic models of sequences:In the 1990s,Bengio combined NNs with probabilistic models of sequences,such as HMM.A system used
9、 by AT&T/NCR for reading handwritten checks,a pinnacle of NN research in the 1990sModern DL speech recognition sys.are extending these concepts.High-dimensional word embeddings and attention:In 2000,the landmark paper,“A Neural Probabilistic Language Model”Introduced high-dimension word embeddings a
10、s a representation of word meaning.Had a huge and lasting impact on NLP tasks including MT,QA,and visual QA.His group also introduced a form of attention mechanism which led to breakthroughs in MT and form a key component of sequential processing with deep learning.Generative adversarial networks:Si
11、nce 2010,Generative Adversarial Networks(GANs),Ian Goodfellow,BengioA revolution in computer vision and computer graphics.Computers can actually create original images,reminiscent of the creativity that is considered a hallmark of human intelligence.Select Technical Accomplishments:Select Technical
12、Accomplishments:Yann Yann LeCunLeCun8Convolutional neural networks:1980s,developed CNN(Uni.ofToronto and Bell Labs)A foundational principle in the field,which,among other advantages,have been essential in making DL more efficient.Today,CNN are an industry standard in computer vision,as well as in sp
13、eech recognition,speech synthesis,image synthesis,and NLP.Used in a wide variety of applications,including autonomous driving,medical image analysis,voice-activated assistants,and information filtering.Improving BP alg.s:An early version of the BP alg.(backprop),gave a clean derivation based on vari
14、ational principles.Speed up BP alg.s:two simple methods to accelerate learning time.Broadening the vision of NNs:Developing a broader vision for NNs as a computational model for a wide range of tasksIntroducing in early work a number of concepts now fundamental in AILearning hierarchical feature rep
15、resentation;(Together with Lon Bottou)Learning systems can be built as complex networks of modules where BP is performed through automatic differentiation.Deep learning architectures can manipulate structured data,such as graphs从左至右:LeCun、Hinton、Bengio、吴恩达9Topic 6.ML Theory-I:Evaluating Hypotheses学习
16、理论 I:假设的评估问题Min Zhang Introduction to Machine Learning:Decision Tree Learning11Review:Inductive learning hypothesisMuch of the learning involves acquiring general concept from specific training examples.Inductive learning algorithms can at best guarantee that the output hypothesis fits the target co
17、ncept over the training data.Notice:over-fitting problemIntroduction to Machine Learning:Decision Tree Learning12Review:Inductive learning hypothesisThe Inductive Learning Hypothesis:Any hypothesis found to approximate the target function well over a sufficiently large set of training examples will
18、also approximate the target function well over unobserved examples.(任一假设若在足够大的训练样例集中很好地逼近目标函数,它也能在未见实例中很好地逼近目标函数)introduction to machine learning:Theory(I)Evaluating Hypotheses13Motivation Question 1Performance estimationGiven the observed accuracy of a hypothesis over a limited sample of data how w
19、ell does it estimate the accuracy over additional data?introduction to machine learning:Theory(I)Evaluating Hypotheses14Motivation Question 2h1 outperforms h2 over some sample of dataHow probable is it that h1 is better in general?introduction to machine learning:Theory(I)Evaluating Hypotheses15Moti
20、vation Question 3When data is limitedwhat is the best way to use this data to both learn a hypothesis and estimate its accuracy?The mathematical study of the likelihood and probability of events occurring based on known information and inferred by taking a limited number of samples.introduction to m
21、achine learning:Theory(I)Evaluating Hypotheses16Background Knowledge on Statisticsintroduction to machine learning:Theory(I)Evaluating Hypotheses17Basics of Sampling TheoryBernoulli experimentsOnly 2 outputs:success probability:p,fail probability:q=1-pUse random variable X to record the number of su
22、ccesslBinomial Distribution:lToss a coin:probability of heads side up p,toss n times,observed heads up r timeslIf X B(n,p)then Pr(X=r)=P(r)introduction to machine learning:Theory(I)Evaluating Hypotheses18Binomial distributionintroduction to machine learning:Theory(I)Evaluating Hypotheses19Where the
23、Binomial Distribution AppliesTwo possible outcomes(success and failure)(Y=0 or Y=1)The probability of success is the same on each trialPr(Y=1)=p,where p is a constantThere are n independent trailsRandom variables Y1,Yn,iid(independent identically distribution)R:random variable,count of Yiwhere Yi=1
24、on n trails,Pr(R=r)Binomial distributionMean(expected value):ER,Binomial distribution:=npVariance:VarR=E(R-ER)2,s2(Standard deviations)Binomial distribution:s2=np(1-p)introduction to machine learning:Theory(I)Evaluating Hypotheses20Discussions on Question 1introduction to machine learning:Theory(I)E
25、valuating Hypotheses21Review Question 1Performance estimationGiven the observed accuracy of a hypothesis over a limited sample of data how well does it estimate the accuracy over additional data?introduction to machine learning:Theory(I)Evaluating Hypotheses22Estimating Hypothesis Accuracy:Define Pr
- 配套讲稿:
如PPT文件的首页显示word图标,表示该PPT已包含配套word讲稿。双击word图标可打开word文档。
- 特殊限制:
部分文档作品中含有的国旗、国徽等图片,仅作为作品整体效果示例展示,禁止商用。设计者仅对作品中独创性部分享有著作权。
- 关 键 词:
- 机器学习概论机器学习概论 5 机器 学习 概论
![提示](https://www.taowenge.com/images/bang_tan.gif)
限制150内