欢迎来到淘文阁 - 分享文档赚钱的网站! | 帮助中心 好文档才是您的得力助手!
淘文阁 - 分享文档赚钱的网站
全部分类
  • 研究报告>
  • 管理文献>
  • 标准材料>
  • 技术资料>
  • 教育专区>
  • 应用文书>
  • 生活休闲>
  • 考试试题>
  • pptx模板>
  • 工商注册>
  • 期刊短文>
  • 图片设计>
  • ImageVerifierCode 换一换

    9人工智能导论 (17).pdf

    • 资源ID:4060227       资源大小:1.23MB        全文页数:30页
    • 资源格式: PDF        下载积分:2金币
    快捷下载 游客一键下载
    会员登录下载
    微信登录下载
    三方登录下载: 微信开放平台登录   QQ登录  
    二维码
    微信扫一扫登录
    下载资源需要2金币
    邮箱/手机:
    温馨提示:
    快捷下载时,用户名和密码都是您填写的邮箱或者手机号,方便查询和重复下载(系统自动生成)。
    如填写123,账号就是123,密码也是123。
    支付方式: 支付宝    微信支付   
    验证码:   换一换

     
    账号:
    密码:
    验证码:   换一换
      忘记密码?
        
    友情提示
    2、PDF文件下载后,可能会被浏览器默认打开,此种情况可以点击浏览器菜单,保存网页到桌面,就可以正常下载了。
    3、本站不支持迅雷下载,请使用电脑自带的IE浏览器,或者360浏览器、谷歌浏览器下载即可。
    4、本站资源下载后的文档和图纸-无水印,预览文档经过压缩,下载后原文更清晰。
    5、试题试卷类文档,如果标题没有明确说明有答案则都视为没有答案,请知晓。

    9人工智能导论 (17).pdf

    Networked Models Artificial Intelligence 52 12. Models in Machine Learning 12.1. Probabilistic Models 12.2. Geometric Models 12.3. Logical Models 12.4. Networked Models Contents: Artificial Intelligence : Learning : Models 53 The networked models here refer to as the models of artificial neural network (ANN). 这里的网络化模型指的是人工神经网络模型 (ANN)。 An ANN is an artificial representation of the human brain that tries to simulate its learning processing. 一个ANN是人脑的一种人工表征,试图模拟人类的学习过程。 ANN can be constructed a system by interconnected “neurons” which send messages to each other. ANN可以通过互联的“神经元”构建一个系统,神经元之间相互发送消息。 The connections between neurons have numeric weights that can be tuned based on experience, making ANN adaptive to inputs and capable of learning. 神经元之间的连接具有数值权重,可以通过经验调整,使ANN适应输入并且能够学习。 What are Networked Models 什么是网络化模型 12.4. Networked Models Artificial Intelligence 54 12.4. Networked Models 12.4.1. Artificial Neural Networks 12.4.2. Deep Neural Networks Contents: Artificial Intelligence : Learning : Models 55 Biological Neuron 生物神经元 12.4.1. Artificial Neural Networks Input 输入 (Stimulus) 刺激 Output 输出 (Response) 反应 Dendrite 树突 Nucleus 细胞核 Axon 轴突 Synapse 突触 Artificial Intelligence : Learning : Models 56 Artificial Neuron 人工神经元 12.4.1. Artificial Neural Networks Neuron i 神经元 Input 输入 (Stimulus) 刺激 Output 输出 (Response) 反应 yi Activation Function 激活函数 Activation function: 激活函数 or x1 x2 x5 x3 x4 wi1 wi3 wi2 wi4 wi5 ui Artificial Intelligence : Learning : Models 57 Artificial Neural Network (ANN) 人工神经网络 12.4.1. Artificial Neural Networks ANN is a family of learning models inspired by biological neural networks The interconnection between the different layers of neurons The learning process for updating the weights of the interconnections The activation function that converts a neurons weighted input to its output Input 输入 (Stimulus) 刺激 Output 输出 (Response) 反应 Hidden layers 隐藏层 Input layers 输入层 Output layers 输出层 ANN是受生物神经网络启发的一系列学习模型 不同的神经元层次之间互联 学习过程是为了更新互联权重 激活函数将神经元的加权输入转换为其输出 Artificial Intelligence : Learning : Models 58 1943, McCulloch and Pitts 马卡洛和匹茨 created a computational model for neural networks based on mathematics and algorithms called threshold logic. 基于称之为阈值逻辑的数学和算法创建了神经网络的计算模型。 1954, Farley and Clark 法利和克拉克 first used computational machines, then called calculators, to simulate a Hebbian network. 首次利用计算的机器、后来称其为计算器,来仿真赫布网络。 1958, Rosenblatt 罗森布莱特 created perceptron, an algorithm for pattern recognition, which is with only one output layer, so also called “single layer perceptron”. 创建了感知机,一种模式识别算法,它仅有一个输出 层,也被称为“单层感知机”。 History of Artificial Neural Networks 人工神经网络的发展史 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 59 1969, Minsky and Papert 明斯基和帕伯特 Published a famous book entitled “Perceptrons”. 出版了一本名为“感知机”的著名书籍。 It pointed in this book that the single layer perceptrons are only capable of learning linearly separable patterns, but not possible to learn an XOR function. 书中指出,单层感知机仅能学习线性可分模式,而不能用于学习异或功能。 1974, Werbos 韦伯斯 Proposed the back-propagationalgorithm, a method for training ANNs and used in conjunction with an optimization method such as gradient descent. 提出了反向传播算法,一种用于训练ANNs的方法,并且与 梯度下降等优化方法结合使用。 Regenerates interest in the 1980s. 1980年代才引起重视。 History of Artificial Neural Networks 人工神经网络的发展史 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 60 1989, Yann LeCun et al 雅恩 勒昆等人 Published LeNet-5, a pioneering 7-level convolutional neural network (CNN) is applied to recognize hand-written numbers on checks. 发表了LeNet-5,一种开拓性的7层卷积神经网络 (CNN),用于检查支票 上的手写数字。 1992, Schmidhuber 施米德胡贝 Proposed recurrent neural network (RNN), this creates an internal state which allows it to exhibit dynamic temporal behavior. 提出了循环神经网络,它创建网络的内部状态,得以展现 动态时间行为。 2006, Hinton and Salakhutdinov 辛顿和萨拉赫丁诺 夫 Renewed interest in neural nets was sparked by the advent of deep learning. 深度学习的出现,再次引发了对神经网络的兴趣。 History of Artificial Neural Networks 人工神经网络的发展史 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 61 2012, Andrew Ng and Jeff Dean 吴恩达和杰夫 迪 恩 Google Brain team created a neural network that learned to recognize higher-level concepts, such as cats, from watching unlabeled images. Google大脑团队创建了一个神经网络,学会观看未标注图像来识别高层 次概念,例如猫。 2012, Krizhevsky et al 克利则夫斯基等 With Deep CNNs won the large-scale ImageNet competition by a significant margin over shallow machine learning methods. 采用深度CNNs获得了大规模ImageNet比赛的胜利,比浅 层学习方法有显著优势。 2014, Ian Goodfellow et al 伊恩 古德菲勒等 Proposed generative adversarial network (GAN) which has two neural networks competing against each other in a zero-sum game framework. 提出了生成对抗网络(GAN),其中有两个神经网 络,彼此以“零和”博弈方式相互竞争。 History of Artificial Neural Networks 人工神经网络的发展史 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 62 Structures of Neural Networks 神经网络的结构 12.4.1. Artificial Neural Networks Neural Networks 神经网络 Feedforward Neural Networks 前馈神经网络 Recurrent Neural Networks 循环神经网络 Artificial Intelligence : Learning : Models 63 Feedforward neural network 前馈神经网络 information moves in only one direction, forward, from input nodes, through hidden nodes and to the output nodes. 信息从输入结点仅仅以一个方向,即前进方向,穿过隐 藏层并抵达输出节点。 Recurrent neural network 循环神经网络 connections form a directed cycle. 连接形成有向循环。 creating an internal state of the network which allows it to exhibit dynamic temporal behavior. 建立网络的内部状态,使之展现动态的时间特性。 Structures of Neural Network Models 神经网络模型的结构 12.4.1. Artificial Neural Networks Feedforward neural network 前馈神经网络 Recurrent neural network 循环神经网络 Artificial Intelligence : Learning : Models 64 Back-propagation (BP) is an abbreviation for “Backward propagation of errors”. 反向传播 (BP) 是“反向误差传播”的缩略语。 It is a common method of training Artificial Neural Networks, and used in conjunction with an optimization method such as gradient descent. 是训练人工神经网络的常用方法,与梯度下降优化方法结合使用。 The algorithm repeats a two phase cycle: 该算法重复两个阶段的循环: Back-propagation 反向传播 12.4.1. Artificial Neural Networks phase 1: propagation 传播 phase 2: weight update 权值更新 Repeat phase 1 and phase 2 until the performance of the network is satisfactory. 重复阶段1和阶段2的操作,直到 网络的性能得到满足。 Artificial Intelligence : Learning : Models 65 Phase 1: Propagation第1阶段:传播 Feedforward propagation前馈传播 the input of training data through the neural network in order to generate output activations. 输入的训练数据穿过神经网络,从而生成输出激活值。 Back-propagation反向传播 the output activations through the neural network using the training data target in order to generate the deltasof all output and hidden neurons. 输出激活再使用训练数据目标穿过神经网络,生成所有的输出层和隐藏层神经元的差值。 deltas = expected output - actual output values 差值 = 期待输出 - 实际输出 Algorithm of Back-propagation 反向传播算法 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 66 Phase 2: Weight update第2阶段:权值更新 For each weight: 对每个权值: Multiply its output delta and input activation, to get the gradient of the weight. 将其输出差值与输入激活相乘,以便得到该权值梯度。 Subtract a ratio(percentage) of the gradient from the weight. The ratio is called learning rate. 从权值中减去梯度的比值(百分比)。该比值被称为学习率。 The greater the ratio, the faster the neuron trains; 比值越大,神经元训练越快; the lower the ratio, the more accurate the training is. 比值越低,训练精度越高。 Algorithm of Back-propagation 反向传播算法 12.4.1. Artificial Neural Networks Artificial Intelligence : Learning : Models 67 A Stochastic Gradient Descent Algorithm 随机梯度下降算法 12.4.1. Artificial Neural Networks functionSTOCHASTIC-GRADIENT-DESCENT() return the network initialize network weights (often small random values) do for each training example named ex prediction = neural-net-output(network, ex) / forward pass actual = teacher-output(ex) compute error (prediction - actual) at the output units compute whfor all weights from hidden layer to output layer / backward pass compute wifor all weights from input layer to hidden layer update network weights / input layer not modified by error estimate until all examples classified correctly or another stopping criterion satisfied return the network For training a three-layer network (only one hidden layer) 用于训练一个三层网络(仅有一个隐藏层) Artificial Intelligence : Learning : Models 68 (c) Good fit of the data 数据良拟合 Comparison of Training Results 训练结果的比较 12.4.1. Artificial Neural Networks (a) Under-fit of the data 数据低拟合 (b) Over-fit of the data 数据过拟合 Artificial Intelligence : Learning : Models 69 There is no universally agreed upon threshold of depth dividing shallow neural networks from deep neural networks. 就划分浅层神经网络与深层神经网络的深度而言,尚 未有公认的观点。 Shallow vs. Deep Neural Network 浅层与深层神经网络 12.4.1. Artificial Neural Networks But most researchers agree that deep neural networks have more than 2 of hidden layers, and hidden layers 10 to be very deep neural networks. 但大多数研究人员认为,深度神经网络的隐藏层 超过2、而隐藏层大于10的为超深度神经网络 。 Artificial Neural Network 人工神经网络 Input layer 输入层 Output layer 输出层 Hidden layers 隐藏层 Artificial Intelligence 70 12.4. Networked Models 12.4.1. Artificial Neural Networks 12.4.2. Deep Neural Networks Contents: Artificial Intelligence : Learning : Models 71 Biological: Visual cortex is Deep Hierarchical 生物学:视觉皮层是深层次的 Why Deep Hierarchy 为什么深度层次 12.4.2. Deep Neural Networks Artificial Intelligence : Learning : Models 72 DNNs use many layers of nonlinear processing units for feature extraction and transformation. DNNs使用许多层非线性处理单元,用于特征提取和转换。 Able to learn multiple levels of features or representations of the data. Higher level features are derived from lower level features. 能够学习数据的多层特征或表征。高层特征来自于低层特征。 Be part of the broader machine learning field: learning representations of data. 成为更广泛的机器学习领域的一部分:学习数据表征。 Learning multiple levels of representations that correspond to different levels of abstraction; the levels form a hierarchy of concepts. 学习多层级表征,对应于不同的抽象层级;这种层级形成了一种概念的层次结构。 Deep Neural Networks (DNNs) 深度神经网络 12.4.2. Deep Neural Networks Artificial Intelligence : Learning : Models 73 Deep belief networks (DBN) Convolutional neural networks (CNN) Deep Boltzmann machines (DBM) Recurrent neural networks (RNN) Long short-term memory (LSTM) Auto-Encoders Generative Adversarial Network (GAN) Typical Deep Neural Networks 代表性的深度神经网络 12.4.2. Deep Neural Networks 深度信念网络 (DBN) 卷积神经网络 (CNN) 深度波兹曼机 (DBM) 循环网络 (RNN) 长短期记忆 (LSTM) 自动编码器 生成对抗网络 (GAN) Artificial Intelligence : Learning : Models 74 CNN is a type of feed-forward artificial neural network that uses at least one of convolution in place of general matrix multiplication. CNN是一种前馈式人工神经网络,使用至少一个卷积层来代替一般的矩阵乘法。 Case Study: Convolutional neural network (CNN) 卷积神经网络 12.4.2. Deep Neural Networks Four key ideas: 四个关键思想: local connections (convolution) 局部连接(卷积) shared weights 共享权值 pooling (sampling) 池化(采样) many layers. 多层 . . Input 输入 Convolution 卷积 Pooling 池化 Fully conn. 全连接 Softmax 分类器 . . Artificial Intelligence : Learning : Models 75 Convolution with a 3x3 filter 用3x3滤波器进行卷积 Convolution layer卷积层 Consist of a set of learnable filters, each filter is convolved across the width and height of the input volume, computing the dot product between the entries of the filter and the input, and producing a 2-dimensional activation map of that filter. 包含一组学习滤波器,每个滤波器对输入的宽和高进行卷积,计算滤波器 和输入之间的点积,生成一个该滤波器的2维活动图。 Pooling layer池化层 A form of non-linear down-sampling. It partitions the input image into a set of non-overlapping rectangles and, for each such sub-region, outputs the maximum. 是一种非线性下采样。它将输入图像分割成一组不重叠的矩形,对每个这 样的子区域,再产出其最大值。 Case Study: Convolutional neural network (CNN) 卷积神经网络 12.4.2. Deep Neural Networks Max pooling with a 2x2 filter and stride = 2 用2x2滤波器进行最大池化,步长=2 Source: Artificial Intelligence : Learning : Models 76 Case Study: Generative Adversarial Network (GAN) 生成对抗网络 12.4.2. Deep Neural Networks GAN is pioneered by Ian Goodfellow et al at University of Montreal in 2014. GAN是由Goodfellow等人于2014年在蒙特利尔大学开创的。 Adversarial Network, inspired by Adversarial game. 对抗网络,灵感源于对抗博弈。 Generator maps from a latent space to a particular data distribution of interest. 生成器从潜在空间映射到所关注的特定数据分布。 Discriminator discriminate between real samples and samples produced by generator. 判别器在真实样本和生成样本之间进行判别。 Training a model in a worst-case scenario, with inputs chosen by an adversary. 在最坏场景下训练模型,由adversary选择输入。 Source: Goodfellow NIPS 2016 Workshop Artificial Intelligence : Learning : Models 77 Case Study: Generative Adversarial Network (GAN) 生成对抗网络 12.4.2. Deep Neural Networks The generator worked well with digits (a) and faces (b), but it created very fuzzy and vague images (c) and (d) when using the CIFAR-10 dataset. 该生成器对数字(a)和人脸(b)效果好,但使用CIFAR-10 数据集时,却生成了非常模糊和含糊的图像(c)和(d)。 Generative Adversarial Network (GAN) 生成对抗网络 Artificial Intelligence : Learning : Models 78 Speech recognition Object recognition Image retrieval Image understanding Natural language processing Recommendation systems Drug discovery Biomedical informatics Typical Applications of Deep Neural Networks 深度神经网络的主要应用 12.4.2. Deep Neural Networks 语音识别 物体识别 图像检索 图像理解 自然语言处理 推荐系统 药物发现 生物医学信息学 Artificial Intelligence : Learning : Models 79 Deep Learning is a new area of Machine Learning research, which has been introduced with the objective of moving Machine Learning closer to one of its original goals: Artificial Intelligence. 12.4. Networked Models Source: 深度学习是机器学习研究的一个新领域,其研究目的是 使机器学习更接近其原始目标之一:人工智能。 Postscript 后记

    注意事项

    本文(9人工智能导论 (17).pdf)为本站会员(奉***)主动上传,淘文阁 - 分享文档赚钱的网站仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知淘文阁 - 分享文档赚钱的网站(点击联系客服),我们立即给予删除!

    温馨提示:如果因为网速或其他原因下载失败请重新下载,重复下载不扣分。




    关于淘文阁 - 版权申诉 - 用户使用规则 - 积分规则 - 联系我们

    本站为文档C TO C交易模式,本站只提供存储空间、用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。本站仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知淘文阁网,我们立即给予删除!客服QQ:136780468 微信:18945177775 电话:18904686070

    工信部备案号:黑ICP备15003705号 © 2020-2023 www.taowenge.com 淘文阁 

    收起
    展开