作为波前斜率预测器的简单神经网络:训练和性能问题

P. Gallant, G. Aitken
{"title":"作为波前斜率预测器的简单神经网络:训练和性能问题","authors":"P. Gallant, G. Aitken","doi":"10.1364/adop.1996.athc.8","DOIUrl":null,"url":null,"abstract":"Artificial neural networks have gained significant popularity over the past several years in a wide variety of engineering applications. This popularity is due in part to the ability of a neural network that is trained using a supervised training rule such as error backpropagation to acquire a nonparametric representation of the mapping between a set of inputs and outputs without any specific knowledge of the application domain. Given a sufficient number of nonlinear terms, represented by a number of hidden-layer neurons, a multilayer neural network can model any mathematical function that is continuous and differentiable (Hecht-Nielsen, 1990). Difficulties can arise however when a network is trained with a limited amount of noisy “real” data and is then expected to operate as part of a system for a specific application. The network must acquire an internal representation, as stored in its weights, during the training phase that subsequently generalizes well to unseen data. In the case of a prediction application, generalization capability becomes the paramount design criteria. The generalization performance of a trained network is a strong function of several factors, including: the architecture and complexity of the network, the type of supervised training rule employed, and the manner in which data is preprocessed and presented to the network.","PeriodicalId":256393,"journal":{"name":"Adaptive Optics","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Simple Neural Networks as Wavefront Slope Predictors: Training and Performance Issues\",\"authors\":\"P. Gallant, G. Aitken\",\"doi\":\"10.1364/adop.1996.athc.8\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Artificial neural networks have gained significant popularity over the past several years in a wide variety of engineering applications. This popularity is due in part to the ability of a neural network that is trained using a supervised training rule such as error backpropagation to acquire a nonparametric representation of the mapping between a set of inputs and outputs without any specific knowledge of the application domain. Given a sufficient number of nonlinear terms, represented by a number of hidden-layer neurons, a multilayer neural network can model any mathematical function that is continuous and differentiable (Hecht-Nielsen, 1990). Difficulties can arise however when a network is trained with a limited amount of noisy “real” data and is then expected to operate as part of a system for a specific application. The network must acquire an internal representation, as stored in its weights, during the training phase that subsequently generalizes well to unseen data. In the case of a prediction application, generalization capability becomes the paramount design criteria. The generalization performance of a trained network is a strong function of several factors, including: the architecture and complexity of the network, the type of supervised training rule employed, and the manner in which data is preprocessed and presented to the network.\",\"PeriodicalId\":256393,\"journal\":{\"name\":\"Adaptive Optics\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Adaptive Optics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1364/adop.1996.athc.8\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Adaptive Optics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1364/adop.1996.athc.8","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

人工神经网络在过去几年中在各种工程应用中获得了显著的普及。这种流行部分是由于神经网络使用监督训练规则(如误差反向传播)进行训练的能力,可以在不需要任何应用领域的特定知识的情况下获得一组输入和输出之间映射的非参数表示。给定足够数量的非线性项,由一些隐藏层神经元表示,多层神经网络可以模拟任何连续和可微的数学函数(Hecht-Nielsen, 1990)。然而,当一个网络用有限数量的有噪声的“真实”数据进行训练,然后期望作为特定应用系统的一部分运行时,就会出现困难。在训练阶段,网络必须获得一个内部表示,存储在它的权重中,随后可以很好地推广到看不见的数据。在预测应用程序中,泛化能力成为最重要的设计标准。训练网络的泛化性能是几个因素的强大函数,包括:网络的架构和复杂性,所采用的监督训练规则的类型,以及数据预处理和呈现给网络的方式。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Simple Neural Networks as Wavefront Slope Predictors: Training and Performance Issues
Artificial neural networks have gained significant popularity over the past several years in a wide variety of engineering applications. This popularity is due in part to the ability of a neural network that is trained using a supervised training rule such as error backpropagation to acquire a nonparametric representation of the mapping between a set of inputs and outputs without any specific knowledge of the application domain. Given a sufficient number of nonlinear terms, represented by a number of hidden-layer neurons, a multilayer neural network can model any mathematical function that is continuous and differentiable (Hecht-Nielsen, 1990). Difficulties can arise however when a network is trained with a limited amount of noisy “real” data and is then expected to operate as part of a system for a specific application. The network must acquire an internal representation, as stored in its weights, during the training phase that subsequently generalizes well to unseen data. In the case of a prediction application, generalization capability becomes the paramount design criteria. The generalization performance of a trained network is a strong function of several factors, including: the architecture and complexity of the network, the type of supervised training rule employed, and the manner in which data is preprocessed and presented to the network.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信