基于Langevin-Neelakanta机器的新型深度学习人工神经网络

D. De Groff, P. Neelakanta
{"title":"基于Langevin-Neelakanta机器的新型深度学习人工神经网络","authors":"D. De Groff, P. Neelakanta","doi":"10.14738/tmlai.104.13100","DOIUrl":null,"url":null,"abstract":"In the contexts of deep learning (DL) considered in artificial intelligence (AI) efforts, relevant machine learning (ML) algorithms adopted refer to using a class of deep artificial neural network (ANN) that supports a learning process exercised with an enormous set of input data (labeled and/or unlabeled) so to predict at the output details on accurate features of labeled data present in the input data set. In the present study, a deep ANN is proposed thereof conceived with certain novel considerations: The proposed deep architecture consists of a large number of consequently placed structures of paired-layers. Each layer hosts identical number of neuronal units for computation and the neuronal units are massively interconnected across the entire network. Further, each paired-layer is independently subjected to unsupervised learning (USL). Hence, commencing from the input layer-pair, the excitatory (input) data supplied flows across the interconnected neurons of paired layers, terminating eventually at the final pair of layers, where the output is recovered. That is, the converged neuronal states at any given pair is iteratively passed on to the next pair and so on. The USL suite involves collectively gathering the details of neural information across a pair of the layers constituting the network. This summed data is then limited with a specific choice of a squashing (sigmoidal) function; and, the resulting scaled value is used to adjust the coefficients of interconnection weights seeking a convergence criterion. The associated learning rate on weight adjustment is uniquely designed to facilitate fast learning towards convergence. The unique aspects of deep learning proposed here refer to: (i) Deducing the learning coefficient with a compatible algorithm so as to realize a fast convergence; and, (ii) the adopted sigmoidal function in the USL loop conforms to the heuristics of the so-called Langevin-Neelakanta machine. The paper describes the proposed deep ANN architecture with necessary details on structural considerations, sigmoidal selection, prescribing required learning rate and operational (training and predictive phase) routines. Results are furnished to demonstrate the performance efficacy of the test ANN.","PeriodicalId":119801,"journal":{"name":"Transactions on Machine Learning and Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Novel Deep Learning ANN Supported on Langevin-Neelakanta Machine\",\"authors\":\"D. De Groff, P. Neelakanta\",\"doi\":\"10.14738/tmlai.104.13100\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the contexts of deep learning (DL) considered in artificial intelligence (AI) efforts, relevant machine learning (ML) algorithms adopted refer to using a class of deep artificial neural network (ANN) that supports a learning process exercised with an enormous set of input data (labeled and/or unlabeled) so to predict at the output details on accurate features of labeled data present in the input data set. In the present study, a deep ANN is proposed thereof conceived with certain novel considerations: The proposed deep architecture consists of a large number of consequently placed structures of paired-layers. Each layer hosts identical number of neuronal units for computation and the neuronal units are massively interconnected across the entire network. Further, each paired-layer is independently subjected to unsupervised learning (USL). Hence, commencing from the input layer-pair, the excitatory (input) data supplied flows across the interconnected neurons of paired layers, terminating eventually at the final pair of layers, where the output is recovered. That is, the converged neuronal states at any given pair is iteratively passed on to the next pair and so on. The USL suite involves collectively gathering the details of neural information across a pair of the layers constituting the network. This summed data is then limited with a specific choice of a squashing (sigmoidal) function; and, the resulting scaled value is used to adjust the coefficients of interconnection weights seeking a convergence criterion. The associated learning rate on weight adjustment is uniquely designed to facilitate fast learning towards convergence. The unique aspects of deep learning proposed here refer to: (i) Deducing the learning coefficient with a compatible algorithm so as to realize a fast convergence; and, (ii) the adopted sigmoidal function in the USL loop conforms to the heuristics of the so-called Langevin-Neelakanta machine. The paper describes the proposed deep ANN architecture with necessary details on structural considerations, sigmoidal selection, prescribing required learning rate and operational (training and predictive phase) routines. Results are furnished to demonstrate the performance efficacy of the test ANN.\",\"PeriodicalId\":119801,\"journal\":{\"name\":\"Transactions on Machine Learning and Artificial Intelligence\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Transactions on Machine Learning and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.14738/tmlai.104.13100\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transactions on Machine Learning and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.14738/tmlai.104.13100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在人工智能(AI)工作中考虑的深度学习(DL)背景下,采用的相关机器学习(ML)算法是指使用一类深度人工神经网络(ANN),该网络支持使用大量输入数据(标记和/或未标记)进行学习过程,以便预测输入数据集中存在的标记数据的准确特征的输出细节。在本研究中,提出了一个深度人工神经网络,其中考虑了一些新颖的因素:所提出的深度架构由大量成对层的相应放置结构组成。每一层承载相同数量的用于计算的神经元单元,并且神经元单元在整个网络中大规模互连。此外,每个配对层都独立地进行无监督学习(USL)。因此,从输入层对开始,提供的兴奋性(输入)数据流经成对层的相互连接的神经元,最终终止于最后一层对,在那里输出被恢复。也就是说,任意给定对的收敛神经元状态迭代地传递给下一对,以此类推。USL套件包括在构成网络的一对层上集体收集神经信息的细节。然后,这个汇总的数据被限制为一个特定的压缩(s型)函数的选择;并利用得到的缩放值来调整互连权值系数,寻求收敛准则。权重调整的相关学习率设计独特,以促进快速学习趋于收敛。这里提出的深度学习的独特之处在于:(i)用兼容的算法推导学习系数,从而实现快速收敛;(ii) USL回路中采用的s型函数符合所谓Langevin-Neelakanta机的启发式。本文描述了所提出的深度人工神经网络架构,并提供了有关结构考虑、s型选择、规定所需学习率和操作(训练和预测阶段)例程的必要细节。实验结果验证了测试人工神经网络的性能有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Novel Deep Learning ANN Supported on Langevin-Neelakanta Machine
In the contexts of deep learning (DL) considered in artificial intelligence (AI) efforts, relevant machine learning (ML) algorithms adopted refer to using a class of deep artificial neural network (ANN) that supports a learning process exercised with an enormous set of input data (labeled and/or unlabeled) so to predict at the output details on accurate features of labeled data present in the input data set. In the present study, a deep ANN is proposed thereof conceived with certain novel considerations: The proposed deep architecture consists of a large number of consequently placed structures of paired-layers. Each layer hosts identical number of neuronal units for computation and the neuronal units are massively interconnected across the entire network. Further, each paired-layer is independently subjected to unsupervised learning (USL). Hence, commencing from the input layer-pair, the excitatory (input) data supplied flows across the interconnected neurons of paired layers, terminating eventually at the final pair of layers, where the output is recovered. That is, the converged neuronal states at any given pair is iteratively passed on to the next pair and so on. The USL suite involves collectively gathering the details of neural information across a pair of the layers constituting the network. This summed data is then limited with a specific choice of a squashing (sigmoidal) function; and, the resulting scaled value is used to adjust the coefficients of interconnection weights seeking a convergence criterion. The associated learning rate on weight adjustment is uniquely designed to facilitate fast learning towards convergence. The unique aspects of deep learning proposed here refer to: (i) Deducing the learning coefficient with a compatible algorithm so as to realize a fast convergence; and, (ii) the adopted sigmoidal function in the USL loop conforms to the heuristics of the so-called Langevin-Neelakanta machine. The paper describes the proposed deep ANN architecture with necessary details on structural considerations, sigmoidal selection, prescribing required learning rate and operational (training and predictive phase) routines. Results are furnished to demonstrate the performance efficacy of the test ANN.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信