教程1:神经网络和支持向量机

C. Chandra Sekhar, S. Thamarai Selvi, C. N. Rao
{"title":"教程1:神经网络和支持向量机","authors":"C. Chandra Sekhar, S. Thamarai Selvi, C. N. Rao","doi":"10.1109/ICOAC.2011.6165238","DOIUrl":null,"url":null,"abstract":"Discriminative learning based approaches to pattern classification are important for tasks that involve nonlinearly separable classes and overlapping classes as in many real world pattern classification tasks. Multilayer feedforward neural networks built using the computational models of neurons with sigmoidal activation function, and trained using the error back propagation learning algorithm have been explored for complex pattern classification tasks. The main limitations of theses models are the slow convergence of the learning method, the local minima problem and the poor generalization ability of trained models. Support vector machines overcome these limitations by using the principles of kernel methods. In the kernel methods for pattern classification, the first stage involves nonlinear transformation of representation of an example in a low dimensional input feature space to a representation in a high dimensional feature space induced by a kernel function, so that the nonlinearly separable classes in the input feature space are likely be linearly separable classes in the kernel feature space. The second stage in the kernel methods involves constructing an optimal linear solution in the kernel feature space that corresponds to an optimal nonlinear solution in the input feature space. The main advantages of the support vector machines are the good generalization ability and the requirement of small size training data sets. The main issue in the design of support vector machines is the choice of kernel function that induces the nonlinear transformation. Support vector machines can be used for vectorial representations of data as well as for non-vectorial representations of data, whereas multilayer feed forward neural networks can be used mainly for vectorial representations of data. The tutorial presents the underlying principles of approaches to pattern classification using multilayer feedforward neural networks and support vector machines, and discusses the issues in developing pattern classification models using these approaches. Some applications of support vector machines to pattern classification tasks in speech and image processing are also presented.","PeriodicalId":369712,"journal":{"name":"2011 Third International Conference on Advanced Computing","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Tutorial I: Neural networks and support vector machines\",\"authors\":\"C. Chandra Sekhar, S. Thamarai Selvi, C. N. Rao\",\"doi\":\"10.1109/ICOAC.2011.6165238\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Discriminative learning based approaches to pattern classification are important for tasks that involve nonlinearly separable classes and overlapping classes as in many real world pattern classification tasks. Multilayer feedforward neural networks built using the computational models of neurons with sigmoidal activation function, and trained using the error back propagation learning algorithm have been explored for complex pattern classification tasks. The main limitations of theses models are the slow convergence of the learning method, the local minima problem and the poor generalization ability of trained models. Support vector machines overcome these limitations by using the principles of kernel methods. In the kernel methods for pattern classification, the first stage involves nonlinear transformation of representation of an example in a low dimensional input feature space to a representation in a high dimensional feature space induced by a kernel function, so that the nonlinearly separable classes in the input feature space are likely be linearly separable classes in the kernel feature space. The second stage in the kernel methods involves constructing an optimal linear solution in the kernel feature space that corresponds to an optimal nonlinear solution in the input feature space. The main advantages of the support vector machines are the good generalization ability and the requirement of small size training data sets. The main issue in the design of support vector machines is the choice of kernel function that induces the nonlinear transformation. Support vector machines can be used for vectorial representations of data as well as for non-vectorial representations of data, whereas multilayer feed forward neural networks can be used mainly for vectorial representations of data. The tutorial presents the underlying principles of approaches to pattern classification using multilayer feedforward neural networks and support vector machines, and discusses the issues in developing pattern classification models using these approaches. Some applications of support vector machines to pattern classification tasks in speech and image processing are also presented.\",\"PeriodicalId\":369712,\"journal\":{\"name\":\"2011 Third International Conference on Advanced Computing\",\"volume\":\"4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 Third International Conference on Advanced Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICOAC.2011.6165238\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 Third International Conference on Advanced Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICOAC.2011.6165238","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

基于判别学习的模式分类方法对于涉及非线性可分类和重叠类的任务非常重要,就像许多现实世界的模式分类任务一样。利用具有s型激活函数的神经元计算模型构建多层前馈神经网络,并使用误差反向传播学习算法进行训练,探索了用于复杂模式分类任务的多层前馈神经网络。这些模型的主要局限性是学习方法收敛速度慢,存在局部极小问题,训练模型泛化能力差。支持向量机通过使用核方法的原理克服了这些限制。在模式分类的核方法中,第一阶段是将样本在低维输入特征空间中的表示非线性转换为由核函数诱导的高维特征空间中的表示,从而使输入特征空间中的非线性可分类很可能是核特征空间中的线性可分类。核方法的第二阶段涉及在核特征空间中构造一个最优线性解,该解对应于输入特征空间中的最优非线性解。支持向量机的主要优点是良好的泛化能力和对训练数据集的要求较小。支持向量机设计中的主要问题是核函数的选择,从而引起非线性变换。支持向量机既可以用于数据的向量表示,也可以用于数据的非向量表示,而多层前馈神经网络主要用于数据的向量表示。本教程介绍了使用多层前馈神经网络和支持向量机进行模式分类方法的基本原理,并讨论了使用这些方法开发模式分类模型的问题。本文还介绍了支持向量机在语音和图像处理中模式分类任务中的一些应用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Tutorial I: Neural networks and support vector machines
Discriminative learning based approaches to pattern classification are important for tasks that involve nonlinearly separable classes and overlapping classes as in many real world pattern classification tasks. Multilayer feedforward neural networks built using the computational models of neurons with sigmoidal activation function, and trained using the error back propagation learning algorithm have been explored for complex pattern classification tasks. The main limitations of theses models are the slow convergence of the learning method, the local minima problem and the poor generalization ability of trained models. Support vector machines overcome these limitations by using the principles of kernel methods. In the kernel methods for pattern classification, the first stage involves nonlinear transformation of representation of an example in a low dimensional input feature space to a representation in a high dimensional feature space induced by a kernel function, so that the nonlinearly separable classes in the input feature space are likely be linearly separable classes in the kernel feature space. The second stage in the kernel methods involves constructing an optimal linear solution in the kernel feature space that corresponds to an optimal nonlinear solution in the input feature space. The main advantages of the support vector machines are the good generalization ability and the requirement of small size training data sets. The main issue in the design of support vector machines is the choice of kernel function that induces the nonlinear transformation. Support vector machines can be used for vectorial representations of data as well as for non-vectorial representations of data, whereas multilayer feed forward neural networks can be used mainly for vectorial representations of data. The tutorial presents the underlying principles of approaches to pattern classification using multilayer feedforward neural networks and support vector machines, and discusses the issues in developing pattern classification models using these approaches. Some applications of support vector machines to pattern classification tasks in speech and image processing are also presented.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信