A Comparative Analysis of Progressive Loss Functions with Multi Layered Resnet Model

Sravanthi Kantamaneni, Charles, T. Babu
{"title":"A Comparative Analysis of Progressive Loss Functions with Multi Layered Resnet Model","authors":"Sravanthi Kantamaneni, Charles, T. Babu","doi":"10.1109/CENTCON52345.2021.9687884","DOIUrl":null,"url":null,"abstract":"In this paper, one of the proposed ResNet model is used for denoising of RB noise. In fact, ResNet is one of the advanced deep learning methods for analysing and improving various 1D and 2D signals. Accuracy decreases due to the vanishing gradients in plain networks. The model Mozilla common speech data set is used. These are 48kHz recordings of all short sentence speaking subjects. They are all fixed at the same length and the same sampling frequency. The training course for this model uses an Adam optimizer/solver. This model is implemented in scheduling the learning rate “with a division” of 0.9 drop factor and a period of one. About 50 noise samples are available in the data set. Similarly, noise signals are acquired under various environmental conditions. Therefore, one separate data set is prepared for the T&T of the signal. When the T&T data set is small, the problem of overcompliance arises. In other words, since we are only trying to collect all data points from our dataset, we have used one proposed model to manage this dataset more efficiently. In the RMSE and precision validation values, you can feel the over- compliance issues here. Overfitting means that by 1 point of travel, the learning plot starts to deteriorate after loss and an increase in accuracy in terms of the identification. Similarly, if we are trying to pick a simple model for denoising, i.e. there is another problem - underfitting. Underfitting means that the model is either oversized or this model is oversized so that it doesn't learn enough about the dataset using that model. Each time various types of noises tries to rip off the amount added to the voice signal. Improvements in terms of denoising, RMSE and validation precision with the help of this model was given in the following sections.","PeriodicalId":103865,"journal":{"name":"2021 International Conference on Disruptive Technologies for Multi-Disciplinary Research and Applications (CENTCON)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Disruptive Technologies for Multi-Disciplinary Research and Applications (CENTCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CENTCON52345.2021.9687884","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, one of the proposed ResNet model is used for denoising of RB noise. In fact, ResNet is one of the advanced deep learning methods for analysing and improving various 1D and 2D signals. Accuracy decreases due to the vanishing gradients in plain networks. The model Mozilla common speech data set is used. These are 48kHz recordings of all short sentence speaking subjects. They are all fixed at the same length and the same sampling frequency. The training course for this model uses an Adam optimizer/solver. This model is implemented in scheduling the learning rate “with a division” of 0.9 drop factor and a period of one. About 50 noise samples are available in the data set. Similarly, noise signals are acquired under various environmental conditions. Therefore, one separate data set is prepared for the T&T of the signal. When the T&T data set is small, the problem of overcompliance arises. In other words, since we are only trying to collect all data points from our dataset, we have used one proposed model to manage this dataset more efficiently. In the RMSE and precision validation values, you can feel the over- compliance issues here. Overfitting means that by 1 point of travel, the learning plot starts to deteriorate after loss and an increase in accuracy in terms of the identification. Similarly, if we are trying to pick a simple model for denoising, i.e. there is another problem - underfitting. Underfitting means that the model is either oversized or this model is oversized so that it doesn't learn enough about the dataset using that model. Each time various types of noises tries to rip off the amount added to the voice signal. Improvements in terms of denoising, RMSE and validation precision with the help of this model was given in the following sections.
逐级损失函数与多层Resnet模型的比较分析
本文采用所提出的一种ResNet模型对RB噪声进行去噪。事实上,ResNet是一种先进的深度学习方法,用于分析和改进各种一维和二维信号。在平面网络中,由于梯度消失导致精度降低。使用Mozilla通用语音数据集模型。这些是所有短句说话对象的48kHz录音。它们都固定在相同的长度和相同的采样频率上。该模型的训练课程使用Adam优化/求解器。该模型实现在学习率“除”为0.9下降因子,周期为1的调度中。数据集中大约有50个噪声样本。同样,噪声信号是在各种环境条件下采集的。因此,为信号的T&T准备了一个单独的数据集。当T&T数据集很小时,就会出现过度遵从的问题。换句话说,由于我们只是试图从数据集中收集所有数据点,因此我们使用了一个提议的模型来更有效地管理该数据集。在RMSE和精度验证值中,您可以感受到这里的过度遵从性问题。过拟合是指每走1个点,学习图在丢失后开始恶化,在识别方面的准确性增加。同样,如果我们试图选择一个简单的模型去噪,也就是说,存在另一个问题——欠拟合。欠拟合意味着模型要么过大,要么这个模型过大,以至于它没有充分了解使用该模型的数据集。每次都有各种各样的噪音试图把添加到语音信号中的量扯掉。在此模型的帮助下,在去噪、RMSE和验证精度方面的改进将在以下章节中给出。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信