2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)最新文献

筛选
英文 中文
Real-Time Optimizing Weighted Gaussian Curvature for 4K Videos 4K视频实时优化加权高斯曲率
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596473
Wenming Tang, Lebin Zhou, Yuanhao Gong
{"title":"Real-Time Optimizing Weighted Gaussian Curvature for 4K Videos","authors":"Wenming Tang, Lebin Zhou, Yuanhao Gong","doi":"10.1109/mlsp52302.2021.9596473","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596473","url":null,"abstract":"Weighted Gaussian curvature is an important measurement for surfaces, images and videos. However, since it is a second order quantity, its optimization usually leads to high order geometric flows that cause difficulties for practical applications. In this paper, we propose a novel optimization method for weighted Gaussian curvature. Our method does not require the image (video) to be second-order differentiable, thus, avoiding the high order geometric flows. In addition, we propose a new 4-D look-up table method to accelerate the optimization of weighted Gaussian curvature. Therefore, our algorithm is very efficient and can achieve real-time processing for high resolution videos. For example, our method can process 4K videos with 50 frames per second on a single graphic card (NVIDIA 3090). Several numerical experiments are carried out to confirm the efficiency and effectiveness of the proposed method. Thanks to the high performance, our method can be applied in a large range of applications that involve weighted Gaussian curvature, such as image restoration, registration, enhancement, etc.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124651925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Introducing K-Anonymity Principles to Adversarial Attacks for Privacy Protection in Image Classification Problems 将k -匿名原理引入图像分类问题中隐私保护的对抗性攻击
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596565
V. Mygdalis, A. Tefas, I. Pitas
{"title":"Introducing K-Anonymity Principles to Adversarial Attacks for Privacy Protection in Image Classification Problems","authors":"V. Mygdalis, A. Tefas, I. Pitas","doi":"10.1109/mlsp52302.2021.9596565","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596565","url":null,"abstract":"The network output activation values for a given input can be employed to produce a sorted ranking. Adversarial attacks typically generate the least amount of perturbation required to change the classifier label. In that sense, generated adversarial attack perturbation only affects the output in the 1st sorted ranking position. We argue that meaningful information about the adversarial examples i.e., their original labels, is still encoded in the network output ranking and could potentially be extracted, using rule-based reasoning. To this end, we introduce a novel adversarial attack methodology inspired by the K-anonymity principles, that generates adversarial examples that are not only misclassified, but their output sorted ranking spreads uniformly along K different positions. Any additional perturbation arising from the strength of the proposed objectives, is regularized by a visual similarity-based term. Experimental results denote that the proposed approach achieves the optimization goals inspired by K-anonymity with reduced perturbation as well.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132104106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Optimizing Time Domain Fully Convolutional Networks for 3D Speech Enhancement in a Reverberant Environment Using Perceptual Losses 利用感知损失优化时域全卷积网络在混响环境下的三维语音增强
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596103
Heitor R. Guimarães, Wesley Beccaro, M. A. Ramírez
{"title":"Optimizing Time Domain Fully Convolutional Networks for 3D Speech Enhancement in a Reverberant Environment Using Perceptual Losses","authors":"Heitor R. Guimarães, Wesley Beccaro, M. A. Ramírez","doi":"10.1109/mlsp52302.2021.9596103","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596103","url":null,"abstract":"Noise in 3D reverberant environment is detrimental to several downstream applications. In this work, we propose a novel approach to 3D speech enhancement directly in the time domain through the usage of Fully Convolutional Networks (FCN) with a custom loss function based on the combination of a perceptual loss, built on top of the wav2vec model and a soft version of the short-time objective intelligibility (STOI) metric. The dataset and experiments were based on Task 1 of the L3DAS21 challenge. Our model achieves a STOI score of 0.82, word error rate (WER) equal to 0.36, and a score of 0.73 in the metric proposed by the challenge based on STOI and WER combination using as reference the development set. Our submission, based on this method, was ranked second in Task 1 of the L3DAS21 challenge.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127394501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Bag of Groups of Convolutional Features Model for Visual Object Recognition 用于视觉目标识别的卷积特征包模型
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596432
Jaspreet Singh, Chandan Singh
{"title":"Bag of Groups of Convolutional Features Model for Visual Object Recognition","authors":"Jaspreet Singh, Chandan Singh","doi":"10.1109/mlsp52302.2021.9596432","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596432","url":null,"abstract":"Deep convolutional neural networks (CNNs) are only equivariant to translation. Recently, equivariant CNNs are proposed for the task of image classification which are not only equivariant to translation but also to other affine geometric transformations. Moreover, CNNs and equivariant CNNs require a large amount of labeled training data to generalize its parameters which also limit their application areas. We propose a bag of groups of convolutional features (BoGCFs) model for the CNNs and group-equivariant CNNs (G-CNNs)[1], which preserves the fundamental property of equivariance of G-CNNs and generate the global invariant features by dividing the convolutional feature maps of the deeper layers of the network into groups. The proposed model for CNNs and G-CNNs, referred as CNN-BoGCFs and G-CNN-BoGCFs, performs significantly high when trained on a small amount of labeled data for image classification. The proposed method is evaluated using rotated MNIST, SIMPLIcity and Oxford flower 17 datasets.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115930222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Zero-Shot Motion Pattern Recognition from 4D Point-Clouds 4D点云的零射击运动模式识别
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596375
Dariush Salami, S. Sigg
{"title":"Zero-Shot Motion Pattern Recognition from 4D Point-Clouds","authors":"Dariush Salami, S. Sigg","doi":"10.1109/mlsp52302.2021.9596375","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596375","url":null,"abstract":"We address a timely and relevant problem in signal processing: The recognition of patterns from spatial data in motion through a zero-shot learning scenario. We introduce a neural network architecture based on Siamese networks to recognize unseen classes of motion patterns. The approach uses a graph-based technique to achieve permutation invariance and also encodes moving point clouds into a representation space in a computationally efficient way. We evaluated the model on an open dataset with twenty-one gestures. The model out-performes state-of-the-art architectures with a considerable margin in four different settings in terms of accuracy while reducing the computational complexity up to 60 times.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"133 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124273970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Compression of DNNs Using Magnitude Pruning and Nonlinear Information Bottleneck Training 基于幅度修剪和非线性信息瓶颈训练的深度神经网络压缩
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596128
Morten Østergaard Nielsen, Jan Østergaard, J. Jensen, Z. Tan
{"title":"Compression of DNNs Using Magnitude Pruning and Nonlinear Information Bottleneck Training","authors":"Morten Østergaard Nielsen, Jan Østergaard, J. Jensen, Z. Tan","doi":"10.1109/mlsp52302.2021.9596128","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596128","url":null,"abstract":"As Deep Neural Networks (DNNs) have achieved state-of-the-art performance in various scientific fields and applications, the memory and computational complexity of DNNs have increased concurrently. The increased complexity required by DNNs prohibits them from running on platforms with limited computational resources. This has sparked a renewed interest in parameter pruning. We propose to replace the standard cross-entropy objective - typically used in classification problems - with the Nonlinear Information Bottleneck (NIB) objective to improve the accuracy of a pruned network. We demonstrate, that our proposal outperforms cross-entropy combined with global magnitude pruning for high compression rates on VGG-nets trained on CIFAR10. With approximately 97% of the parameters pruned, we obtain an accuracy of 87.63% and 88.22% for VGG-16 and VGG-19, respectively, where the baseline accuracy is 91.5% for the unpruned networks. We observe that the majority of biases are pruned completely, and pruning parameters globally outperforms layer-wise pruning.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116953041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
About the Equivalence Between Complex-Valued and Real-Valued Fully Connected Neural Networks - Application to Polinsar Images 关于复值与实值全连接神经网络的等价性——在Polinsar图像中的应用
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596542
J. A. Barrachina, C. Ren, G. Vieillard, C. Morisseau, J. Ovarlez
{"title":"About the Equivalence Between Complex-Valued and Real-Valued Fully Connected Neural Networks - Application to Polinsar Images","authors":"J. A. Barrachina, C. Ren, G. Vieillard, C. Morisseau, J. Ovarlez","doi":"10.1109/mlsp52302.2021.9596542","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596542","url":null,"abstract":"In this paper we provide an exhaustive statistical comparison between Complex-Valued MultiLayer Perceptron (CV-MLP) and Real-Valued MultiLayer Perceptron (RV-MLP) on Oberpfaffenhofen Polarimetric and Interferometric Synthetic Aperture Radar (PolInSAR) database. In order to compare both networks in a fair manner, the need to define the equivalence between the models arises. A novel definition for an equivalent Real-Valued Neural Network (RVNN) is proposed in terms of its real-valued trainable parameters that maintain the aspect ratio and analyze its dynamics. We show that CV-MLP gets a slightly better statistical performance for classification on the PolInSAR image than a capacity equivalent RV-MLP.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"154 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115708228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Distributed Dictionary Learning Over Heterogeneous Clients Using Local Adaptive Dictionaries 使用本地自适应字典的异构客户端分布式字典学习
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596408
You-De Huang, Yao-Win Peter Hong
{"title":"Distributed Dictionary Learning Over Heterogeneous Clients Using Local Adaptive Dictionaries","authors":"You-De Huang, Yao-Win Peter Hong","doi":"10.1109/mlsp52302.2021.9596408","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596408","url":null,"abstract":"This work examines the use of dictionary learning among distributed clients with heterogeneous tasks. We propose a distributed dictionary learning algorithm that enables collaborative training of a shared global dictionary among clients while adaptively constructing local dictionary elements to address the heterogeneity of local tasks. The proposed distributed dictionary learning with local adaptive dictionaries (DDL-LAD) algorithm consists of two parts: a distributed optimization procedure that enables joint training of the dictionaries without sharing of the local datasets with the server, and a splitting and elimination procedure that is used to adaptively construct local dictionary elements. The splitting procedure identifies elements in the global dictionary that exhibit discriminative features for the local tasks. The elements are split and appended to the local dictionaries. Then, to avoid overgrowing of the local dictionaries, an elimination procedure is adopted to prune elements with less usage. Experiments on a distributed EMNIST dataset is provided to demonstrate the effectiveness of the proposed DDL-LAD algorithm compared to existing schemes that adopt only a global shared dictionary.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129055462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward Minimal-Sufficiency in Regression Tasks: An Approach Based on a Variational Estimation Bottleneck 回归任务的最小充分性:基于变分估计瓶颈的方法
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596368
Zhaoyang Lyu, Gholamali Aminian, M. Rodrigues
{"title":"Toward Minimal-Sufficiency in Regression Tasks: An Approach Based on a Variational Estimation Bottleneck","authors":"Zhaoyang Lyu, Gholamali Aminian, M. Rodrigues","doi":"10.1109/mlsp52302.2021.9596368","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596368","url":null,"abstract":"We propose a new variational estimation bottleneck based on a mean-squared error metric to aid regression tasks. In particular, this bottleneck - which draws inspiration from a variational information bottleneck for classification counterparts - consists of two components: (1) one captures the notion of $mathcal{V}_{r}$ -sufficiency that quantifies the ability for an estimator in some class of estimators $mathcal{V}_{r}$ to infer the quantity of interest; (2) the other component appears to capture a notion of $mathcal{V}_{r}$ - minimality that quantifies the ability of the estimator to generalize to new data. We demonstrate how to train this bottleneck for regression problems. We also conduct various experiments in image denoising and deraining applications showcasing that our proposed approach can lead to neural network regressors offering better performance without suffering from overfitting.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130244638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hyphylearn: A Domain Adaptation-Inspired Approach to Classification Using Limited Number of Training Samples Hyphylearn:一种基于领域自适应的有限训练样本分类方法
2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP) Pub Date : 2021-10-25 DOI: 10.1109/mlsp52302.2021.9596469
Alireza Nooraiepour, W. Bajwa, N. Mandayam
{"title":"Hyphylearn: A Domain Adaptation-Inspired Approach to Classification Using Limited Number of Training Samples","authors":"Alireza Nooraiepour, W. Bajwa, N. Mandayam","doi":"10.1109/mlsp52302.2021.9596469","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596469","url":null,"abstract":"The fundamental task of classification given a limited number of training data samples is considered for physical systems with known parametric statistical models. As a solution, a hybrid classification method-termed HYPHYLEARN-is proposed that exploits both the physics-based statistical models and the learning-based classifiers. The proposed solution is based on the conjecture that HYPHYLEARN would alleviate the challenges associated with the individual approaches of learning-based and statistical model-based classifiers by fusing their respective strengths. The proposed hybrid approach first estimates the unobservable model parameters using the available (suboptimal) statistical estimation procedures, and subsequently uses the physics-based statistical models to generate synthetic data. Next, the training data samples are incorporated with the synthetic data in a learning-based classifier that is based on domain-adversarial training of neural networks. Numerical results on multiuser detection, a concrete communication problem, demonstrate that HYPHYLEARN leads to major classification improvements compared to the existing stand-alone and hybrid classification methods.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121561557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信