{"title":"Real-Time Optimizing Weighted Gaussian Curvature for 4K Videos","authors":"Wenming Tang, Lebin Zhou, Yuanhao Gong","doi":"10.1109/mlsp52302.2021.9596473","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596473","url":null,"abstract":"Weighted Gaussian curvature is an important measurement for surfaces, images and videos. However, since it is a second order quantity, its optimization usually leads to high order geometric flows that cause difficulties for practical applications. In this paper, we propose a novel optimization method for weighted Gaussian curvature. Our method does not require the image (video) to be second-order differentiable, thus, avoiding the high order geometric flows. In addition, we propose a new 4-D look-up table method to accelerate the optimization of weighted Gaussian curvature. Therefore, our algorithm is very efficient and can achieve real-time processing for high resolution videos. For example, our method can process 4K videos with 50 frames per second on a single graphic card (NVIDIA 3090). Several numerical experiments are carried out to confirm the efficiency and effectiveness of the proposed method. Thanks to the high performance, our method can be applied in a large range of applications that involve weighted Gaussian curvature, such as image restoration, registration, enhancement, etc.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124651925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Introducing K-Anonymity Principles to Adversarial Attacks for Privacy Protection in Image Classification Problems","authors":"V. Mygdalis, A. Tefas, I. Pitas","doi":"10.1109/mlsp52302.2021.9596565","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596565","url":null,"abstract":"The network output activation values for a given input can be employed to produce a sorted ranking. Adversarial attacks typically generate the least amount of perturbation required to change the classifier label. In that sense, generated adversarial attack perturbation only affects the output in the 1st sorted ranking position. We argue that meaningful information about the adversarial examples i.e., their original labels, is still encoded in the network output ranking and could potentially be extracted, using rule-based reasoning. To this end, we introduce a novel adversarial attack methodology inspired by the K-anonymity principles, that generates adversarial examples that are not only misclassified, but their output sorted ranking spreads uniformly along K different positions. Any additional perturbation arising from the strength of the proposed objectives, is regularized by a visual similarity-based term. Experimental results denote that the proposed approach achieves the optimization goals inspired by K-anonymity with reduced perturbation as well.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132104106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Heitor R. Guimarães, Wesley Beccaro, M. A. Ramírez
{"title":"Optimizing Time Domain Fully Convolutional Networks for 3D Speech Enhancement in a Reverberant Environment Using Perceptual Losses","authors":"Heitor R. Guimarães, Wesley Beccaro, M. A. Ramírez","doi":"10.1109/mlsp52302.2021.9596103","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596103","url":null,"abstract":"Noise in 3D reverberant environment is detrimental to several downstream applications. In this work, we propose a novel approach to 3D speech enhancement directly in the time domain through the usage of Fully Convolutional Networks (FCN) with a custom loss function based on the combination of a perceptual loss, built on top of the wav2vec model and a soft version of the short-time objective intelligibility (STOI) metric. The dataset and experiments were based on Task 1 of the L3DAS21 challenge. Our model achieves a STOI score of 0.82, word error rate (WER) equal to 0.36, and a score of 0.73 in the metric proposed by the challenge based on STOI and WER combination using as reference the development set. Our submission, based on this method, was ranked second in Task 1 of the L3DAS21 challenge.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127394501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Bag of Groups of Convolutional Features Model for Visual Object Recognition","authors":"Jaspreet Singh, Chandan Singh","doi":"10.1109/mlsp52302.2021.9596432","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596432","url":null,"abstract":"Deep convolutional neural networks (CNNs) are only equivariant to translation. Recently, equivariant CNNs are proposed for the task of image classification which are not only equivariant to translation but also to other affine geometric transformations. Moreover, CNNs and equivariant CNNs require a large amount of labeled training data to generalize its parameters which also limit their application areas. We propose a bag of groups of convolutional features (BoGCFs) model for the CNNs and group-equivariant CNNs (G-CNNs)[1], which preserves the fundamental property of equivariance of G-CNNs and generate the global invariant features by dividing the convolutional feature maps of the deeper layers of the network into groups. The proposed model for CNNs and G-CNNs, referred as CNN-BoGCFs and G-CNN-BoGCFs, performs significantly high when trained on a small amount of labeled data for image classification. The proposed method is evaluated using rotated MNIST, SIMPLIcity and Oxford flower 17 datasets.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115930222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Zero-Shot Motion Pattern Recognition from 4D Point-Clouds","authors":"Dariush Salami, S. Sigg","doi":"10.1109/mlsp52302.2021.9596375","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596375","url":null,"abstract":"We address a timely and relevant problem in signal processing: The recognition of patterns from spatial data in motion through a zero-shot learning scenario. We introduce a neural network architecture based on Siamese networks to recognize unseen classes of motion patterns. The approach uses a graph-based technique to achieve permutation invariance and also encodes moving point clouds into a representation space in a computationally efficient way. We evaluated the model on an open dataset with twenty-one gestures. The model out-performes state-of-the-art architectures with a considerable margin in four different settings in terms of accuracy while reducing the computational complexity up to 60 times.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"133 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124273970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Morten Østergaard Nielsen, Jan Østergaard, J. Jensen, Z. Tan
{"title":"Compression of DNNs Using Magnitude Pruning and Nonlinear Information Bottleneck Training","authors":"Morten Østergaard Nielsen, Jan Østergaard, J. Jensen, Z. Tan","doi":"10.1109/mlsp52302.2021.9596128","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596128","url":null,"abstract":"As Deep Neural Networks (DNNs) have achieved state-of-the-art performance in various scientific fields and applications, the memory and computational complexity of DNNs have increased concurrently. The increased complexity required by DNNs prohibits them from running on platforms with limited computational resources. This has sparked a renewed interest in parameter pruning. We propose to replace the standard cross-entropy objective - typically used in classification problems - with the Nonlinear Information Bottleneck (NIB) objective to improve the accuracy of a pruned network. We demonstrate, that our proposal outperforms cross-entropy combined with global magnitude pruning for high compression rates on VGG-nets trained on CIFAR10. With approximately 97% of the parameters pruned, we obtain an accuracy of 87.63% and 88.22% for VGG-16 and VGG-19, respectively, where the baseline accuracy is 91.5% for the unpruned networks. We observe that the majority of biases are pruned completely, and pruning parameters globally outperforms layer-wise pruning.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116953041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. A. Barrachina, C. Ren, G. Vieillard, C. Morisseau, J. Ovarlez
{"title":"About the Equivalence Between Complex-Valued and Real-Valued Fully Connected Neural Networks - Application to Polinsar Images","authors":"J. A. Barrachina, C. Ren, G. Vieillard, C. Morisseau, J. Ovarlez","doi":"10.1109/mlsp52302.2021.9596542","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596542","url":null,"abstract":"In this paper we provide an exhaustive statistical comparison between Complex-Valued MultiLayer Perceptron (CV-MLP) and Real-Valued MultiLayer Perceptron (RV-MLP) on Oberpfaffenhofen Polarimetric and Interferometric Synthetic Aperture Radar (PolInSAR) database. In order to compare both networks in a fair manner, the need to define the equivalence between the models arises. A novel definition for an equivalent Real-Valued Neural Network (RVNN) is proposed in terms of its real-valued trainable parameters that maintain the aspect ratio and analyze its dynamics. We show that CV-MLP gets a slightly better statistical performance for classification on the PolInSAR image than a capacity equivalent RV-MLP.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"154 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115708228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Distributed Dictionary Learning Over Heterogeneous Clients Using Local Adaptive Dictionaries","authors":"You-De Huang, Yao-Win Peter Hong","doi":"10.1109/mlsp52302.2021.9596408","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596408","url":null,"abstract":"This work examines the use of dictionary learning among distributed clients with heterogeneous tasks. We propose a distributed dictionary learning algorithm that enables collaborative training of a shared global dictionary among clients while adaptively constructing local dictionary elements to address the heterogeneity of local tasks. The proposed distributed dictionary learning with local adaptive dictionaries (DDL-LAD) algorithm consists of two parts: a distributed optimization procedure that enables joint training of the dictionaries without sharing of the local datasets with the server, and a splitting and elimination procedure that is used to adaptively construct local dictionary elements. The splitting procedure identifies elements in the global dictionary that exhibit discriminative features for the local tasks. The elements are split and appended to the local dictionaries. Then, to avoid overgrowing of the local dictionaries, an elimination procedure is adopted to prune elements with less usage. Experiments on a distributed EMNIST dataset is provided to demonstrate the effectiveness of the proposed DDL-LAD algorithm compared to existing schemes that adopt only a global shared dictionary.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129055462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward Minimal-Sufficiency in Regression Tasks: An Approach Based on a Variational Estimation Bottleneck","authors":"Zhaoyang Lyu, Gholamali Aminian, M. Rodrigues","doi":"10.1109/mlsp52302.2021.9596368","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596368","url":null,"abstract":"We propose a new variational estimation bottleneck based on a mean-squared error metric to aid regression tasks. In particular, this bottleneck - which draws inspiration from a variational information bottleneck for classification counterparts - consists of two components: (1) one captures the notion of $mathcal{V}_{r}$ -sufficiency that quantifies the ability for an estimator in some class of estimators $mathcal{V}_{r}$ to infer the quantity of interest; (2) the other component appears to capture a notion of $mathcal{V}_{r}$ - minimality that quantifies the ability of the estimator to generalize to new data. We demonstrate how to train this bottleneck for regression problems. We also conduct various experiments in image denoising and deraining applications showcasing that our proposed approach can lead to neural network regressors offering better performance without suffering from overfitting.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130244638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hyphylearn: A Domain Adaptation-Inspired Approach to Classification Using Limited Number of Training Samples","authors":"Alireza Nooraiepour, W. Bajwa, N. Mandayam","doi":"10.1109/mlsp52302.2021.9596469","DOIUrl":"https://doi.org/10.1109/mlsp52302.2021.9596469","url":null,"abstract":"The fundamental task of classification given a limited number of training data samples is considered for physical systems with known parametric statistical models. As a solution, a hybrid classification method-termed HYPHYLEARN-is proposed that exploits both the physics-based statistical models and the learning-based classifiers. The proposed solution is based on the conjecture that HYPHYLEARN would alleviate the challenges associated with the individual approaches of learning-based and statistical model-based classifiers by fusing their respective strengths. The proposed hybrid approach first estimates the unobservable model parameters using the available (suboptimal) statistical estimation procedures, and subsequently uses the physics-based statistical models to generate synthetic data. Next, the training data samples are incorporated with the synthetic data in a learning-based classifier that is based on domain-adversarial training of neural networks. Numerical results on multiuser detection, a concrete communication problem, demonstrate that HYPHYLEARN leads to major classification improvements compared to the existing stand-alone and hybrid classification methods.","PeriodicalId":156116,"journal":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121561557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}