{"title":"A Novel Multi-Scale Residual Dense Dehazing Network (MSRDNet) for Single Image Dehazing✱","authors":"Chippy M. Manu, G. SreeniK.","doi":"10.1145/3571600.3571601","DOIUrl":"https://doi.org/10.1145/3571600.3571601","url":null,"abstract":"Dehazing is a difficult process because of the damage caused by the non-uniform fog and haze distribution in images. To address these issues, a Multi-Scale Residual dense Dehazing Network (MSRDNet) is proposed in this paper. A Contextual feature extraction module (CFM) for extracting multi-scale features and an Adaptive Residual Dense Module (ARDN) are used as sub-modules of MSRDNet. Moreover, all the hierarchical features extracted by each ARDN are fused, which helps to detect hazy maps of varying lengths with multi-scale features. This framework outperforms the state-of-the-art dehazing methods in removing haze while maintaining and restoring image detail in real-world and synthetic images captured under various scenarios.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"122 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78584695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust Brain State Decoding using Bidirectional Long Short Term Memory Networks in functional MRI.","authors":"Anant Mittal, Priya Aggarwal, Luiz Pessoa, Anubha Gupta","doi":"10.1145/3490035.3490269","DOIUrl":"10.1145/3490035.3490269","url":null,"abstract":"<p><p>Decoding brain states of the underlying cognitive processes via learning discriminative feature representations has recently gained a lot of interest in brain imaging studies. Particularly, there has been an impetus to encode the dynamics of brain functioning by analyzing temporal information available in the fMRI data. Long-short term memory (LSTM), a class of machine learning model possessing a \"memory\" component, to retain previously seen temporal information, is increasingly being observed to perform well in various applications with dynamic temporal behavior, including brain state decoding. Because of the dynamics and inherent latency in fMRI BOLD responses, future temporal context is crucial. However, it is neither encoded nor captured by the conventional LSTM model. This paper performs robust brain state decoding via information encapsulation from both the past and future instances of fMRI data via bi-directional LSTM. This allows for explicitly modeling the dynamics of BOLD response without any delay adjustment. To this end, we utilize a bidirectional LSTM, wherein, the input sequence is fed in normal time-order for one LSTM network, and in the reverse time-order, for another. The two hidden activations of forward and reverse directions in bi-LSTM are collated to build the \"memory\" of the model and are used to robustly predict the brain states at every time instance. Working memory data from the Human Connectome Project (HCP) is utilized for validation and was observed to perform 18% better than it's unidirectional counterpart in terms of accuracy in predicting the brain states.</p>","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9639335/pdf/nihms-1789845.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40673961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ICVGIP 2018: 11th Indian Conference on Computer Vision, Graphics and Image Processing, Hyderabad, India, 18-22 December, 2018","authors":"","doi":"10.1145/3293353","DOIUrl":"https://doi.org/10.1145/3293353","url":null,"abstract":"","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81854985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bharti, A. Juneja, M. Saxena, S. Gudwani, S. Kumaran, R. Agrawal, M. Behari
{"title":"Voxel-based morphometry and minimum redundancy maximum relevance method for classification of Parkinson's disease and controls from T1-weighted MRI","authors":"Bharti, A. Juneja, M. Saxena, S. Gudwani, S. Kumaran, R. Agrawal, M. Behari","doi":"10.1145/3009977.3009998","DOIUrl":"https://doi.org/10.1145/3009977.3009998","url":null,"abstract":"Parkinson's disease (PD) is a neurodegenerative disorder, which needs to be accurately diagnosed in early stage. Voxel-based morphometry (VBM) has been extensively utilized to determine focal changes between PD patients and controls. However, it is not much utilized in differential diagnosis of an individual subject. Thus, in this study, VBM findings in conjunction with minimum redundancy maximum relevance (mRMR) method are utilized to obtain a set of relevant and non-redundant features for computer-aided diagnosis (CAD) of PD using T1-weighted MRI. In the proposed method, firstly, statistical features are extracted from the clusters obtained from statistical maps, generated using VBM, of gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF) independently and their different combinations. Then mRMR, a multivariate feature selection method, is utilized to find a minimal set of relevant and non-redundant features. Finally, support vector machine is utilized to learn a decision model using the selected features. Experiments are performed on newly acquired T1-weighted MRI of 30 PD patients and 30 age & gender matched controls. The performance is evaluated using leave-one out cross-validation scheme in terms of sensitivity, specificity and classification accuracy. The maximum accuracy of 88.33% is achieved for GM+WM and GM+WM+CSF. In addition, the proposed method outperforms the existing methods. It is also observed that the selected clusters belong to regions namely middle and superior frontal gyrus for GM, inferior, middle frontal gyrus and insula for WM and lateral ventricle for CSF. Further, correlation of UPDRS/H&Y staging scale with GM/WM/CSF volume is observed to be not significant. Appreciable classification performance of the proposed method highlights the potential of the proposed method in CAD support system for the clinicians in PD diagnosis.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"4 1","pages":"22:1-22:6"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73493078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Babu Kishore Subramanian, Ashutosh Gupta, C. Seelamantula
{"title":"A distribution-independent risk estimator for image denoising","authors":"Babu Kishore Subramanian, Ashutosh Gupta, C. Seelamantula","doi":"10.1145/3009977.3010025","DOIUrl":"https://doi.org/10.1145/3009977.3010025","url":null,"abstract":"We address the problem of image denoising for an additive white noise model without placing any restrictions on the statistical distribution of noise. We assume knowledge of only the first- and second-order noise statistics. In the recent mean-square error (MSE) minimization approaches for image denoising, one considers a particular noise distribution and derives an expression for the unbiased risk estimate of the MSE. For additive white Gaussian noise, an unbiased estimate of the MSE is Stein's unbiased risk estimate (SURE), which relies on Stein's lemma. We derive an unbiased risk estimate without using Stein's lemma or its counterparts for additive white noise model irrespective of the noise distribution. We refer to the MSE estimate as the generic risk estimate (GenRE). We demonstrate the effectiveness of GenRE using shrinkage in the undecimated Haar wavelet transform domain as the denoising function. The estimated peak-signal-to-noise-ratio (PSNR) using GenRE is typically within 1% of the PSNR obtained when optimizing with the oracle MSE. The performance of the proposed method is on par with SURE for Gaussian noise distribution, and better than SURE-based methods for other noise distributions such as uniform and Laplacian distribution in terms of both PSNR and structural similarity (SSIM).","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"220 1","pages":"52:1-52:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75624899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Informed multimodal latent subspace learning via supervised matrix factorization","authors":"Ramashish Gaurav, Mridula Verma, K. K. Shukla","doi":"10.1145/3009977.3010012","DOIUrl":"https://doi.org/10.1145/3009977.3010012","url":null,"abstract":"Matrix factorization technique has been widely used as a popular method to learn a joint latent-compact subspace, when multiple views or modals of objects (belonging to single-domain or multiple-domain) are available. Our work confronts the problem of learning an informative latent subspace by imparting supervision to matrix factorization for fusing multiple modals of objects, where we devise simpler supervised additive updates instead of multiplicative updates, thus scalable to large scale datasets. To increase the classification accuracy we integrate the label information of images with the process of learning a semantically enhanced subspace. We perform extensive experiments on two publicly available standard image datasets of NUS WIDE and compare the results with state-of-the-art subspace learning and fusion techniques to evaluate the efficacy of our framework. Improvement obtained in the classification accuracy confirms the effectiveness of our approach. In essence, we propose a novel method for supervised data fusion thus leading to supervised subspace learning.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"76 1","pages":"36:1-36:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78214159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. NimishaT., K. Seemakurthy, A. Rajagopalan, N. Vedachalam, Ramesh Raju
{"title":"Color restoration in turbid medium","authors":"M. NimishaT., K. Seemakurthy, A. Rajagopalan, N. Vedachalam, Ramesh Raju","doi":"10.1145/3009977.3010028","DOIUrl":"https://doi.org/10.1145/3009977.3010028","url":null,"abstract":"Light scattering and color distortions are two major issues with underwater imaging. Scattering occurs due to turbidity of the medium and color distortions are caused by differential attenuation of wavelengths as a function of depth. As a result, underwater images taken in a turbid medium have low contrast, color cast, and color loss. The main objective of this work is color restoration of underwater images i.e, produce its equivalent image as seen outside of the water surface. As a first step, we account for low contrast by employing dark channel prior based dehazing. These images are then color corrected by learning a mapping function between a pair of color chart images, one taken inside water and another taken outside. The mapping thus learned is with respect to a reference distance from the water surface. We also propose a color modulation scheme that is applied prior to color mapping to accommodate the same mapping function for different depths as well. Color restoration results are given on several images to validate the efficacy of the proposed methodology.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"14 1","pages":"26:1-26:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78483761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A framework to assess Sun salutation videos","authors":"Hiteshi Jain, Gaurav Harit","doi":"10.1145/3009977.3010045","DOIUrl":"https://doi.org/10.1145/3009977.3010045","url":null,"abstract":"There are many exercises which are repetitive in nature and are required to be done with perfection to derive maximum benefits. Sun Salutation or Surya Namaskar is one of the oldest yoga practice known. It is a sequence of ten actions or 'asanas' where the actions are synchronized with breathing and each action and its transition should be performed with minimal jerks. Essentially, it is important that this yoga practice be performed with Grace and Consistency. In this context, Grace is the ability of a person to perform an exercise with smoothness i.e. without sudden movements or jerks during the posture transition and Consistency measures the repeatability of an exercise in every cycle. We propose an algorithm that assesses how well a person practices Sun Salutation in terms of grace and consistency. Our approach works by training individual HMMs for each asana using STIP features[11] followed by automatic segmentation and labeling of the entire Sun Salutation sequence using a concatenated-HMM. The metric of grace and consistency are then laid down in terms of posture transition times. The assessments made by our system are compared with the assessments of the yoga trainer to derive the accuracy of the system. We introduce a dataset for Sun Salutation videos comprising 30 sequences of perfect Sun Salutation performed by seven experts and used this dataset to train our system. While Sun Salutation can be judged on multiple parameters, we focus mainly on judging Grace and Consistency.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"43 1","pages":"29:1-29:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86528823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Susheel Suresh, Tarun Sharma, K. PrashanthT., V. Subramaniam, D. Sitaram, M. Nirupama
{"title":"Towards quantifying the amount of uncollected garbage through image analysis","authors":"Susheel Suresh, Tarun Sharma, K. PrashanthT., V. Subramaniam, D. Sitaram, M. Nirupama","doi":"10.1145/3009977.3010061","DOIUrl":"https://doi.org/10.1145/3009977.3010061","url":null,"abstract":"Civic authorities in many Indian cities have a tough time in garbage collection and as a result there is a pile up of garbage in the cities. In order to manage the situation, it is first required to be able to quantify the issue. In this paper, we address the problem of quantification of garbage in a dump using a two step approach. In the first step, we build a mobile application that allows citizens to capture images of garbage and upload them to a server. In the second step, back-end performs analysis on these images to estimate the amount of garbage using computer vision techniques. Our approach to volume estimation uses multiple images of the same dump (provided by the mobile application) from different perspectives, segments the dump from the background, reconstructs a three dimensional view of the dump and then estimates its volume. Using our novel pipeline, our experiments indicate that with 8 different perspectives, we are able to achieve an accuracy of about 85 % for estimating the volume.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"26 1","pages":"73:1-73:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82623766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Generating synthetic handwriting using n-gram letter glyphs","authors":"Arka Ujjal Dey, Gaurav Harit","doi":"10.1145/3009977.3010042","DOIUrl":"https://doi.org/10.1145/3009977.3010042","url":null,"abstract":"We propose a framework for synthesis of natural semi cursive handwritten Latin script that can find application in text personalization, or in generation of synthetic data for recognition systems. Our method is based on the generation of synthetic n-gram letter glyphs and their subsequent concatenation. We propose a non-parametric data driven generation scheme that is able to mimic the variation observed in handwritten glyph samples to synthesize natural looking synthetic glyphs. These synthetic glyphs are then stitched together to form complete words, using a spline based concatenation scheme. Further, as a refinement, our method is able to generate pen-lifts, giving our results a natural semi-cursive look. Through subjective experiments and detailed analysis of the results, we demonstrate the effectiveness of our formulation in being able to generate natural looking synthetic script.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"27 1","pages":"49:1-49:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89111162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}