{"title":"Face Recognition against Mouth Shape Variations","authors":"Mustafa M. Alrjebi, Wanquan Liu, Ling Li","doi":"10.1109/DICTA.2015.7371259","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371259","url":null,"abstract":"In this paper, face recognition against mouth shape variations is investigated. In order to detect possible mouth variations, the inner mouth landmarks are first detected by a landmark detector and then used to estimate the connectivity between the upper lip and the lower lip of a face image. The vertical distance between the middle inner points of upper lip and the lower lip is calculated, and then used with appropriate threshold to decide whether the two lips are connected or separated. If the two lips are not connected, we further detect the teeth positions based on the colour pixel values, and then a face can be classified into four classes: closed mouth (C), closed mouth with teeth (Ct), open mouth (O), and open mouth with teeth (Ot). Next we attempt to transform face images in classes Ct and Ot into classes C and O respectively, and this is done by shrinking the areas with the upper and lower parts of the mouth by a proposed alignment approach. In this mouth closing process, both face areas located above and below the mouth are changed dramatically and the whole face image is vertically stretched to the original size in order to change the face image into neutral appearance. Extensive experiments on AR database and BU database show that the proposed shape correction approach to closing an opened mouth can significantly increase the recognition rate up to 21.5% by using PCA and 17.5% by using LDA.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134140618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Learning Efficiently- The Deep CNNs-Tree Network","authors":"Fu-Chun Hsu, J. Gubbi, M. Palaniswami","doi":"10.1109/DICTA.2015.7371277","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371277","url":null,"abstract":"In recent years, deep feature learning has been successfully applied in many fields such as visual recognition, speech recognition, and natural language processing. Based on the recent rapid development in deep learning community, applying Convolutional Neural Network (CNN) has impacted several fields. However, the number of parameters required to develop a sophisticated large CNN model becomes a problem. We aimed at this problem and presented the Deep CNNs-Tree Network model as our solution. By clustering similar channel features in the feature maps, we were able to create a tree of CNNs and replace the original CNN layer with the proposed model. Experiments on popular image datasets, the MNIST and CIFAR-10, has shown that the proposed network achieve similar performance of accuracy when compared to the traditional CNN, and only less than 5% of accuracy loss. A reduction of more than 70% parameters was observed using the proposed method.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"285 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114954613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jack Valmadre, S. Sridharan, S. Denman, C. Fookes, S. Lucey
{"title":"Closed-Form Solutions for Low-Rank Non-Rigid Reconstruction","authors":"Jack Valmadre, S. Sridharan, S. Denman, C. Fookes, S. Lucey","doi":"10.1109/DICTA.2015.7371247","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371247","url":null,"abstract":"Recovering the motion of a non-rigid body from a set of monocular images permits the analysis of dynamic scenes in uncontrolled environments. However, the extension of factorisation algorithms for rigid structure from motion to the low-rank non- rigid case has proved challenging. This stems from the comparatively hard problem of finding a linear ``corrective transform'' which recovers the projection and structure matrices from an ambiguous factorisation. We elucidate that this greater difficulty is due to the need to find multiple solutions to a non-trivial problem, casting a number of previous approaches as alleviating this issue by either a) introducing constraints on the basis, making the problems non- identical, or b) incorporating heuristics to encourage a diverse set of solutions, making the problems inter-dependent. While it has previously been recognised that finding a single solution to this problem is sufficient to estimate cameras, we show that it is possible to bootstrap this partial solution to find the complete transform in closed-form. However, we acknowledge that our method minimises an algebraic error and is thus inherently sensitive to deviation from the low-rank model. We compare our closed-form solution for non-rigid structure with known cameras to the closed-form solution of Dai et al.~cite{Dai2012}, which we find to produce only coplanar reconstructions. We therefore make the recommendation that 3D reconstruction error always be measured relative to a trivial reconstruction such as a planar one.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116769693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Quick Atmospheric Correction (QUAC) Algorithm for Hyperspectral Image Processing: Extending QUAC to a Coastal Scene","authors":"S. Carr, L. Bernstein, S. Adler-Golden","doi":"10.1109/DICTA.2015.7371314","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371314","url":null,"abstract":"The quick atmospheric correction (QUAC) algorithm is a relatively fast and robust atmospheric compensation algorithm for hyperspectral image processing utilizing in scene information. An adjustment of some key parameters in QUAC is made leading to improved results for coastal scenes. In general the QUAC results compare well with two first principles radiative transfer (RT) model based algorithms. Some suggestions for future work are made including automating the setting of key QUAC parameters and accounting for the coastal zone aerosols more accurately in the RT algorithms.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"200 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134041548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Retinal Minimum Distance Band (MDB)Computation from SD-OCT Images","authors":"Md. Akter Hussain, A. Bhuiyan, K. Ramamohanarao","doi":"10.1109/DICTA.2015.7371238","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371238","url":null,"abstract":"In this paper, we have proposed a novel automatic method to determine the Minimum Distance Band (MDB) from the retinal SD-OCT image. MDB is the minimum distance between the retinal pigment epithelium (RPE) layer and the Optic Nerve Head (ONH) surface. It is an effective biomarker for early detection of glaucoma. Our proposed method is the first automatic method for computing the Minimum Distance Band (MDB). This method uses the approximate location of three benchmark reference (TBMR) layers of the retina that help to reduce the search space. This approach is highly accurate in detecting the boundary layers even in the presence of pathology. The terminal points of the RPE serve as the ONH region. The accuracies of ONH and MDB are tested against 13 manually graded optic disc centred SD-OCT volumes (11 B-scan per volume) of Glaucoma patients as a gold standard. Our method is very effective with mean and standard deviation of the error of the ONH width and MDB of 5.36±5.55 and 14.96±17.75 respectively. The precision, recall and F1 score for the ONH region are 94.90, 98.14 and 96.49 respectively.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124870436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Relative Depth Estimation from Hyperspectral Data","authors":"Ali Zia, J. Zhou, Yongsheng Gao","doi":"10.1109/DICTA.2015.7371299","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371299","url":null,"abstract":"This paper addresses the problem of relative depth estimation using spatial defocus and spectral chromatic aberration presented in hyperspectral data. Our approach produces merged relative sparse depth map using two different methods. The first method constructs a histogram descriptor for edge pixels in each spectral band image. Due to the spectral chromatic aberration, different edge statistical information can be extracted from each band even at the same location. Variance among histogram bins provides input data for band-wise spatial defocus calculation. These band-wise statistical data are later combined to give the first sparse depth map. The second approach uses difference of neighboring spectral vectors to estimate relative depth. The two sparse maps with distinguishing features are finally combined and optimized to generate final sparse depth map. During the last step, normalization and smoothing are used to guarantee better consistency among edge pixels. Experimental results show that our method can generate better sparse depth map than alternative methods which operate on RGB images.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127323592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Performance Evaluation of a Newly Proposed Novel L-SPECT System for SPECT Imaging","authors":"Tasneem Rahman, M. Tahtali, M. Pickering","doi":"10.1109/DICTA.2015.7371294","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371294","url":null,"abstract":"The importance of pre-clinical research of single photon emission computed tomography(SPECT) imaging is now widely recognized. For the demand of high resolution and high detection efficiency, SPECT has been developed in several ways and it is simultaneously making demand for high quality imaging. This paper introduces the newly proposed dual-head L-SPECT system and investigates the initial performance of the system having an array of pinholes as a collimator. The proposed L-SPECT system is based on the concept of light field imaging allowing the refocusing of images after exposure. A microlens array is placed just before the imaging sensor to simultaneously record the direction of incident light rays and their intensities. We are proposing a detector module with 48mm by 48mm of active area behind an array of 100×100 pinholes for gamma rays instead of microlenses. The system is based on a pixelated array of NaI crystals (10×10×10 mm elements) coupled with an array of position sensitive photomultiplier tubes (PSPMTs). The basic characteristics of this system were evaluated with pinhole radii of 50μm, 60μm and 100μm. The measurements of system sensitivity, system spatial resolution, energy resolution, volume sensitivity and uniformity were evaluated for 99mTc (140keV) solution where reconstructed images are well visualized. Monte Carlo simulation studies using the Geant4 Application for Tomographic Emission (GATE) software package validate the performance of this novel dual head L-SPECT where a general cylindrical water phantom is used to evaluate its performance. The analysis results show the combination of excellent spatial resolution and high detection efficiency over an energy range between 20-160 keV.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129086006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Bio-Cell Image Segmentation Using Bayes Graph-Cut Model","authors":"Maedeh Beheshti, J. Faichney, A. Gharipour","doi":"10.1109/DICTA.2015.7371241","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371241","url":null,"abstract":"The accurate segmentation of biomedical images has become increasingly important for recognizing cells that have the phenotype of interest in biomedical applications. In order to improve the conventional deterministic segmentation models, this paper proposes a novel graph-cut cell image segmentation algorithm based on Bayes theorem. There are two segmentation phases in this method. The first phase is an interactive process to specify a preliminary set of regional pixels and the background based on the interactive graph-cut model. In the second phase, final segmentation is calculated based on the idea of Bayes theorem, combining prior information with data. Our idea can be considered an integration of graph-cut methods and Bayes theorem for cell image segmentation. Experimental results show that the proposed model performs better in comparison with several existing methods.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127985089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Plane-Tree Low-Bitrate Mesh Compression","authors":"Luke Lincoln, R. González","doi":"10.1109/DICTA.2015.7371295","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371295","url":null,"abstract":"In this paper we present a novel low-bitrate 3D model compression algorithm called the Plane-Tree. This algorithm is based on octree subdivision and stores a plane at each leaf node which better approximates the surface within the node. Quantitative evaluations show that this method is competitive with state of the art transform based methods and outperforms them at low-bitrates.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124327917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hayden Faulkner, Ergnoor Shehu, Zygmunt L. Szpak, W. Chojnacki, J. Tapamo, A. Dick, A. Hengel
{"title":"A Study of the Region Covariance Descriptor: Impact of Feature Selection and Image Transformations","authors":"Hayden Faulkner, Ergnoor Shehu, Zygmunt L. Szpak, W. Chojnacki, J. Tapamo, A. Dick, A. Hengel","doi":"10.1109/DICTA.2015.7371222","DOIUrl":"https://doi.org/10.1109/DICTA.2015.7371222","url":null,"abstract":"We analyse experimentally the region covariance descriptor which has proven useful in numerous computer vision applications. The properties of the descriptor--despite its widespread deployment--are not well understood or documented. In an attempt to uncover key attributes of the descriptor, we characterise the interdependence between the choice of features and distance measures through a series of meticulously designed and performed experiments. Our results paint a rather complex picture and underscore the necessity for more extensive empirical and theoretical work. In light of our findings, there is reason to believe that the region covariance descriptor will prove useful for methods that perform image super-resolution, deblurring, and denoising based on matching and retrieval of image patches from an image dictionary.","PeriodicalId":214897,"journal":{"name":"2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126614843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}