Mengyao Yang, K. Xie, Tong Li, Yonghua Ye, Zepeng Yang
{"title":"Color constancy using AlexNet convolutional neural network","authors":"Mengyao Yang, K. Xie, Tong Li, Yonghua Ye, Zepeng Yang","doi":"10.1117/12.2604686","DOIUrl":"https://doi.org/10.1117/12.2604686","url":null,"abstract":"Color constancy usually refers to the adaptive ability of people to correctly perceive the color of objects under any light source and is an important prerequisite for advanced tasks such as recognition, segmentation and 3D vision. The purpose of color constancy calculation is to estimate the illumination color of the image. In this work, we established the Alexnet network model to accurately estimate the lighting in the scene. The AlexNet model includes an input layer, 8 convolutional layers, AlexNet takes a 512x512 3-channel image patch as input. Compared with the previous network models, the AlexNet model contains several relatively new technical points. For the first time, ReLU, Dropout have been successfully applied in CNN. At the same time, AlexNet also uses GPU for computing acceleration. The illumination color estimation is more robust and stable, and can be combined with the field of color correction of image processing and computer vision.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"26 1","pages":"119130M - 119130M-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76864643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detection of Wagyu beef sources with image classification using convolutional neural network","authors":"Nattakorn Kointarangkul, Y. Limpiyakorn","doi":"10.1117/12.2604971","DOIUrl":"https://doi.org/10.1117/12.2604971","url":null,"abstract":"Wagyu beef originated in Japan. There are many types of Wagyu beef in the market around the globe, though. Primary sources may include Australia, the United States of America, Canada, and the United Kingdom. The authentic Japanese Wagyu is well known for its intense marbling, juicy rich flavor, and tenderness. And there are differences in flavor, texture, and quality between the different types of Wagyu. Nowadays, there is a growing interest in deep learning as a remarkable solution for several domain problems such as computer vision and image classification. In this study, we thus present an AI-based approach to identifying Wagyu beef sources with image classification. A deep neural network, CNN, was constructed to detect the marbled fat patterns of two sources, Japanese Wagyu and Australian Wagyu. The images were collected from reliable sources on the internet and augmented with DCGAN. The prediction of Wagyu sources achieved high accuracy of 95%. The learning model of Convolutional Neural Networks was found to be a promising method for the rapid characterization of the unique patterns of marbled fat layers. The classifier would benefit the customers for buying what they expect from the products in terms of quality and taste.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"23 1","pages":"119130J - 119130J-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76719291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The generalized covariance union fusion approach for distributed sensors with different fields of view","authors":"Feng Ma, Huan-zhang Lu, Luping Zhang, Xinglin Sheng","doi":"10.1117/12.2604692","DOIUrl":"https://doi.org/10.1117/12.2604692","url":null,"abstract":"This paper proposes a generalized covariance union(GCU) approach to solve the distributed fusion problem of sensors with different fields of view (FoVs). It uses the fusion results within the intersection of the FoVs (IoF)to estimate the(target positioning) measurement error, and then employs this estimated error to correct the multitarget densities outside the IoF. Compared with the current approach, GCU approach is more robust to the sensor-related measurement error. Simulation experiments verified the effectiveness of the proposed approaches.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"120 9 1","pages":"119130E - 119130E-11"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81304472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Single frame shadow segmentation based on image enhancement for video SAR","authors":"Yiming Xu, Dongsheng Li, Jushi Tang","doi":"10.1117/12.2604767","DOIUrl":"https://doi.org/10.1117/12.2604767","url":null,"abstract":"As Video synthetic aperture radar (SAR) technology has been developing rapidly in recent years, moving target detection and tracking has gradually become a research hotspot in the field of SAR. Since moving targets in Video SAR produce relatively clear shadows at their real locations, the shadow-based approach provides a new method for ground moving target detection. In this paper, a new approach based on image fusion enhancement is proposed to improve the extraction effect of target shadow in single frame Video SAR image, and the process of shadow segmentation is studied accordingly. First, we use Median Filter to denoise the image, and then use a variety of image enhancement methods to improve the contrast between shadows and background, including piecewise linear stretching, histogram specification, and S-curve enhancement, then use adaptive threshold segmentation algorithm to realize the separation of background and target shadow, finally use morphological processing method to further highlight the target shadow. The effectiveness of the proposed approach is verified on the Video SAR dataset published by Sandia Lab.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"38 1","pages":"1191306 - 1191306-8"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81229078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Research on infrared and visible image registration of substation equipment based on multi-scale Retinex and ASIFT features","authors":"Ning Yang, Yang Yang, Peng Li, Fei Gao","doi":"10.1117/12.2605000","DOIUrl":"https://doi.org/10.1117/12.2605000","url":null,"abstract":"Infrared and visible image registration of substation equipment is of great significance for power equipment detection and fault diagnosis. The scene of substation is complex, and the background of equipment image is usually messy, and the feature points of visible image are easy to fall on the background. The metal has good thermal conductivity, and its temperature is close to the ambient temperature. The metal part in the infrared image with metal tower as the background can not be clearly displayed, which is easy to cause the image mismatch or even unable to match. The existing registration methods such as SIFT, SURF and ASIFT are difficult to effectively solve this kind of image registration problem of substation equipment with complex background. To solve this problem, this paper proposes an infrared and visible image registration algorithm based on Multi-scale Retinex and ASIFT features. Firstly, the Multi-scale Retinex algorithm is used to separate the components representing the properties of the object in the visible image, so as to weaken the influence of the clutter background. Then, the ASIFT algorithm is used to do affine transformation to simulate the affine deformation under all parallax, and the feature points are roughly matched Finally, the random sampling consistent algorithm is added to eliminate the mismatching points. Experimental results show that the algorithm can increase the number of matching points by at least 4 times, the average matching accuracy is improved by 13%, and the average matching time is shortened by 183ms.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"1 1","pages":"1191303 - 1191303-8"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87190004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Nonnegative matrix factorization of DCE-MRI for prostate cancer classification","authors":"Aijie Hou, Yahui Peng, Xinchun Li","doi":"10.1117/12.2604770","DOIUrl":"https://doi.org/10.1117/12.2604770","url":null,"abstract":"The purpose of the study is to analyze whether certain components can be extracted in dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) for the classification of prostate cancer (PCa). Nonnegative matrix factorization (NMF) was used to extract the characteristic curve from DCE-MRI. The peak sharpness of the characteristic curve was evaluated to classify prostates with and without PCa. Results showed that the peak sharpness of the characteristic curve was significantly different in prostates with and without PCa (p = 0.008) and the area under the receiver operating characteristic curve was 0.86 ± 0.08. We conclude that the NMF can decompose DCE-MRI into components and the peak sharpness of the characteristic curve has the promise to classify prostates with and without PCa accurately.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"1 1","pages":"1191305 - 1191305-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79973871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Quality perception and discrimination thresholds in quantised triangle meshes","authors":"Aeshah Almutairi, T. Saarela, I. Ivrissimtzis","doi":"10.1117/12.2604720","DOIUrl":"https://doi.org/10.1117/12.2604720","url":null,"abstract":"At certain stages of the graphics pipeline, and most notably during compression for transmission and storage, triangle meshes may undergo a fixed-point arithmetic quantisation of their vertex coordinates. This paper presents the results of a psychophysical experiment, where discrimination thresholds between the original unquantised triangle meshes, and the quantised at various levels of quantisation versions of them, were estimated. The experiment had a two-alternative forced choice design. Our results show that the amount of geometric information of a mesh, as measured by its filesize after compression, correlates with the discrimination threshold. On the other hand, we did not find any correlation between the discrimination thresholds and the quality of the underlying meshing, as measured by the mean aspect ration of the mesh triangles.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"66 1","pages":"1191308 - 1191308-11"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91235524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xiaohong Cai, Ming Li, H. Cao, Jin-gang Ma, Xiaoyan Wang, Xuqiang Zhuang
{"title":"Image classification based on self-attention convolutional neural network","authors":"Xiaohong Cai, Ming Li, H. Cao, Jin-gang Ma, Xiaoyan Wang, Xuqiang Zhuang","doi":"10.1117/12.2604788","DOIUrl":"https://doi.org/10.1117/12.2604788","url":null,"abstract":"Image classification technology is the most basic and important technical branch of computer vision. How to effectively extract effective information from images has become more and more urgent. First, we use the self-attention module to use the correlation between the features to weight and sum the features to get the image category. The self-attention mechanism is simpler to calculate, which greatly reduces the complexity of the model. Secondly, we have also made an optimization strategy for the complex CNN (Convolutional Neural Network) model. This article uses the global average pooling method to replace the fully connected method, which reduces the complexity of the model and generates fewer features. Finally, we verified the feasibility and effectiveness of our model on two data sets.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"57 1","pages":"1191307 - 1191307-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80531116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jessie R. Balbin, Marianne M. Sejera, Ziad N. Al-Sagheer, Jann Amiel Nidehn B. Castañeda, Von Andrine V. Francisco
{"title":"Mobile geo-tagging and cloud-based underwater garbage identification using convolutional neural network","authors":"Jessie R. Balbin, Marianne M. Sejera, Ziad N. Al-Sagheer, Jann Amiel Nidehn B. Castañeda, Von Andrine V. Francisco","doi":"10.1117/12.2605058","DOIUrl":"https://doi.org/10.1117/12.2605058","url":null,"abstract":"Water is the essence of life, and water pollution is a major threat to all living things on this planet. To provide solutions to help combat water pollution, we have created a device that would help locate and identify the different garbage types underwater. This paper focused on the detection and identification of cans, plastics, polystyrenes, and glass underwater using object detection and object identification by Convolutional Neural Network and Geotagging. The system set-up comprises the following: a webcam, power bank, Raspberry Pi, GPS module, and an improvise floater. The GUI will display the camera's captured video, the number of garbage identified, and its location in coordinates. The testing was done in two ways: different water visibility and different water levels. The identification accuracy of our program is 94.33% for plastics, 97.34% for glass, 96.89% for polystyrenes, 98.22% for cans, and 96.88% for random garbage, reliability for identification is 100% for plastics, 91.67% for glass, 91.67% for polystyrenes, 95.83% for cans, and 91.67% for random garbage. The mean, median, and mode for the visibility levels are 96.375, 98, and 99, and the depth level is 96.385, 98, and 99.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"6 1","pages":"119130N - 119130N-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89936207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Learning human-object interactions by attention aggregation","authors":"Dongzhou Gu, Shuang Cai, Shiwei Ma","doi":"10.1117/12.2604708","DOIUrl":"https://doi.org/10.1117/12.2604708","url":null,"abstract":"Recent years, deep neural networks have achieved impressive progress in object detection. However, detecting the interactions between objects is still challenging. Many researchers pay attention to human-object interaction (HOI) detection as a basic task in detailed scene understanding. Most conventional HOI detectors are in a two-stage manner and usually slow in inference. One-stage methods for direct parallel detection of HOI triples breaks through the limitation of object detection, but the extracted features are still insufficient. To overcome these drawbacks above, we propose an improved one-stage HOI detection approach, in which attention aggregation module and dynamic point matching strategy play key roles. The attention aggregation enhances the semantic expression ability of interaction points explicitly by aggregating contextually important information, while the matching strategy can filter the negative HOI pairs effectively in the inference stage. Extensive experiments on two challenging HOI detection benchmarks: VCOCO and HICO-DET show that our method achieves considerable performance compared to state-of-the-art performance without any additional human pose and language features.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"68 1","pages":"119130H - 119130H-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74934336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}