{"title":"Machine Learning for Multiscale Video Coding","authors":"M. V. Gashnikov","doi":"10.3103/S1060992X23030037","DOIUrl":"10.3103/S1060992X23030037","url":null,"abstract":"<p>The research concerns the use of machine learning algorithms for multiscale coding of digital video sequences. Based on machine learning, the digital image coder is generalized to the coding of video sequences. To this end, we offer an algorithm that allows for videoframes interdependency by using linear regression. The generalized image coder uses multiscale representation of videoframes, neural network three-dimensional interpolation of multiscale videoframe interpretation levels and generative-adversarial neural net replacement of homogeneous portions of a videoframe by synthetic video data. The method of coding the entire video and method of coding videoframes are exemplified by block diagrams. Formalized description of how videoframe correlation is taken into account is given. Real video sequences are used to carry out numerical experiments. The experimental data allow us to make a conclusion about the promise of using the algorithm in video coding and processing.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 3","pages":"189 - 196"},"PeriodicalIF":0.9,"publicationDate":"2023-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41079807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"English-Afaan Oromo Machine Translation Using Deep Attention Neural Network","authors":"Ebisa A. Gemechu, G. R. Kanagachidambaresan","doi":"10.3103/S1060992X23030049","DOIUrl":"10.3103/S1060992X23030049","url":null,"abstract":"<p>Attention-based neural machine translation (attentional NMT), which jointly aligns and translates, has got much popularity in recent years. Besides, a language model needs an accurate and larger bilingual dataset_ from the source to the target, to boost translation performance. There are many such datasets publicly available for well-developed languages for model training. However, currently, there is no such dataset available for the English-Afaan Oromo pair to build NMT language models. To alleviate this problem, we manually prepared a 25K English-Afaan Oromo new dataset for our model. Language experts evaluate the prepared corpus for translation accuracy. We also used the publicly available English-French, and English-German datasets to see the translation performances among the three pairs. Further, we propose a deep attentional NMT model to train our models. Experimental results over the three language pairs demonstrate that the proposed system and our new dataset yield a significant gain. The result from the English-Afaan Oromo model achieved 1.19 BLEU points over the previous English-Afaan Oromo Machine Translation (MT) models. The result also indicated that the model could perform as closely as the other developed language pairs if supplied with a larger dataset. Our 25K new dataset also set a baseline for future researchers who have curiosity about English-Afaan Oromo machine translation.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 3","pages":"159 - 168"},"PeriodicalIF":0.9,"publicationDate":"2023-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41079707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Skin Cancer Detection and Classification System by Applying Image Processing and Machine Learning Techniques","authors":"Dr. A. Rasmi, Dr. A. Jayanthiladevi","doi":"10.3103/S1060992X23030086","DOIUrl":"10.3103/S1060992X23030086","url":null,"abstract":"<p>In these modern days, cancers like Skin cancers is the general type of cancer that alters the life style of millions of citizens in each time. Around three million people are identified with it in each and every year in the US alone. Skin cancer related to the irregular enlargement of cells. On account of malignancy characteristic skin type cancer is termed as melanoma. Melanoma seems on skins because of the contact to ultraviolet emission and hereditary reasons. Thus melanoma seems like brown and black colour, but also occurs anyplace of the patient. Mostly the skin type cancers could be treatable at the earliest phases of beginning. So a fast recognition of skin cancer could rescue the life of patient. However, identifying skin cancer in its starting phases may be difficult and moreover it is expensive. Thus in the paper, they try to cope with such types of problems by making a wise decision scheme for skin lesion identification like the starting phase, that should be set into a smart robot for physical condition monitoring in our present surroundings to support early on detection. The scheme is enhanced to classify benign and malignant skin lesions with different procedures, comprising of pre-processing for instance noise elimination, segmentation, and feature extraction from lesion sections, feature collection and labelling. Following the separation of lots of raw images, colour and texture characteristics from the lesion regions, is employed to categorize the largely prejudiced noteworthy subsets for fit and cancerous circumstances. In it SVM has been applied to carry out benign and malignant lesion detection.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 3","pages":"197 - 203"},"PeriodicalIF":0.9,"publicationDate":"2023-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41079703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Comparison of the 2011 and 2020 Stratospheric Ozone Events at Arctic and Northern Eurasian Latitudes Using TEMIS and Aura MLS Data","authors":"O. E. Bazhenov","doi":"10.3103/S1060992X23030025","DOIUrl":"10.3103/S1060992X23030025","url":null,"abstract":"<p>Winters-springs 2019–2020 and 2010–2011 became the periods of the severest ozone events in the Arctic throughout the satellite era. They stemmed from extremely cold and persistent polar stratospheric cloud (PSC) seasons, conducive to record strong chemical ozone destruction. TEMIS observations indicate that the total ozone (TO) column diverged from long-term norm by 45 to 55% in 2020 and by 37 to 44% in 2011 at Arctic sites; and by 27 to 32% in 2020 and by 27 to 36% in 2011 at midlatitudes. Aura MLS profiles showed that the minimum temperature was 8–13% lower than norm over the Arctic in 2020 and 8–12% lower in 2011. The ozone mixing ratios were 4% of the long-term mean at height of 20 km on March 27, 2020 and 25% at height of 21 km on March 20, 2011 for Eureka; and 7% at 19 km on April 19, 2020 and 24% at 20 km on March 20, 2011 for Ny-Ålesund. The divergences in water vapor and ozone mixing ratios, water vapor mixing ratio and temperature, and ozone mixing ratio and temperature show stronger correlations in 2020 than 2011. The correlations weaken equatorward, until becoming almost insignificant at extra-vortex latitudes.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 3","pages":"182 - 188"},"PeriodicalIF":0.9,"publicationDate":"2023-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41079700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. A. Nevzorov, A. V. Nevzorov, A. I. Nadeev, N. G. Zaitsev, Ya. O. Romanovskii
{"title":"Algorithm for Data Processing from Ozone Lidar Sensing in the Atmosphere","authors":"A. A. Nevzorov, A. V. Nevzorov, A. I. Nadeev, N. G. Zaitsev, Ya. O. Romanovskii","doi":"10.3103/S1060992X23030050","DOIUrl":"10.3103/S1060992X23030050","url":null,"abstract":"<p>We developed an algorithm of software product for processing the data from lidar sensing at the wavelengths of 299/341 nm for a vertical path of atmospheric sensing with the spatial resolution from 1.5 to 150 m. The main options of the software include: recording the atmospheric lidar sensing data, conversion of DAT to TXT file format, and retrieval of ozone concentration profiles. The software complex, developed on the basis of our algorithm to process the lidar sensing data, makes it possible to obtain the ozone concentration profiles from 4 to 20 km. The blocks of recording the data from atmospheric lidar sensing and retrieving the ozone concentration profiles allow for a visual control of the recorded lidar returns and retrieved ozone concentration profiles. We present an example of retrieving the ozone concentration profile from lidar data, which was obtained in 2022.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 3","pages":"169 - 181"},"PeriodicalIF":0.9,"publicationDate":"2023-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41079702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Application of Deep Neural Network Structures in Semantic Segmentation for Road Scene Understanding","authors":"Qusay Sellat, Kanagachidambaresan Ramasubramanian","doi":"10.3103/S1060992X23020108","DOIUrl":"10.3103/S1060992X23020108","url":null,"abstract":"<p>Semantic segmentation is crucial for autonomous driving as the pixel-wise classification of the surrounding scene images is the main input in the scene understanding stage. With the development of deep learning technology and the impressive hardware capabilities, semantic segmentation has seen an important improvement towards higher segmentation accuracy. However, an efficient sematic segmentation model is needed for real-time applications such as autonomous driving. In this paper, we discover the potential of employing the design principles of two deep learning models, namely PSPNet and EfficientNet to produce a high accurate and efficient convolutional autoencoder model for semantic segmentation. Also, we benefit from data augmentation for better model training. Our experiment on CamVid dataset produces optimistic results and the comparison with other mainstream semantic segmentation models justifies the used approach.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 2","pages":"137 - 146"},"PeriodicalIF":0.9,"publicationDate":"2023-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"4898548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detection and Prediction of Breast Cancer Using Improved Faster Regional Convolutional Neural Network Based on Multilayer Perceptron’s Network","authors":"Poonam Rana, Pradeep Kumar Gupta, Vineet Sharma","doi":"10.3103/S1060992X23020054","DOIUrl":"10.3103/S1060992X23020054","url":null,"abstract":"<p>One of the most frequent causes of death for women worldwide is breast cancer. In most cases, breast cancer can be quickly identified if certain symptoms emerge. But many women with breast cancer don’t show any symptoms. So, it is very critical to detect this disease in early stage also numerous radiologists are needed to diagnose this disease which is quite expensive for the majority of cancer hospitals. To address these concerns, the proposed methodology creates a Faster-Regional Convolutional Neural Network (Faster-RCNN) for recognizing breast cancer. Ultrasound images are collected and pre-processed utilizing resizing, adaptive median filter, histogram global contrast enhancement and high boost filtering. Image resizing is utilized to change the image size without cutting anything out. Adaptive median filter is utilized to remove unwanted noise present in the resized image. Histogram global contrast enhancement is used to enhancing the contrast level of the image. High boost filtering is utilized to sharpening the edges present in the image. After that, pre-processed images are fetched as an input to Faster R-CNN, which extract the features and segment the accurate region of the tumour. These segmented regions are classified using Multilayer Perceptron’s for detecting whether the patients are affected by breast cancer or not. According to the experimental study, the proposed approach achieves 97.1% correctness, 0.03% error, 91% precision and 93% specificity. Therefore, the developed approach attains better performance compared to other existing approaches. This prediction model helps to detect breast cancer at early stage and improve patient’s living standard.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 2","pages":"86 - 100"},"PeriodicalIF":0.9,"publicationDate":"2023-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"4895873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Improving the Performance of Human Part Segmentation Based on Swin Transformer","authors":"Juan Du, Tao Yang","doi":"10.3103/S1060992X23020030","DOIUrl":"10.3103/S1060992X23020030","url":null,"abstract":"<p>One of the current challenges in deep learning is semantic segmentation. Moreover, human part segmentation is a sub-task in image segmentation, which differs from traditional segmentation to understand the human body’s intrinsic connections. Convolutional Neural Network (CNN) has always been a standard feature extraction network in human part segmentation. Recently, the proposed Swin Transformer surpasses CNN for many image applications. However, few articles have explored the performance of Swin Transformer in human part segmentation compared to CNN. In this paper, we make a comparison experiment on this issue, and the experimental results prove that even in the area of human part segmentation and without any additional trick, the Swin Transformer has good results compared with CNN. At the same time, this paper also combines the Edge Perceiving Module (EPM) currently commonly used in CNN with Swin Transformer to prove that Swin Transformer can see the intrinsic connection of segmented parts. This research demonstrates the feasibility of applying Swin Transformer to the part segmentation of images, which is conducive to advancing image segmentation technology in the future.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 2","pages":"101 - 107"},"PeriodicalIF":0.9,"publicationDate":"2023-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"4898556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Ben-Ali, Y. Errouas, I. El Kadmiri, Z. Rahou, A. Hallaoui, D. Bria
{"title":"Filters Based on Defect Modes by 1D Star Waveguides Defective System","authors":"Y. Ben-Ali, Y. Errouas, I. El Kadmiri, Z. Rahou, A. Hallaoui, D. Bria","doi":"10.3103/S1060992X23020029","DOIUrl":"10.3103/S1060992X23020029","url":null,"abstract":"<p>In this paper, we present a theoretical study of the properties of defect modes in one-dimensional defective photonic star waveguides (SWGs). Both symmetrically and asymmetrically stacked defective SWGs are considered in the analysis. The properties of the defect modes are studied through the calculation of the frequency-dependent transmittance. We have also explored the effects of defect length and permittivity, as well as the position of the defect and the number of resonators grafted onto the SWGs on the number of defect modes, their transmission, and their quality factor. The findings demonstrate that geometric defects result in two modes of maximum transmission with a high-quality factor (<span>({{Q}_{1}} = 358)</span> and <span>({{Q}_{2}} = 1550)</span>). However, when there are both geometric and material defects, the quality factor of the second filter <span>({{Q}_{2}})</span> improved from 1550 to 2203, while the quality factor of the first filter <span>({{Q}_{1}})</span>remains almost unchanged. Moreover, when only material defects are present, two electromagnetic filters with maximum transmission and high-quality factor (<span>({{Q}_{1}} = 1708)</span> and <span>({{Q}_{2}} = 9085)</span>) can be obtained.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 2","pages":"108 - 125"},"PeriodicalIF":0.9,"publicationDate":"2023-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"4895869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Text-Text Neural Machine Translation: A Survey","authors":"Ebisa Gemechu, G. R. Kanagachidambaresan","doi":"10.3103/S1060992X23020042","DOIUrl":"10.3103/S1060992X23020042","url":null,"abstract":"<p>We present a review of Neural Machine Translation (NMT), which has got much popularity in recent decades. Machine translation eased the way we do massive language translation in the new digital era. Otherwise, language translation would have been manually done by human experts. However, manual translation is very costly, time-consuming, and prominently inefficient. So far, three main Machine Translation (MT) techniques have been developed over the past few decades. Viz rule-based, statistical, and neural machine translations. We have presented the merits and demerits of each of these methods and discussed a more detailed review of articles under each category. In the present survey, we conducted an in-depth review of existing approaches, basic architecture, and models for MT systems. Our effort is to shed light on the existing MT systems and assist potential researchers, in revealing related works in the literature. In the process, critical research gaps have been identified. This review intrinsically helps researchers who are interested in the study of MT.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"32 2","pages":"59 - 72"},"PeriodicalIF":0.9,"publicationDate":"2023-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"4898744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}