{"title":"Improving the Quality of Image Captioning using CNN and LSTM Method","authors":"M. Pradeepan Lala, D. Kumar","doi":"10.1109/ICIIET55458.2022.9967570","DOIUrl":null,"url":null,"abstract":"In image captioning improving the content of the image by describing the meaning of the picture is a challenge as it should not only be understandable to the user but also described in a short and clear sentence. The proposed solution uses CNN and LSTM as captioning models in an Encoder and the Decoder methodology is used to translate an image into a sentence. The Xception architecture is modified by adding a depth wise convolution layer. A custom activation function is created based on swish and mish. CNN is used for feature extraction, RNN is used for sequence prediction, and LSTM for framing the words into a sentence. The proposed work is validated on two-dimensional image datasets such as dog category data extracted from the flicker8k dataset and real-time images captured through a webcam. The training/testing shows improved loss value, caption prediction time, and an increase in the quality of caption in terms of the BLEU@I parameter.","PeriodicalId":341904,"journal":{"name":"2022 International Conference on Intelligent Innovations in Engineering and Technology (ICIIET)","volume":"74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Intelligent Innovations in Engineering and Technology (ICIIET)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIIET55458.2022.9967570","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In image captioning improving the content of the image by describing the meaning of the picture is a challenge as it should not only be understandable to the user but also described in a short and clear sentence. The proposed solution uses CNN and LSTM as captioning models in an Encoder and the Decoder methodology is used to translate an image into a sentence. The Xception architecture is modified by adding a depth wise convolution layer. A custom activation function is created based on swish and mish. CNN is used for feature extraction, RNN is used for sequence prediction, and LSTM for framing the words into a sentence. The proposed work is validated on two-dimensional image datasets such as dog category data extracted from the flicker8k dataset and real-time images captured through a webcam. The training/testing shows improved loss value, caption prediction time, and an increase in the quality of caption in terms of the BLEU@I parameter.