Eliyah Immanuel Thavaraj A, S. Juliet, Anila Sharon J
{"title":"图像自动标注优化器的比较研究","authors":"Eliyah Immanuel Thavaraj A, S. Juliet, Anila Sharon J","doi":"10.1109/ICECA55336.2022.10009435","DOIUrl":null,"url":null,"abstract":"In the field of artificial intelligence, computer vision and natural language processing are used to automatically generate an image's contents. The regenerative neuronal model is developed and is dependent on machine translation and computer vision. Using this technique, natural phrases are produced that finally explain the image. This architecture also includes recurrent neural networks (RNN) and convolutional neural networks (CNN). The RNN is used to create phrases, whereas the CNN is used to extract characteristics from images. The model has been taught to produce captions that, when given an input image, almost exactly describe the image. The outcome of these algorithms is determined by several factors, including feature extraction, caption generation, and optimizer selection. Our goal is to conduct a comparative analysis of several optimizers to determine the optimizer that achieves highest accuracy for a deep learning model. The deep learning model is subsequently trained with various optimizers on the Flicker dataset. The accuracy of the results of the model using optimizers are achieved as follows: RMSprop optimizer has a 92% accuracy, SGD has a 12% accuracy, Adam optimizer has 53% accuracy, and Adadelta has a 12% per cent.","PeriodicalId":356949,"journal":{"name":"2022 6th International Conference on Electronics, Communication and Aerospace Technology","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"A Comparative Study on Optimizers for Automatic Image Captioning\",\"authors\":\"Eliyah Immanuel Thavaraj A, S. Juliet, Anila Sharon J\",\"doi\":\"10.1109/ICECA55336.2022.10009435\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the field of artificial intelligence, computer vision and natural language processing are used to automatically generate an image's contents. The regenerative neuronal model is developed and is dependent on machine translation and computer vision. Using this technique, natural phrases are produced that finally explain the image. This architecture also includes recurrent neural networks (RNN) and convolutional neural networks (CNN). The RNN is used to create phrases, whereas the CNN is used to extract characteristics from images. The model has been taught to produce captions that, when given an input image, almost exactly describe the image. The outcome of these algorithms is determined by several factors, including feature extraction, caption generation, and optimizer selection. Our goal is to conduct a comparative analysis of several optimizers to determine the optimizer that achieves highest accuracy for a deep learning model. The deep learning model is subsequently trained with various optimizers on the Flicker dataset. The accuracy of the results of the model using optimizers are achieved as follows: RMSprop optimizer has a 92% accuracy, SGD has a 12% accuracy, Adam optimizer has 53% accuracy, and Adadelta has a 12% per cent.\",\"PeriodicalId\":356949,\"journal\":{\"name\":\"2022 6th International Conference on Electronics, Communication and Aerospace Technology\",\"volume\":\"42 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 6th International Conference on Electronics, Communication and Aerospace Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICECA55336.2022.10009435\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 6th International Conference on Electronics, Communication and Aerospace Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICECA55336.2022.10009435","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Comparative Study on Optimizers for Automatic Image Captioning
In the field of artificial intelligence, computer vision and natural language processing are used to automatically generate an image's contents. The regenerative neuronal model is developed and is dependent on machine translation and computer vision. Using this technique, natural phrases are produced that finally explain the image. This architecture also includes recurrent neural networks (RNN) and convolutional neural networks (CNN). The RNN is used to create phrases, whereas the CNN is used to extract characteristics from images. The model has been taught to produce captions that, when given an input image, almost exactly describe the image. The outcome of these algorithms is determined by several factors, including feature extraction, caption generation, and optimizer selection. Our goal is to conduct a comparative analysis of several optimizers to determine the optimizer that achieves highest accuracy for a deep learning model. The deep learning model is subsequently trained with various optimizers on the Flicker dataset. The accuracy of the results of the model using optimizers are achieved as follows: RMSprop optimizer has a 92% accuracy, SGD has a 12% accuracy, Adam optimizer has 53% accuracy, and Adadelta has a 12% per cent.