Meme Expressive Classification in Multimodal State with Feature Extraction in Deep Learning

A. Barveen, S. Geetha, Mohamad Faizal
{"title":"Meme Expressive Classification in Multimodal State with Feature Extraction in Deep Learning","authors":"A. Barveen, S. Geetha, Mohamad Faizal","doi":"10.1109/ICEEICT56924.2023.10157066","DOIUrl":null,"url":null,"abstract":"Memes are a socially interactive way to communicate online. Memes are used by users to communicate with one another on social networking sites and other forums. Memes essentially focus on speech recognition and image macros. While a meme is being created, it focuses on the semiotic type of resources that the internet community interprets with other resources, which facilitates the interaction among the internet and meme creators. Memes recreate based on various approaches, which fall under various acts such as existing speech acts. Based on the expressive face with captioned short texts, even the short text is exaggerated. Every year, meme mimicking applications are created that allow users to use the imitated meme expressions. Memes represent the shared texts of the younger generations on various social platforms. The classifications of sentiment based on the various memetic expressions are the most efficient way to analyse those feelings and emotions. HOG feature extraction allows the images to be segmented into blocks of smaller size by using a single feature vector for dimension, which characterizes the local object appearances to characterize the meme classification. The existence of specific characteristics, including such edges, angles, or patterns, is then analyzed by combining HOG features using multi-feature analysis on patches. Based upon the classification methodology, it classifies the sentiments, which tend to improve the learning process in an efficient manner. By combining a deep learning approach with a recurrent neural network, the extended LSTM-RNN can identify subtle nuances in memes, allowing for more accurate and detailed meme classification. This proposed method effectively evaluates several classification techniques, including CNN and Extended LSTM-RNN for meme image characterization. Through training and validation, Extended LSTM-RNN achieved 0.98% accuracy with better performance than CNN.","PeriodicalId":345324,"journal":{"name":"2023 Second International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 Second International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEEICT56924.2023.10157066","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Memes are a socially interactive way to communicate online. Memes are used by users to communicate with one another on social networking sites and other forums. Memes essentially focus on speech recognition and image macros. While a meme is being created, it focuses on the semiotic type of resources that the internet community interprets with other resources, which facilitates the interaction among the internet and meme creators. Memes recreate based on various approaches, which fall under various acts such as existing speech acts. Based on the expressive face with captioned short texts, even the short text is exaggerated. Every year, meme mimicking applications are created that allow users to use the imitated meme expressions. Memes represent the shared texts of the younger generations on various social platforms. The classifications of sentiment based on the various memetic expressions are the most efficient way to analyse those feelings and emotions. HOG feature extraction allows the images to be segmented into blocks of smaller size by using a single feature vector for dimension, which characterizes the local object appearances to characterize the meme classification. The existence of specific characteristics, including such edges, angles, or patterns, is then analyzed by combining HOG features using multi-feature analysis on patches. Based upon the classification methodology, it classifies the sentiments, which tend to improve the learning process in an efficient manner. By combining a deep learning approach with a recurrent neural network, the extended LSTM-RNN can identify subtle nuances in memes, allowing for more accurate and detailed meme classification. This proposed method effectively evaluates several classification techniques, including CNN and Extended LSTM-RNN for meme image characterization. Through training and validation, Extended LSTM-RNN achieved 0.98% accuracy with better performance than CNN.
基于深度学习特征提取的多模态模因表达分类
模因是一种在线交流的社交互动方式。模因是用户在社交网站和其他论坛上相互交流的工具。模因主要关注语音识别和图像宏。在模因产生的过程中,它关注的是网络社区用其他资源解释的资源的符号学类型,这有利于互联网和模因创造者之间的互动。模因基于各种方法进行再现,这些方法属于各种行为,例如现有的语言行为。从这张带字幕的表情脸来看,就连短文都被夸大了。每年都会出现模因模仿应用程序,允许用户使用模仿的模因表情。表情包代表了年轻一代在各种社交平台上的共享文本。基于各种模因表达的情绪分类是分析这些感觉和情绪的最有效方法。HOG特征提取允许使用单个特征向量作为维度,将图像分割成较小尺寸的块,特征向量表征局部物体的外观,从而表征模因分类。然后通过对patch进行多特征分析,结合HOG特征来分析特定特征(包括边缘、角度或图案)的存在性。在分类方法的基础上,对情感进行分类,有利于有效地改进学习过程。通过将深度学习方法与递归神经网络相结合,扩展的LSTM-RNN可以识别模因中的细微差别,从而实现更准确和详细的模因分类。该方法有效地评估了几种分类技术,包括CNN和扩展LSTM-RNN用于模因图像表征。经过训练和验证,扩展LSTM-RNN准确率达到0.98%,优于CNN。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信