{"title":"利用机器学习技术探索汉字雕塑的文化属性","authors":"Zhen Luo","doi":"10.32629/jai.v7i4.1471","DOIUrl":null,"url":null,"abstract":"The article employs machine learning, specifically the CLIP (Contrastive Language-Image Pretraining) model, to analyze Chinese character sculptures’ cultural attributes. It overcomes challenges in multi-dimensional data processing and high digitization costs. The process involves normalizing sculpture images, using FastText for vector representations of Chinese characters, and mapping text to the same embedding space as images for word embedding. The CLIP model, through unsupervised training, minimizes the negative logarithmic likelihood loss between image and text embeddings to establish cultural attribute representations. Key findings include the CLIP model’s improved performance over the M3 model, with a 5.4% higher average AUC. The model demonstrates high efficiency and accuracy, evident in its low RMSE (0.034) and MAE (0.025) and fast analysis time of 182 ms. This approach effectively and accurately analyzes the cultural attributes of Chinese character sculptures, addressing existing research gaps.","PeriodicalId":307060,"journal":{"name":"Journal of Autonomous Intelligence","volume":"77 3","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Exploration of the cultural attributes of Chinese character sculpture using machine learning technology\",\"authors\":\"Zhen Luo\",\"doi\":\"10.32629/jai.v7i4.1471\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The article employs machine learning, specifically the CLIP (Contrastive Language-Image Pretraining) model, to analyze Chinese character sculptures’ cultural attributes. It overcomes challenges in multi-dimensional data processing and high digitization costs. The process involves normalizing sculpture images, using FastText for vector representations of Chinese characters, and mapping text to the same embedding space as images for word embedding. The CLIP model, through unsupervised training, minimizes the negative logarithmic likelihood loss between image and text embeddings to establish cultural attribute representations. Key findings include the CLIP model’s improved performance over the M3 model, with a 5.4% higher average AUC. The model demonstrates high efficiency and accuracy, evident in its low RMSE (0.034) and MAE (0.025) and fast analysis time of 182 ms. This approach effectively and accurately analyzes the cultural attributes of Chinese character sculptures, addressing existing research gaps.\",\"PeriodicalId\":307060,\"journal\":{\"name\":\"Journal of Autonomous Intelligence\",\"volume\":\"77 3\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Autonomous Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.32629/jai.v7i4.1471\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Autonomous Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.32629/jai.v7i4.1471","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Exploration of the cultural attributes of Chinese character sculpture using machine learning technology
The article employs machine learning, specifically the CLIP (Contrastive Language-Image Pretraining) model, to analyze Chinese character sculptures’ cultural attributes. It overcomes challenges in multi-dimensional data processing and high digitization costs. The process involves normalizing sculpture images, using FastText for vector representations of Chinese characters, and mapping text to the same embedding space as images for word embedding. The CLIP model, through unsupervised training, minimizes the negative logarithmic likelihood loss between image and text embeddings to establish cultural attribute representations. Key findings include the CLIP model’s improved performance over the M3 model, with a 5.4% higher average AUC. The model demonstrates high efficiency and accuracy, evident in its low RMSE (0.034) and MAE (0.025) and fast analysis time of 182 ms. This approach effectively and accurately analyzes the cultural attributes of Chinese character sculptures, addressing existing research gaps.