{"title":"混合现实中的多媒体检索:利用实时查询实现沉浸式体验","authors":"Rahel Arnold, H. Schuldt","doi":"10.1109/AIxVR59861.2024.00048","DOIUrl":null,"url":null,"abstract":"Recent advancements in Mixed Reality (MR) technology and the exponential growth of multimedia data production have led to the emergence of innovative approaches for efficient content retrieval. This paper introduces Mixed Reality Multimedia Retrieval ((MR)2), a groundbreaking concept at the convergence of MR and multimedia retrieval. At its core, (MR)2 leverages MR’s transformative capabilities with an innovative live query option, allowing users to initiate queries intuitively through real-world object interactions. By autonomously generating queries based on object recognition in the user’s field of view, (MR)2 facilitates the retrieval of similar multimedia content from a connected database. The technical backbone of the (MR)2 framework includes object detection (YOLOv8), semantic similarity search (CLIP), and data management (Cottontail DB). Our research redefines user interactions with multimedia databases, seamlessly bridging the physical and digital domains. A successful iOS prototype application demonstrates promising results, paving the way for immersive and context-aware multimedia retrieval in the MR era.","PeriodicalId":518749,"journal":{"name":"2024 IEEE International Conference on Artificial Intelligence and eXtended and Virtual Reality (AIxVR)","volume":"199 2","pages":"289-293"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multimedia Retrieval in Mixed Reality: Leveraging Live Queries for Immersive Experiences\",\"authors\":\"Rahel Arnold, H. Schuldt\",\"doi\":\"10.1109/AIxVR59861.2024.00048\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent advancements in Mixed Reality (MR) technology and the exponential growth of multimedia data production have led to the emergence of innovative approaches for efficient content retrieval. This paper introduces Mixed Reality Multimedia Retrieval ((MR)2), a groundbreaking concept at the convergence of MR and multimedia retrieval. At its core, (MR)2 leverages MR’s transformative capabilities with an innovative live query option, allowing users to initiate queries intuitively through real-world object interactions. By autonomously generating queries based on object recognition in the user’s field of view, (MR)2 facilitates the retrieval of similar multimedia content from a connected database. The technical backbone of the (MR)2 framework includes object detection (YOLOv8), semantic similarity search (CLIP), and data management (Cottontail DB). Our research redefines user interactions with multimedia databases, seamlessly bridging the physical and digital domains. A successful iOS prototype application demonstrates promising results, paving the way for immersive and context-aware multimedia retrieval in the MR era.\",\"PeriodicalId\":518749,\"journal\":{\"name\":\"2024 IEEE International Conference on Artificial Intelligence and eXtended and Virtual Reality (AIxVR)\",\"volume\":\"199 2\",\"pages\":\"289-293\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2024 IEEE International Conference on Artificial Intelligence and eXtended and Virtual Reality (AIxVR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AIxVR59861.2024.00048\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 IEEE International Conference on Artificial Intelligence and eXtended and Virtual Reality (AIxVR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AIxVR59861.2024.00048","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multimedia Retrieval in Mixed Reality: Leveraging Live Queries for Immersive Experiences
Recent advancements in Mixed Reality (MR) technology and the exponential growth of multimedia data production have led to the emergence of innovative approaches for efficient content retrieval. This paper introduces Mixed Reality Multimedia Retrieval ((MR)2), a groundbreaking concept at the convergence of MR and multimedia retrieval. At its core, (MR)2 leverages MR’s transformative capabilities with an innovative live query option, allowing users to initiate queries intuitively through real-world object interactions. By autonomously generating queries based on object recognition in the user’s field of view, (MR)2 facilitates the retrieval of similar multimedia content from a connected database. The technical backbone of the (MR)2 framework includes object detection (YOLOv8), semantic similarity search (CLIP), and data management (Cottontail DB). Our research redefines user interactions with multimedia databases, seamlessly bridging the physical and digital domains. A successful iOS prototype application demonstrates promising results, paving the way for immersive and context-aware multimedia retrieval in the MR era.