{"title":"EEG-based affect states classification using Deep Belief Networks","authors":"Haiyan Xu, K. Plataniotis","doi":"10.1109/DMIAF.2016.7574921","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574921","url":null,"abstract":"Affective states classification has become an important part of the Brain-Computer Interface (HCI) study. In recent years, affective computing systems using physiological signals, such as ECG, GSR and EEG has shown very promising results. However, like many other machine learning studies involving physiological signals, the bottle neck is always around the database acquisition and the annotation process. To investigate potential ways to address this small sample problem, this paper introduces a Deep Belief Networks (DBN) based learning system for the EEG-based affective processing system. Through the greedy-layer pretraining using unlabeled data as well as a supervised fine-tuning process, the DBN-based approaches significantly reduced the number of labeled samples required. The DBN methods also acted as an application specific feature selector, by examining the weight vector between the input feature vector and the first invisible layer, we can gain much needed insights on the spatial or spectral locations of the most discriminating features. In this study, DBNs are trained on the narrow-band spectral features extracted from multichannel EEG recordings. To evaluate the efficacy of the proposed DBN-based learning system, we carried out an subject-independent affective states classification experiments on the DEAP database to classify 2-dimensional affect states. As a baseline to the proposed DBN approach, the same classification problem was also carried out using support vector machines (SVMs) and one-way ANOVA based feature selection process. The classification results shown that the proposed framework using Deep Belief Networks not only provided better classification performance, but also significantly lower the number of labeled data required to train such machine learning systems.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128439921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Vision-based engagement detection in Virtual Reality","authors":"Ghassem Tofighi, Haisong Gu, K. Raahemifar","doi":"10.1109/DMIAF.2016.7574933","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574933","url":null,"abstract":"User engagement modeling for manipulating actions in vision-based interfaces is one of the most important case studies of user mental state detection. In a Virtual Reality environment that employs camera sensors to recognize human activities, we have to know were user intend to perform an action and when he/she is disengaged. Without a proper algorithm for recognizing engagement status, any kind of activities could be interpreted as manipulating actions, called “Midas Touch” problem. Baseline approach for solving this problem is activating gesture recognition system using some focus gestures such as waiving or raising hand. However, a desirable natural user interface should be able to understand user's mental status automatically. In this paper, a novel multi-modal model for engagement detection, DAIA 1, is presented. using DAIA, the spectrum of mental status for performing an action is quantized in a finite number of engagement states. For this purpose, a Finite State Transducer (FST) is designed. This engagement framework shows how to integrate multi-modal information from user biometric data streams such as 2D and 3D imaging. FST is employed to make the state transition smoothly using combination of several boolean expressions. Our FST true detection rate is 92.3% in total for four different states. Results also show FST can segment user hand gestures more robustly.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130047595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assessing unreliability in OTT video QoE subjective evaluations using clustering with idealized data","authors":"Jie Jiang, P. Spachos, M. Chignell, L. Zucherman","doi":"10.1109/DMIAF.2016.7574940","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574940","url":null,"abstract":"In this paper, we describe an Over-The-Top (OTT) video Quality of Experience (QoE) subjective evaluation experiment that was carried out to examine variations in the way subjects assess viewing experiences. The experiment focuses on different level of impairment and failure types, using 5-point measurement scales. Clustering is used to differentiate between unreliable and reliable participants, where reliability is defined in terms of criteria such as consistency of rating and ability to distinguish between qualitative differences in level of impairments. The results show that clustering a data set that is augmented with unreliable pseudo-participants can provide a new and improved perspective on individual differences in video QoE assessment.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134270013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Enabling enterprise-scale systems using cloud-based personal media","authors":"S. Fels, J. C. A. Silva","doi":"10.1109/DMIAF.2016.7574938","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574938","url":null,"abstract":"Personal cloud services are emerging as a disruptive technology for tools and services that use digital media assets. As the shift to a bring-your-own-device and app workplace becomes commonplace, the opportunity exists to leverage these personal services to create enterprise scale approaches. However, a key limiting feature is the ability to provide private and secure assets, whether they are photos, videos, audio and text in a unified way layered on top of personal cloud services. We discuss an approach to support this shift and illustrate its feasibility using a prototype secure email service that is layered on top of a popular unsecured cloud service. One of the primary benefits of our approach is that companies can create secure enterprise services on top of a third party cloud used by individuals alleviating the need for a complex IT infrastructure for digital media assets. These solutions are particularly relevant for emerging markets.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115352458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"HDR Video Coding based on a temporally constrained Tone Mapping Operator","authors":"C. Ozcinar, Paul Lauga, G. Valenzise, F. Dufaux","doi":"10.1109/DMIAF.2016.7574900","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574900","url":null,"abstract":"Given its potential for more realistic rendering and enhanced user experience, High Dynamic Range (HDR) imaging is raising a lot of interest both in industry and academia. In this context, efficient representation and coding techniques are needed, as HDR video entails significantly higher raw data rate. In this paper, we present a temporally constrained content-adaptive Tone Mapping Operator (TMO) in order to convert the input HDR video into a reduced bit depth video sequence which is then encoded using High Efficiency Video Coding (HEVC). As the proposed TMO simultaneously takes into account the statistical characteristics of the input frame while better preserving temporal coherence of the tone mapped video sequence, it leads to improved coding efficiency. Experimental results show that the proposed technique compares favorably with existing methods in terms of rate-distortion when using the HDR-VDP-2.2.1 quality metric.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133953369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"High Dynamic Range versus Standard Dynamic Range compression efficiency","authors":"Ronan Boitard, M. Pourazad, P. Nasiopoulos","doi":"10.1109/DMIAF.2016.7574890","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574890","url":null,"abstract":"High Dynamic Range (HDR) image and video technology aims at conveying the full range of perceptible shadow and highlight details with sufficient precision. HDR is regarded by many experts as the next evolution in digital media. However, industrial broadcasters have concerns regarding the bandwidth overhead that this new technology entails. While many consider that broadcasting HDR content would increase bandwidth requirements by around 20%, this number is based on studies where, in addition to the SDR main stream, HDR-related side information is conveyed. A recent subjective evaluation reported that encoding HDR video content in a single layer might require less bandwidth than its associated SDR version. Similar results were discussed in the MPEG ad-hoc group on High Dynamic Range and Wide Color Gamut. In this article, we explain how having more information can result in lower bandwidth requirements. To this end, we describe several limitations of the human vision system that, when exploited, optimize the HDR distribution pipeline for a human observer. Our theoretical assumption about the higher efficiency of HDR is backed up by a statistical analysis of pixel distribution in real images. The Spatial Index objective metric also reconfirms our assumption.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133431676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"New just noticeable coding distortion model for perceptual coding","authors":"Shengyang Xu, Mei Yu, G. Jiang, Shuqing Fang","doi":"10.1109/DMIAF.2016.7574928","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574928","url":null,"abstract":"With the aim of improving the efficiency and perceptual quality in video coding, this paper proposes a novel just-noticeable coding distortion (JNCD) model that considers human visual perception redundancy and unreasonable factors of existing just-noticeable distortion (JND) models in the coding process. First, we design a psycho-physical experiment to analyze the just-noticeable gradient difference (JNGD) and build a JNGD model to filter the gradient components that are imperceptible to human eyes. We use total variation (TV) to decompose an image into a structural image and a textural image, and calculate their gradients. Then, we use JNGD to filter out imperceptible gradient components in each gradient image. Second, human visual sensitivity to different gradient magnitudes is analyzed to model the relationship between the human visual perceptible gradient magnitude and JNCD. Finally, considering the perceived difference of human eye perception in edge, flat, and textural regions of an image, we adjust the JNCD value in each region and establish a JNCD model of the whole image. To verify the efficiency of the proposed JNCD model, we compare it with the classic JND model and test it on the high-efficiency video coding (HEVC) platform. The proposed model has advantages in subjective visual effects, meaning that it is helpful in analysis of human visual perception redundancy and the relevant perceptual video coding.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124138185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"VisQuery: Visual querying of streaming data via pattern matching","authors":"Chenhui Li, G. Baciu, Yunzhe Wang","doi":"10.1109/DMIAF.2016.7574924","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574924","url":null,"abstract":"Querying streaming data is becoming a dominant problem in big data analytics. A practical approach to querying streaming data is through traditional databases that have been modified to support streams, such as MySQL. However, conditional selection for querying data streams is currently an open challenge. We present a new visual framework that provides a more intuitive querying interaction for streaming data by combining visual selections on patterns with image processing techniques in order to better identify regions of interest. The main contribution of this paper is a novel method for matching patterns among normalized frames via feature vector clustering.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129369733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Stelios E. Ploumis, Ronan Boitard, M. Pourazad, P. Nasiopoulos
{"title":"Perception-based Histogram Equalization for tone mapping applications","authors":"Stelios E. Ploumis, Ronan Boitard, M. Pourazad, P. Nasiopoulos","doi":"10.1109/DMIAF.2016.7574892","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574892","url":null,"abstract":"Due to the ever increasing commercial availability of High Dynamic Range (HDR) content and displays, backward compatibility of HDR content with Standard Dynamic Range displays is currently a topic of high importance. Over the years, a significant amount of Tone Mapping Operators (TMOs) have been proposed to adapt HDR content to the restricted capabilities of SDR displays. Among them, the Histogram Equalization (HE) is considered to provide good results for a wide set of images. However, the naïve application of HE results either in banding artifacts or noise amplification when the HDR image has large unified areas (i.e. sky). In order to differentiate relevant information from noise in a uniform background, or in dark areas, the authors proposed a ceiling function. Their method results in noise-free but dim images. In this paper we propose a novel ceiling function which is based on the Perceptual Quantizer (PQ) function. Our method uses as threshold the number of code-words that PQ assigns on a luminance range in the original HDR image and the corresponding number of code-words in the resulting SDR image. We limit the number of code-words on SDR to be equal or less than the HDR. The saved code-words during the ceiling operation are redistributed to increase the contrast as well as the brightness of the final image. Results shows that provided SDR images are noise-free and brighter than the one obtained with prior HE operators. Finally since the proposed method is a Global TMO, it is thereby of low complexity and suitable for real time applications.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125615833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Joint antenna allocation and rate adaption for video transmission in massive MIMO systems","authors":"Bowen Liu, Heli Zhang, Hong Ji, Xi Li, Ke Wang","doi":"10.1109/DMIAF.2016.7574906","DOIUrl":"https://doi.org/10.1109/DMIAF.2016.7574906","url":null,"abstract":"Massive multi-input-multi-output (MIMO) networks could achieve higher data transmission rate benefited from the advantages of space diversity and multiplexing. In recent years, large amounts of research about different service adopted in massive MIMO network have been proposed. In this paper, we investigate instant video communication services requested by users in massive MIMO networks. After defining a detailed system model for video streaming in massive MIMO networks, we jointly consider the problem of antenna allocation and time-average video streaming scheduling. Since the problem is NP-hard, we reformulate it by decomposing the problem into two sub-problems that are antennas allocation and video packets queuing so that some fast common algorithms can be employed. To solve the two sub-problems, Enhanced Hungarian algorithm (EHA) and Enhanced Kuhn-Munkras algorithm (EKM) are designed for antenna allocation, and High Quality Fair Queuing (HQFQ) algorithm is proposed for video streaming scheduling. Consequently, numerical solution can be calculated in the time scale of real-life video streaming sessions. Various results demonstrate that our approach performs well in balance of quality of service and fairness to video streaming users.","PeriodicalId":404025,"journal":{"name":"2016 Digital Media Industry & Academic Forum (DMIAF)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128848604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}