{"title":"HappyFeet: Challenges in Building an Automated Dance Recognition and Assessment Tool","authors":"A. Faridee, S. R. Ramamurthy, Nirmalya Roy","doi":"10.1145/3308755.3308759","DOIUrl":"https://doi.org/10.1145/3308755.3308759","url":null,"abstract":"In this paper, we discuss our experience in building an automated dance assessment tool with IMU and IoT devices and highlight the major challenges of such an endeavor. In a typical dance classroom scenario, where the students frequently outnumber their instructors, such a system can add an immense value to both parties by providing systematic breakdown of the dance moves, comparing the dance moves between the students and the instructors, and pinpointing the places for improvement in an autonomous way. Along that direction, our prototypical work, HappyFeet [1], showcases our initial attempts of developing such an intelligent Dance Activity Recognition (DAR) system. Our CNN based Body Sensor Network proves more effective (by ≈7% margin at 94.20%) at accurately recognizing the micro-steps of the dance activities than traditional feature engineering approaches. These metrics are derived by purposely evaluating the setup on a dance form known for its gentle, smooth and subtle limb movements. In this paper, we articulate how our proposed DAR framework will be generalizable for diverse dance styles involving very pronounced movements, human body kinematics and energy profiles.","PeriodicalId":213775,"journal":{"name":"GetMobile Mob. Comput. Commun.","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125247663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Research on ARM TrustZone","authors":"Wenhao Li, Yubin Xia, Haibo Chen","doi":"10.1145/3308755.3308761","DOIUrl":"https://doi.org/10.1145/3308755.3308761","url":null,"abstract":"ARM TrustZone [1] is a hardware-based security feature that can provide software with a high-privilege and isolated execution environment. Such isolation is ensured by hardware, which is usually considered as more trustworthy than software. Thus the execution environment is also known as trusted execution environment (TEE). TrustZone technology was proposed in 2002, but did not get widely used until 2009, when Apple released iPhone 5s. In iPhone 5s, Apple leveraged TrustZone to protect its Touch ID, which ensures that even if the iOS is fully compromised, the user's fingerprint data can still be safe. In 2017, Google made TEE a mandatory requirement on any Android devices with a fingerprint scanner. Nowadays, almost all mobile phones and tablets have TEE deployed. Meanwhile, ARM integrates TrustZone in ARM64 and ARMv8-M to support a broader range of platforms including servers and IoT devices.","PeriodicalId":213775,"journal":{"name":"GetMobile Mob. Comput. Commun.","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122037997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Ultra-Low-Power Gaze Tracking for Virtual Reality","authors":"Tianxing Li, Qiang Liu, Xia Zhou","doi":"10.1145/3308755.3308765","DOIUrl":"https://doi.org/10.1145/3308755.3308765","url":null,"abstract":"LiGaze is a low-cost, low-power approach to gaze tracking tailored to virtual reality (VR). It relies on a few low-cost photodiodes, eliminating the need for cameras and active infrared emitters. Reusing light from the VR screen, LiGaze leverages photodiodes around a VR lens to measure reflected screen light in different directions. It then infers gaze direction by exploiting the pupil's light absorption property. The core of LiGaze is to deal with screen light dynamics and extract changes in reflected light related to pupil movement. LiGaze infers a 3D gaze vector on the fly using a lightweight regression algorithm. Compared to the eye tracker of an existing VR headset (FOVE), LiGaze achieves 6.3° and 10.1° mean within-user and cross-user accuracy. Its sensing and computation consume 791 W in total and thus can be completely powered by a credit card-size solar cell harvesting energy from indoor lighting. LiGaze's simplicity and ultra-low power make it applicable in a wide range of VR headsets to better unleash VR's potential.","PeriodicalId":213775,"journal":{"name":"GetMobile Mob. Comput. Commun.","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116151879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Mobile Augmented Reality: Exploring a new genre of learning","authors":"Breanne K. Litts, Whitney Lewis","doi":"10.1145/3308755.3308757","DOIUrl":"https://doi.org/10.1145/3308755.3308757","url":null,"abstract":"The proliferation of smartphones provides easy access to experience augmented reality (AR), which has fundamentally shifted the conversation around educational technology both in and out of classrooms. Early research on this new genre of teaching and learning afforded by AR has shown to improve and/or increase learning performance, learning motivation, student engagement, and positive attitudes (Bacca et al., 2014). As mobile technologies reach ubiquity, educators have become exceptionally concerned with designing tools and activities that equip young people to engage with these technologies as producers not just consumers. Of equal significance is the critical issue of ensuring that all young people have equitable access to not only the technologies but also to be able to participate in creating the technologies (Kafai & Burke, 2013). In this column, our goal is to present one approach by which we can empower all young people to produce with mobile augmented reality technologies.","PeriodicalId":213775,"journal":{"name":"GetMobile Mob. Comput. Commun.","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127201081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MOLECULAR COMMUNICATION: Interconnecting Tiny NanoBio Devices","authors":"N. Farsad","doi":"10.1145/3276145.3276147","DOIUrl":"https://doi.org/10.1145/3276145.3276147","url":null,"abstract":"Recent advances in the fi elds of bioengineering and nanotechnology have resulted in the emergence of tiny devices of sub-millimeter and even micron or less dimensions that can perform sensing and actuation. In many cases, the main challenge in moving these devices out of the laboratory and into the real world is not production cost, as they can be produced cost-eff ectively in large volumes, but rather a communication problem. For many applications, these tiny devices need to communicate and collaborate in swarms, or they need to transmit their measurements to other devices. Inspired by nature, chemical signaling (also known as molecular communication) is an eff ective solution to this problem. Th is article explores some of the recent advancements and challenges in engineering molecular communication systems.","PeriodicalId":213775,"journal":{"name":"GetMobile Mob. Comput. Commun.","volume":"EM-17 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132836013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}