{"title":"Web accessibility barriers and their cross-disability impact in eSystems: A scoping review","authors":"","doi":"10.1016/j.csi.2024.103923","DOIUrl":"10.1016/j.csi.2024.103923","url":null,"abstract":"<div><p>Accessibility is an important component in the implementation of Web systems to ensure that these are usable, engaging, and enjoyable by all regardless of the level of ability, condition, or circumstances. Despite manifold efforts, the Web is still largely inaccessible for a plurality of reasons (<em>e.g.</em> poor navigation, lack of/unsuitable alternative text, complex Web forms) with significant impact on disabled users. The impact of Web accessibility barriers varies per disability, but current measures for the impact of barriers treat disabilities as a homogeneous group. In this work, we present a scoping review of the Web accessibility research landscape. Following a structured approach, 112 studies were reviewed, and findings are reported on common Web accessibility barriers and practices within the Web Accessibility Lifecycle. An assessment framework is further proposed to measure the impact of such barriers across disabled groups. Finally, the need for extensive qualitative research into organizational change and multinational studies on Web activity and disturbance by barriers per disabled user group are discussed as future avenues for accessibility research.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0920548924000928/pdfft?md5=f921d53ae3864451cd5b1e92e1a7c0f1&pid=1-s2.0-S0920548924000928-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142229628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"GSParLib: A multi-level programming interface unifying OpenCL and CUDA for expressing stream and data parallelism","authors":"","doi":"10.1016/j.csi.2024.103922","DOIUrl":"10.1016/j.csi.2024.103922","url":null,"abstract":"<div><p>The evolution of Graphics Processing Units (GPUs) has allowed the industry to overcome long-lasting problems and challenges. Many belong to the stream processing domain, whose central aspect is continuously receiving and processing data from streaming data producers such as cameras and sensors. Nonetheless, programming GPUs is challenging because it requires deep knowledge of many-core programming, mechanisms and optimizations for GPUs. Current GPU programming standards do not target stream processing and present programmability and code portability limitations. Among our main scientific contributions resides <span><span>GSParLib</span></span>, a C++ multi-level programming interface unifying <span>CUDA</span> and <span>OpenCL</span> for GPU processing on stream and data parallelism with negligible performance losses compared to manual implementations; <span><span>GSParLib</span></span> is organized in two layers: one for general-purpose computing and another for high-level structured programming based on parallel patterns; a methodology to provide unified and driver agnostic interfaces minimizing performance losses; a set of parallelism strategies and optimizations for GPU processing targeting stream and data parallelism; and new experiments covering GPU performance on applications exposing stream and data parallelism.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142122719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"TDBAMLA: Temporal and dynamic behavior analysis in Android malware using LSTM and attention mechanisms","authors":"","doi":"10.1016/j.csi.2024.103920","DOIUrl":"10.1016/j.csi.2024.103920","url":null,"abstract":"<div><p>The increasing ubiquity of Android devices has precipitated a concomitant surge in sophisticated malware attacks, posing critical challenges to cybersecurity infrastructures worldwide. Existing models have achieved significant strides in malware detection but often suffer from high false-positive rates, lower recall, and computational delays, thus demanding a more efficient and accurate system. Current techniques primarily rely on static features and simplistic learning models, leading to inadequate handling of temporal aspects and dynamic behaviors exhibited by advanced malware. These limitations compromise the detection of modern, evasive malware, and impede real-time analysis. This paper introduces a novel framework for Android malware detection that incorporates Temporal and Dynamic Behavior Analysis using Long Short-Term Memory (LSTM) networks and Attention Mechanisms. We further propose development of an efficient Grey Wolf Optimized (GWO) Decision Trees to find the most salient API call patterns associated with malwares. An Iterative Fuzzy Logic (IFL) layer is also deployed before classification to assess the \"trustworthiness\" of app metadata samples. For Ongoing Learning, we propose use of Deep Q-Networks (DQNs), which helps the reinforcement learning model to adapt more quickly to changes in the threat landscapes. By focusing on crucial system calls and behavioral characteristics in real-time, our model captures the nuanced temporal patterns often exhibited by advanced malwares. Empirical evaluations demonstrate remarkable improvements across multiple performance metrics. Compared to existing models, our approach enhances the precision of malware identification by 8.5 %, accuracy by 5.5 %, and recall by 4.9 %, while also achieving an 8.3 % improvement in the Area Under the Receiver Operating Characteristic Curve (AUC), with higher specificity and a 4.5 % reduction in identification delay. In malware pre-emption tasks, our model outperforms by improving precision by 4.3 %, accuracy by 3.9 %, recall by 4.9 %, AUC by 3.5 %, and increasing specificity by 2.9 %. These gains make our framework highly applicable for real-time detection systems, cloud-based security solutions, and threat intelligence services, thereby contributing to a safer Android ecosystem.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142151842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Who are the best contributors? Designing a multimodal science communication interface based on the ECM, TAM and the Taguchi methods","authors":"","doi":"10.1016/j.csi.2024.103921","DOIUrl":"10.1016/j.csi.2024.103921","url":null,"abstract":"<div><p>Science communication conducted through mobile devices and mobile applications is an efficient and widespread phenomenon that requires communicators and design practitioners to further develop suitable design elements and strategies for such platforms. The effective application of multimodal or multisensory design in interfaces provides users with rich experiences. However, there is a lack of guiding recommendations for user interface design in the citizen science community. This study investigated factors affecting users’ perceptions and behavioral intentions toward multimodal scientific communication interface designs and identified the optimal combinations of such factors for such designs. Through a focus group, we defined three design dimensions of a science communication interface: visual, auditory, and haptic. An online experiment involving 916 participants was then conducted and integrated the technology acceptance model, expectation–confirmation model, and Taguchi method to examine the hierarchical combinations with the greatest influence in each dimension. The results indicated that interface design combinations primarily focusing on visual elements, with auditory and haptic as secondary elements, can serve as effective tools for science communication. Moreover, layout, color tones, vibration intensity, and sound volume significantly affected users’ perceptions and behavioral intentions. As one of the few studies using the Taguchi method to explore the design of science communication interfaces, the present findings enrich the multimodal theory from the perspectives of design and communication, highlighting its value in science communication. This paper simultaneously provides insights into how to select and combine multimodal design elements in science communication interfaces, demonstrating the potential of such designs to affect a user perception, satisfaction, confirmation, and continued usage intention.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142151244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A cellular automata based secured reversible data hiding scheme for dual images using bit-reversal permutation technique","authors":"","doi":"10.1016/j.csi.2024.103919","DOIUrl":"10.1016/j.csi.2024.103919","url":null,"abstract":"<div><p>With the rapid development of advanced communication technology, protection of confidential data during transmission through public channel has become a challenging issue. In this context, the design of a data hiding scheme needs to ensure reversibility, robustness against various malicious attacks, and unaltered visual quality even after embedding high amount of secret data. To meet the above requirements, Cellular Automata along with Bit-Reversal Permutation technique have been utilized on dual-image with the target to enhance the robustness of suggested scheme due to distribution of secret information within two stego images which is hard to extract without both stego simultaneously. The proposed scheme makes a trade-off among visual quality, security and embedding capacity as essential for ensuring innocuous communication. The experimental results and comparison with the state-of-art methods establishes that the proposed scheme ensures high degree of robustness against different venomous attacks. This approach may be beneficial to private and public sector practitioners and government agencies to protect valuable multimedia secret data from adversarial cyber attacks.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Authentication communication by using visualization cryptography for UAV networks","authors":"","doi":"10.1016/j.csi.2024.103918","DOIUrl":"10.1016/j.csi.2024.103918","url":null,"abstract":"<div><p>Utilizing unmanned aerial vehicles (UAVs) to support V2X requirements leverages their versatility and line-of-sight communication. Our prior work explored a multi-agent learning approach for resource optimization, maximizing task offloading while maintaining QoS. This paper focuses on securing UAV communication, particularly authentication. Traditional methods are often unsuitable due to UAV limitations. We propose a novel authentication mechanism for a single ground control station (GCS) interacting with multiple UAVs across flight sessions. The system utilizes a key generation method based on chaotic maps to create unique flight session keys for each pre-defined flight plan. These keys, along with flight plans, are registered in a secure database. During flight, the GCS verifies UAV identity by employing the flight session key and corresponding flight plan for message authentication. This approach reduces computational and communication overhead compared to traditional certificate exchanges and asymmetric cryptography, which are energy-intensive for UAVs. While not a comprehensive security solution, this method provides an initial layer of protection for the UAV network.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142002235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Large language models for code completion: A systematic literature review","authors":"","doi":"10.1016/j.csi.2024.103917","DOIUrl":"10.1016/j.csi.2024.103917","url":null,"abstract":"<div><p>Code completion serves as a fundamental aspect of modern software development, improving developers' coding processes. Integrating code completion tools into an Integrated Development Environment (IDE) or code editor enhances the coding process and boosts productivity by reducing errors and speeding up code writing while reducing cognitive load. This is achieved by predicting subsequent tokens, such as keywords, variable names, types, function names, operators, and more. Different techniques can achieve code completion, and recent research has focused on Deep Learning methods, particularly Large Language Models (LLMs) utilizing Transformer algorithms. While several research papers have focused on the use of LLMs for code completion, these studies are fragmented, and there is no systematic overview of the use of LLMs for code completion. Therefore, we aimed to perform a Systematic Literature Review (SLR) study to investigate how LLMs have been applied for code completion so far. We have formulated several research questions to address how LLMs have been integrated for code completion-related tasks and to assess the efficacy of these LLMs in the context of code completion. To achieve this, we retrieved 244 papers from scientific databases using auto-search and specific keywords, finally selecting 23 primary studies based on an SLR methodology for in-depth analysis. This SLR study categorizes the granularity levels of code completion achieved by utilizing LLMs in IDEs, explores the existing issues in current code completion systems, how LLMs address these challenges, and the pre-training and fine-tuning methods employed. Additionally, this study identifies open research problems and outlines future research directions. Our analysis reveals that LLMs significantly enhance code completion performance across several programming languages and contexts, and their capability to predict relevant code snippets based on context and partial input boosts developer productivity substantially.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0920548924000862/pdfft?md5=8314175bcf57b70427dd5c869ea42978&pid=1-s2.0-S0920548924000862-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142076551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Best of two worlds: Efficient, usable and auditable biometric ABC on the blockchain","authors":"","doi":"10.1016/j.csi.2024.103916","DOIUrl":"10.1016/j.csi.2024.103916","url":null,"abstract":"<div><p>In García-Rodríguez et al. 2024, two generic constructions for biometric-based non-transferable Attribute Based Credentials (biometric ABC) are presented, which offer different trade-offs between efficiency and trust assumptions. In this paper, we focus on the second scheme denoted as BioABC-ZK that tries to remove the strong (and unrealistic) trust assumption on the Reader R, and we show that BioABC-ZK has a security flaw for a colluding R and Verifier V. Besides, BioABC-ZK lacks GDPR-compliance, which requires secure processing of biometrics, for instance in form of Fuzzy Extractors, as opposed to (<span><math><mi>i</mi></math></span>) storing the reference biometric template <span><math><msub><mrow><mi>a</mi></mrow><mrow><mi>B</mi><mi>i</mi><mi>o</mi></mrow></msub></math></span> in the user’s mobile phone and (<span><math><mrow><mi>i</mi><mi>i</mi></mrow></math></span>) processing of biometrics using an external untrusted R, whose foreign manufacturers are unlikely to adjust their products according to GDPR.</p><p>The contributions of this paper are threefold. First, we review efficient biometric ABC schemes to identify the privacy-by-design criteria for them. In view of these principles, we propose a new architecture for <em>biometric ABC</em> of Sarier 2021 by adapting the recently introduced <em>core/helper setting</em>. Briefly, a user in our modified setting is composed of a constrained core device (a SIM card) inside a helper device (a smart phone with dual SIM and face recognition feature), which – as opposed to García-Rodríguez et al. 2024 – does not need to store <span><math><msub><mrow><mi>a</mi></mrow><mrow><mi>B</mi><mi>i</mi><mi>o</mi></mrow></msub></math></span>. This way, the new design provides <em>Identity Privacy</em> without the need for an external R and/or a dedicated hardware per user such as a biometric smart card reader or a tamper proof smart card as in current hardware-bound credential systems. Besides, the new system maintains minimal hardware requirements on the SIM card – only responsible for storing ABC and helper data –, which results in easy adoption and usability without loosing efficiency, if deep face fuzzy vault and our modified ABC scheme are employed together. As a result, a total overhead of 500 ms to a showing of a comparable non-biometric ABC is obtained instead of the 2.1 s in García-Rodríguez et al. 2024 apart from the removal of computationally expensive pairings. Finally, as different from García-Rodríguez et al. 2024, auditing is achieved via Blockchain instead of proving in zero-knowledge the actual biometric matching by the user to reveal malicious behavior of R and V.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141979792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A theory on human factors in DevOps adoption","authors":"","doi":"10.1016/j.csi.2024.103907","DOIUrl":"10.1016/j.csi.2024.103907","url":null,"abstract":"<div><h3>Context:</h3><p>DevOps is a software engineering paradigm that enables faster deliveries and higher quality products. However, DevOps adoption is a complex process that is still insufficiently supported by research. In addition, human factors are the main difficulty for a successful DevOps adoption, although very few studies address this topic.</p></div><div><h3>Objective:</h3><p>This paper addresses two research gaps identified in literature, namely: (1) the characterization of DevOps from the perspective of human factors, i.e. the description of DevOps’ human characteristics to better define it, and (2) the identification and analysis of human factors’ effect in the adoption of DevOps.</p></div><div><h3>Method:</h3><p>We employed a hybrid methodology that included a Systematic Mapping Study followed by the application of a clustering technique. A questionnaire for DevOps practitioners (<span><math><mrow><mi>n</mi><mo>=</mo><mn>15</mn></mrow></math></span>) was employed as an evaluation method.</p></div><div><h3>Results:</h3><p>A total of 59 human factors related to DevOps were identified, described, and synthesized. The results were used to build a theory on DevOps human factors.</p></div><div><h3>Conclusion:</h3><p>The main contribution of this paper is a theory proposal regarding human factors in DevOps adoption. The evaluation results show that almost every human factor identified in the mapping study was found relevant in DevOps adoption. The results of the study represent an extension of DevOps characterization and a first approximation to human factors in DevOps adoption.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S092054892400076X/pdfft?md5=8a197a8035fc3bc559533baac7028e12&pid=1-s2.0-S092054892400076X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141993505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Energy-efficient neural network training through runtime layer freezing, model quantization, and early stopping","authors":"","doi":"10.1016/j.csi.2024.103906","DOIUrl":"10.1016/j.csi.2024.103906","url":null,"abstract":"<div><h3>Background:</h3><p>In the last years, neural networks have been massively adopted by industry and research in a wide variety of contexts. Neural network milestones are generally reached by scaling up computation, completely disregarding the carbon footprint required for the associated computations. This trend has become unsustainable given the ever-growing use of deep learning, and could cause irreversible damage to the environment of our planet if it is not addressed soon.</p></div><div><h3>Objective:</h3><p>In this study, we aim to analyze not only the effects of different energy saving methods for neural networks but also the effects of the moment of intervention, and what makes certain moments optimal.</p></div><div><h3>Method:</h3><p>We developed a novel dataset by training convolutional neural networks in 12 different computer vision datasets and applying runtime decisions regarding layer freezing, model quantization and early stopping at different epochs in each run. We then fit an auto-regressive prediction model on the data collected capable to predict the accuracy and energy consumption achieved on future epochs for different methods. The predictions on accuracy and energy are used to estimate the optimal training path.</p></div><div><h3>Results:</h3><p>Following the predictions of the model can save 56.5% of energy consumed while also increasing validation accuracy by 2.38% by avoiding overfitting.The prediction model developed can predict the validation accuracy with a 8.4% of error, the energy consumed with a 14.3% of error and the trade-off between both with a 8.9% of error.</p></div><div><h3>Conclusions:</h3><p>This prediction model could potentially be used by the training algorithm to decide which methods apply to the model and at what moment in order to maximize the accuracy-energy trade-off.</p></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":null,"pages":null},"PeriodicalIF":4.1,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0920548924000758/pdfft?md5=9fe0b023bddfc875c825b0c53e63af06&pid=1-s2.0-S0920548924000758-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}