{"title":"Analyzing student perceptions and behaviors in the use of an engaging visualization dashboard for peer code review activities","authors":"Hoang-Thanh Duong , Chuan-Lin Huang , Bao-An Nguyen , Hsi-Min Chen","doi":"10.1016/j.scico.2025.103394","DOIUrl":null,"url":null,"abstract":"<div><h3>Background and context</h3><div>Traditional learning management systems often fail to sustain student motivation in computer science education, where consistent practice is crucial for skill development. This necessitates specialized tools that better support technical learning processes, foster engagement, and enable timely interventions for struggling students.</div></div><div><h3>Objectives</h3><div>This study introduces and evaluates a dashboard module integrated into peer code review activities within an \"Object-Oriented Design Laboratory\" course. The research analyzes comprehensive peer code review data collected through this dashboard to examine student review behaviors and identify at-risk students through behavioral indicators.</div></div><div><h3>Method</h3><div>A mixed-methods approach compared an experimental group (<em>n</em> = 75) using the assessment dashboard against a control group (<em>n</em> = 78) without dashboard access. Both cohorts received identical instructions over 17 weeks. Data collection included peer code review processes, programming exam results, and student surveys to analyze review behaviors and perceptions of the dashboard.</div></div><div><h3>Findings</h3><div>Students with dashboard access demonstrated significantly improved engagement and self-awareness in the peer code review activity, along with measurable performance gains. Instructors benefited from more efficient monitoring capabilities across peer code review activities.</div></div><div><h3>Implications</h3><div>Dashboard systems enhance metacognitive awareness through self-reflection and interaction monitoring, showing particular promise for programming education. These systems improve learning outcomes and instructional effectiveness by providing visual feedback that helps students track progress and identify improvement areas while giving instructors valuable insights into engagement patterns and learning behaviors.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"249 ","pages":"Article 103394"},"PeriodicalIF":1.4000,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Science of Computer Programming","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167642325001339","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
Background and context
Traditional learning management systems often fail to sustain student motivation in computer science education, where consistent practice is crucial for skill development. This necessitates specialized tools that better support technical learning processes, foster engagement, and enable timely interventions for struggling students.
Objectives
This study introduces and evaluates a dashboard module integrated into peer code review activities within an "Object-Oriented Design Laboratory" course. The research analyzes comprehensive peer code review data collected through this dashboard to examine student review behaviors and identify at-risk students through behavioral indicators.
Method
A mixed-methods approach compared an experimental group (n = 75) using the assessment dashboard against a control group (n = 78) without dashboard access. Both cohorts received identical instructions over 17 weeks. Data collection included peer code review processes, programming exam results, and student surveys to analyze review behaviors and perceptions of the dashboard.
Findings
Students with dashboard access demonstrated significantly improved engagement and self-awareness in the peer code review activity, along with measurable performance gains. Instructors benefited from more efficient monitoring capabilities across peer code review activities.
Implications
Dashboard systems enhance metacognitive awareness through self-reflection and interaction monitoring, showing particular promise for programming education. These systems improve learning outcomes and instructional effectiveness by providing visual feedback that helps students track progress and identify improvement areas while giving instructors valuable insights into engagement patterns and learning behaviors.
期刊介绍:
Science of Computer Programming is dedicated to the distribution of research results in the areas of software systems development, use and maintenance, including the software aspects of hardware design.
The journal has a wide scope ranging from the many facets of methodological foundations to the details of technical issues andthe aspects of industrial practice.
The subjects of interest to SCP cover the entire spectrum of methods for the entire life cycle of software systems, including
• Requirements, specification, design, validation, verification, coding, testing, maintenance, metrics and renovation of software;
• Design, implementation and evaluation of programming languages;
• Programming environments, development tools, visualisation and animation;
• Management of the development process;
• Human factors in software, software for social interaction, software for social computing;
• Cyber physical systems, and software for the interaction between the physical and the machine;
• Software aspects of infrastructure services, system administration, and network management.