Valentin Baumann, Johannes Dambacher, Marit F L Ruitenberg, Judith Schomaker, Kerstin Krauel
{"title":"Towards a characterization of human spatial exploration behavior.","authors":"Valentin Baumann, Johannes Dambacher, Marit F L Ruitenberg, Judith Schomaker, Kerstin Krauel","doi":"10.3758/s13428-024-02581-3","DOIUrl":"10.3758/s13428-024-02581-3","url":null,"abstract":"<p><p>Spatial exploration is a complex behavior that can be used to gain information about developmental processes, personality traits, or mental disorders. Typically, this is done by analyzing movement throughout an unknown environment. However, in human research, until now there has been no overview on how to analyze movement trajectories with regard to exploration. In the current paper, we provide a discussion of the most common movement measures currently used in human research on spatial exploration, and suggest new indices to capture the efficiency of exploration. We additionally analyzed a large dataset (n = 409) of human participants exploring a novel virtual environment to investigate whether movement measures could be assigned to meaningful higher-order components. Hierarchical clustering of the different measures revealed three different components of exploration (exploratory behavior, spatial shape, and exploration efficiency) that in part replicate components of spatial exploratory behavior identified in animal studies. A validation of our analysis on a second dataset (n = 102) indicated that two of these clusters are stable across different contexts as well as participant samples. For the exploration efficiency cluster, our validation showed that it can be further differentiated into a goal-directed versus a general, area-directed component. By also sharing data and code for our analyses, our results provide much-needed tools for the systematic analysis of human spatial exploration behavior.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"65"},"PeriodicalIF":4.6,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11754322/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143021948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zhongqing Jiang, Yanling Long, Xi'e Zhang, Yangtao Liu, Xue Bai
{"title":"CNEV: A corpus of Chinese nonverbal emotional vocalizations with a database of emotion category, valence, arousal, and gender.","authors":"Zhongqing Jiang, Yanling Long, Xi'e Zhang, Yangtao Liu, Xue Bai","doi":"10.3758/s13428-024-02595-x","DOIUrl":"https://doi.org/10.3758/s13428-024-02595-x","url":null,"abstract":"<p><p>Nonverbal emotional vocalizations play a crucial role in conveying emotions during human interactions. Validated corpora of these vocalizations have facilitated emotion-related research and found wide-ranging applications. However, existing corpora have lacked representation from diverse cultural backgrounds, which may limit the generalizability of the resulting theories. The present paper introduces the Chinese Nonverbal Emotional Vocalization (CNEV) corpus, the first nonverbal emotional vocalization corpus recorded and validated entirely by Mandarin speakers from China. The CNEV corpus contains 2415 vocalizations across five emotion categories: happiness, sadness, fear, anger, and neutrality. It also includes a database containing subjective evaluation data on emotion category, valence, arousal, and speaker gender, as well as the acoustic features of the vocalizations. Key conclusions drawn from statistical analyses of perceptual evaluations and acoustic analysis include the following: (1) the CNEV corpus exhibits adequate reliability and high validity; (2) perceptual evaluations reveal a tendency for individuals to associate anger with male voices and fear with female voices; (3) acoustic analysis indicates that males are more effective at expressing anger, while females excel in expressing fear; and (4) the observed perceptual patterns align with the acoustic analysis results, suggesting that the perceptual differences may stem not only from the subjective factors of perceivers but also from objective expressive differences in the vocalizations themselves. For academic research purposes, the CNEV corpus and database are freely available for download at https://osf.io/6gy4v/ .</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"62"},"PeriodicalIF":4.6,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jane E Bairnsfather, Miriam A Mosing, Margaret S Osborne, Sarah J Wilson
{"title":"Conceptual coherence but methodological mayhem: A systematic review of absolute pitch phenotyping.","authors":"Jane E Bairnsfather, Miriam A Mosing, Margaret S Osborne, Sarah J Wilson","doi":"10.3758/s13428-024-02577-z","DOIUrl":"10.3758/s13428-024-02577-z","url":null,"abstract":"<p><p>Despite extensive research on absolute pitch (AP), there remains no gold-standard task to measure its presence or extent. This systematic review investigated the methods of pitch-naming tasks for the classification of individuals with AP and examined how our understanding of the AP phenotype is affected by variability in the tasks used to measure it. Data extracted from 160 studies (N = 23,221 participants) included (i) the definition of AP, (ii) task characteristics, (iii) scoring method, and (iv) participant scores. While there was near-universal agreement (99%) in the conceptual definition of AP, task characteristics such as stimulus range and timbre varied greatly. Ninety-five studies (59%) specified a pitch-naming accuracy threshold for AP classification, which ranged from 20 to 100% (mean = 77%, SD = 20), with additional variability introduced by 31 studies that assigned credit to semitone errors. When examining participants' performance rather than predetermined thresholds, mean task accuracy (not including semitone errors) was 85.9% (SD = 10.8) for AP participants and 17.0% (SD = 10.5) for non-AP participants. This review shows that the characterisation of the AP phenotype varies based on methodological choices in tasks and scoring, limiting the generalisability of individual studies. To promote a more coherent approach to AP phenotyping, recommendations about the characteristics of a gold-standard pitch-naming task are provided based on the review findings. Future work should also use data-driven techniques to characterise phenotypic variability to support the development of a taxonomy of AP phenotypes to advance our understanding of its mechanisms and genetic basis.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"61"},"PeriodicalIF":4.6,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11750914/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ryan J Fitzgerald, Eva Rubínová, Eva Ribbers, Stefana Juncu
{"title":"Eyewitness Lineup Identity (ELI) database: Crime videos and mugshots for eyewitness identification research.","authors":"Ryan J Fitzgerald, Eva Rubínová, Eva Ribbers, Stefana Juncu","doi":"10.3758/s13428-024-02585-z","DOIUrl":"10.3758/s13428-024-02585-z","url":null,"abstract":"<p><p>There is a long history of experimental research on eyewitness identification, and this typically involves staging a crime for participants to witness and then testing their memory of the \"culprit\" by administering a lineup of mugshots. We created an Eyewitness Lineup Identity (ELI) database, which includes crime videos and mugshot images of 231 identities. We arranged the mugshots into 6-, 9-, and 12-member lineups, and then we tested the stimuli in an eyewitness experiment. Participants (N = 1584) completed six trials of viewing a crime video and completing a lineup identification task. In lineups that included the culprit, the average probability of correction identification was 59.0%, 95% CI [55.9, 62.0]. In lineups that did not include the culprit, the average probability of false alarm was 29.9% [27.8, 32.0]. These outcomes indicate that the ELI database is suitable for eyewitness identification research, and the large number of crime videos would enable stimulus sampling. The database is available for research approved by a research ethics board and can be requested at https://osf.io/vrj3u .</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"63"},"PeriodicalIF":4.6,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11750919/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sonia Amado, Murat C Karataş, Elif Yüvrük, Aycan Kapucu
{"title":"EGEFACE: A new face memory test with static and dynamic images.","authors":"Sonia Amado, Murat C Karataş, Elif Yüvrük, Aycan Kapucu","doi":"10.3758/s13428-024-02592-0","DOIUrl":"10.3758/s13428-024-02592-0","url":null,"abstract":"<p><p>Face memory is a crucial cognitive ability necessary for maintaining a healthy social life. Recent studies reveal large individual differences in face recognition ability. Face memory tests are used to evaluate this ability. The main purpose of this study was to develop a new face memory test (EGEFACE) addressing the limitations of existing tests using both static and dynamic stimuli to increase ecological validity; employing face recognition algorithms to adjust test difficulty; measuring face memory accuracy independently of response bias by including both target-absent and target-present trials and using ROC analysis; and developing a test to measure both ends of the face recognition ability spectrum. After building a new database of static and dynamic faces, we created three difficulty levels using a face recognition algorithm. We collected data from 703 participants in two steps and examined the internal consistency, split-half reliability, and item-total score correlations. The reliability analysis confirmed that both target-absent and target-present trials of EGEFACE were reliable. High EGEFACE performers scored near super recognizer levels on CFMT+, while low performers showed limited overlap with prosopagnosic-level performance on CFMT+, suggesting EGEFACE's sensitivity across different levels of face recognition ability. Overall, results indicated a moderate positive correlation between EGEFACE and CFMT+, showing that both tests assess similar cognitive skills, while a low to moderate correlation with KFMT suggests that EGEFACE measures cognitive ability that is related to yet distinct from face perception. The results suggest that EGEFACE shows promise as an ecologically valid and effective alternative tool for assessing individual differences in face memory.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"64"},"PeriodicalIF":4.6,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11750894/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Estimating correlations in low-reliability settings with constrained hierarchical models.","authors":"Mahbod Mehrvarz, Jeffrey N Rouder","doi":"10.3758/s13428-024-02568-0","DOIUrl":"https://doi.org/10.3758/s13428-024-02568-0","url":null,"abstract":"<p><p>It is popular to study individual differences in cognition with experimental tasks, and the main goal of such approaches is to analyze the pattern of correlations across a battery of tasks and measures. One difficulty is that experimental tasks are often low in reliability as effects are small relative to trial-by-trial variability. Consequently, it remains difficult to accurately estimate correlations. One approach that seems attractive is hierarchical modeling where trial-by-trial variability and variability across conditions, tasks, and individuals are modeled separately. Here we show that hierarchical models may reduce the error in estimating correlations up to 43%, but only if substantive constraint is imposed. The approach here is Bayesian, and we develop novel Bayesian hierarchical factor models for experiments where trials are nested in conditions, tasks, and individuals. The prior on covariances across tasks can either be unconstrained, in which there is little error reduction, or constrained, in which there is substantial error reduction. The constraints are: (1) There is a low-dimension factor structure underlying the covariation across tasks, and (2) all loadings are non-negative leading to a positive manifold on correlations. We argue that both of these assumptions are reasonable in cognitive domains, and that with them, researchers may profitably use hierarchical models to estimate correlations across tasks in low-reliability settings.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"59"},"PeriodicalIF":4.6,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Beijing Sentence Corpus II: A cross-script comparison between traditional and simplified Chinese sentence reading.","authors":"Ming Yan, Jinger Pan, Reinhold Kliegl","doi":"10.3758/s13428-024-02523-z","DOIUrl":"10.3758/s13428-024-02523-z","url":null,"abstract":"<p><p>We introduce a sentence corpus with eye-movement data in traditional Chinese (TC), based on the original Beijing Sentence Corpus (BSC) in simplified Chinese (SC). The most noticeable difference between TC and SC character sets is their visual complexity. There are reaction time corpora in isolated TC character/word lexical decision and naming tasks. However, up to now natural TC sentence reading corpus with recorded eye movements has not been available for general public. We report effects of word frequency, visual complexity, and predictability on eye movements on fixation location and duration based on 60 native TC readers. In addition, because the current BSC-II sentences are nearly identical to the original BSC sentences, we report similarities and differences of the linguistic influences on eye movements for the two varieties of written Chinese. The results shed light on how visual complexity affects eye movements. Together, the two sentence corpora comprise a useful tool to establish cross-script similarities and differences in TC and SC.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 2","pages":"60"},"PeriodicalIF":4.6,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11748476/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142999269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Predicting image memorability from evoked feelings.","authors":"Cheyenne Wakeland-Hart, Mariam Aly","doi":"10.3758/s13428-024-02510-4","DOIUrl":"https://doi.org/10.3758/s13428-024-02510-4","url":null,"abstract":"<p><p>While viewing a visual stimulus, we often cannot tell whether it is inherently memorable or forgettable. However, the memorability of a stimulus can be quantified and partially predicted by a collection of conceptual and perceptual factors. Higher-level properties that represent the \"meaningfulness\" of a visual stimulus to viewers best predict whether it will be remembered or forgotten across a population. Here, we hypothesize that the feelings evoked by an image, operationalized as the valence and arousal dimensions of affect, significantly contribute to the memorability of scene images. We ran two complementary experiments to investigate the influence of affect on scene memorability, in the process creating a new image set (VAMOS) of hundreds of natural scene images for which we obtained valence, arousal, and memorability scores. From our first experiment, we found memorability to be highly reliable for scene images that span a wide range of evoked arousal and valence. From our second experiment, we found that both valence and arousal are significant but weak predictors of image memorability. Scene images were most memorable if they were slightly negatively valenced and highly arousing. Images that were extremely positive or unarousing were most forgettable. Valence and arousal together accounted for less than 8% of the variance in image memorability. These findings suggest that evoked affect contributes to the overall memorability of a scene image but, like other singular predictors, does not fully explain it. Instead, memorability is best explained by an assemblage of visual features that combine, in perhaps unintuitive ways, to predict what is likely to stick in our memory.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"58"},"PeriodicalIF":4.6,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142982532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A note on using random starting values in small sample SEM.","authors":"Julie De Jonckere, Yves Rosseel","doi":"10.3758/s13428-024-02543-9","DOIUrl":"https://doi.org/10.3758/s13428-024-02543-9","url":null,"abstract":"<p><p>Model estimation for SEM analyses in commonly used software typically involves iterative optimization procedures, which can lead to nonconvergence issues. In this paper, we propose using random starting values as an alternative to the current default strategies. By drawing from uniform distributions within data-driven lower and upper bounds (see De Jonckere et al. (2022) Structural Equation Modeling: A Multidisciplinary Journal, 29(3), 412-427), random starting values are generated for each (free) parameter in the model. Through three small simulation studies, we demonstrate that incorporating such bounded random starting values significantly reduces the nonconvergence rate, resulting in increased convergence rates ranging between 87% and 96% in the first two studies. In essence, bounded random starting values seem to offer a promising alternative to the default starting values that are currently used in most software packages.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"57"},"PeriodicalIF":4.6,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142982530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wanjing Anya Ma, Adam Richie-Halford, Amy K Burkhardt, Klint Kanopka, Clementine Chou, Benjamin W Domingue, Jason D Yeatman
{"title":"ROAR-CAT: Rapid Online Assessment of Reading ability with Computerized Adaptive Testing.","authors":"Wanjing Anya Ma, Adam Richie-Halford, Amy K Burkhardt, Klint Kanopka, Clementine Chou, Benjamin W Domingue, Jason D Yeatman","doi":"10.3758/s13428-024-02578-y","DOIUrl":"10.3758/s13428-024-02578-y","url":null,"abstract":"<p><p>The Rapid Online Assessment of Reading (ROAR) is a web-based lexical decision task that measures single-word reading abilities in children and adults without a proctor. Here we study whether item response theory (IRT) and computerized adaptive testing (CAT) can be used to create a more efficient online measure of word recognition. To construct an item bank, we first analyzed data taken from four groups of students (N = 1960) who differed in age, socioeconomic status, and language-based learning disabilities. The majority of item parameters were highly consistent across groups (r = .78-.94), and six items that functioned differently across groups were removed. Next, we implemented a JavaScript CAT algorithm and conducted a validation experiment with 485 students in grades 1-8 who were randomly assigned to complete trials of all items in the item bank in either (a) a random order or (b) a CAT order. We found that, to achieve reliability of 0.9, CAT improved test efficiency by 40%: 75 CAT items produced the same standard error of measurement as 125 items in a random order. Subsequent validation in 32 public school classrooms showed that an approximately 3-min ROAR-CAT can achieve high correlations (r = .89 for first grade, r = .73 for second grade) with alternative 5-15-min individually proctored oral reading assessments. Our findings suggest that ROAR-CAT is a promising tool for efficiently and accurately measuring single-word reading ability. Furthermore, our development process serves as a model for creating adaptive online assessments that bridge research and practice.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"56"},"PeriodicalIF":4.6,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11732908/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142982534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}