{"title":"Predicting high variability in imageability ratings across age groups and its influence on visual word recognition.","authors":"Sandra Aka, Stéphanie Mathey","doi":"10.3758/s13428-024-02520-2","DOIUrl":"10.3758/s13428-024-02520-2","url":null,"abstract":"<p><p>Imageability, an important word characteristic in the psycholinguistic literature, is typically assessed by asking participants to estimate the ease with which a word can evoke a mental image. Our aim was to explore inter-rater disagreement in normative imageability ratings. We examined the predictors of variability around average imageability ratings for young, middle-aged and older adults (Study 1) and assessed its impact on visual word recognition performance in young adults (Study 2). Analyses of French age-related imageability ratings (Ballot et al., Behavior Research Methods, 54, 196-215, 2022) revealed that inter-rater disagreement around the average imageability value was critically high for most words within the imageability norms, thus questioning the construct validity of the average rating for the most variable items. Variability in ratings changed between age groups (18-25, 26-40, 41-59, and over 60 years) and was associated with words that are longer, less frequent, learnt later in life and less emotional (Study 1). To examine the consequences of elevated standard deviations around the average imageability rating on visual word recognition, we entered this factor in a hierarchical regression alongside classic lexico-semantic predictors. The effect of word-imageability on young adults' lexical decision times (Ferrand et al., Behavior Research Methods, 50, 1285-1307, 2018) remained significant after accounting for inter-rater disagreement in imageability ratings, even when considering the least consensual words (Study 2). We conclude that imageability ratings reliably predict visual word recognition performance in young adults for large datasets, but might require caution for smaller ones. Given imageability rating differences across adulthood, further research investigating age-related differences in language processing is necessary.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"8"},"PeriodicalIF":4.6,"publicationDate":"2024-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142799357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward an Asian-based bodily movement database for emotional communication.","authors":"Miao Cheng, Chia-Huei Tseng, Ken Fujiwara, Shoi Higashiyama, Abby Weng, Yoshifumi Kitamura","doi":"10.3758/s13428-024-02558-2","DOIUrl":"10.3758/s13428-024-02558-2","url":null,"abstract":"<p><p>Most current databases for bodily emotion expression are created in Western countries, resulting in culturally skewed representations. To address the obvious risk this bias poses to academic comprehension, we attempted to expand the current repertoire of human bodily emotions by recruiting Asian professional performers to wear whole-body suits with 57 retroreflective markers attached to major joints and body segments, and express seven basic emotions with whole-body movements in a motion-capture lab. For each emotion, actors performed three self-created scenarios that covered a broad range of real-life events to elicit the target emotion within 2-5 seconds. Subsequently, a separate group of participants was invited to judge the perceived emotional category from the extracted biological motions (point-light displays with 18 or 57 markers). The results demonstrated that the emotion discrimination accuracy was comparable to Western databases containing standardized performance scenarios. The results provide a significant step toward establishing a database using a novel emotional induction approach based on personalized scenarios. This database will contribute to a more comprehensive understanding of emotional expression across diverse contexts.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"10"},"PeriodicalIF":4.6,"publicationDate":"2024-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11632091/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142799363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The power of effect size stabilization.","authors":"Benjamin Kowialiewski","doi":"10.3758/s13428-024-02549-3","DOIUrl":"10.3758/s13428-024-02549-3","url":null,"abstract":"<p><p>Determining an appropriate sample size in psychological experiments is a common challenge, requiring a balance between maximizing the chance of detecting a true effect (minimizing false negatives) and minimizing the risk of observing an effect where none exists (minimizing false positives). A recent study proposes using effect size stabilization, a form of optional stopping, to define sample size without increasing the risk of false positives. In effect size stabilization, researchers monitor the effect size of their samples throughout the sampling process and stop sampling when the effect no longer varies beyond predefined thresholds. This study aims to improve our understanding of effect size stabilization properties. Simulations involving effect size stabilization are presented, with parametric modulation of the true effect in the population and the strictness of the stabilization rule. As previously demonstrated, the results indicate that optional stopping based on effect-size stabilization consistently yields unbiased samples over the long run. However, simulations also reveal that effect size stabilization does not guarantee the detection of a true effect in the population. Consequently, researchers adopting effect size stabilization put themselves at risk of increasing type 2 error probability. Instead of using effect-size stabilization procedures for testing, researchers should use them to reach accurate parameter estimates.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"7"},"PeriodicalIF":4.6,"publicationDate":"2024-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142799359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sharif I Kronemer, Victoria E Gobo, Catherine R Walsh, Joshua B Teves, Diana C Burk, Somayeh Shahsavarani, Javier Gonzalez-Castillo, Peter A Bandettini
{"title":"Cross-species real-time detection of trends in pupil size fluctuation.","authors":"Sharif I Kronemer, Victoria E Gobo, Catherine R Walsh, Joshua B Teves, Diana C Burk, Somayeh Shahsavarani, Javier Gonzalez-Castillo, Peter A Bandettini","doi":"10.3758/s13428-024-02545-7","DOIUrl":"10.3758/s13428-024-02545-7","url":null,"abstract":"<p><p>Pupillometry is a popular method because pupil size is easily measured and sensitive to central neural activity linked to behavior, cognition, emotion, and perception. Currently, there is no method for online monitoring phases of pupil size fluctuation. We introduce rtPupilPhase-an open-source software that automatically detects trends in pupil size in real time. This tool enables novel applications of real-time pupillometry for achieving numerous research and translational goals. We validated the performance of rtPupilPhase on human, rodent, and monkey pupil data, and we propose future implementations of real-time pupillometry.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"9"},"PeriodicalIF":4.6,"publicationDate":"2024-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11632003/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142799354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Affective Norms for German as a Second Language (ANGL2).","authors":"Zeming Xu, Jia Liu, Lin Fan","doi":"10.3758/s13428-024-02539-5","DOIUrl":"10.3758/s13428-024-02539-5","url":null,"abstract":"<p><p>The present study introduces affective norms for a set of 880 German words rated by learners of German as a second language (L2), i.e., the Affective Norms for German as a Second Language (ANGL2). The database provides ratings across affective and subjective psycholinguistic dimensions. Besides valence and arousal ratings, ANGL2 features data on emotional prototypicality, which helps to identify emotion-label words and emotion-laden words. Moreover, the database includes two additional semantic variables: concreteness and familiarity. We observed similarities with previous studies, and the ratings provided by L2 speakers demonstrate characteristics that should be noted in studies involving bilinguals, including more moderate valence ratings, and a stronger correlation between valence and arousal, specifically for positive words. ANGL2 is the first set of affective norms that has been rated by L2 speakers for a language other than English. The set of norms is aimed to function as a resource for psycholinguistic experimental studies on the intersection between emotion and language among L2 speakers.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"6"},"PeriodicalIF":4.6,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142779283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Continuous motion tracking for accurate and efficient color vision assessment.","authors":"Chenxi Liang, Jing Chen, Zhongting Chen","doi":"10.3758/s13428-024-02518-w","DOIUrl":"10.3758/s13428-024-02518-w","url":null,"abstract":"<p><p>The assessment of color vision is crucial in both fundamental visual research and clinical diagnosis. However, existing tools for color vision assessment are limited by various factors. This study introduces a novel, efficient method for color vision assessment, which is based on a continuous motion tracking task and a Kalman filter model. The effectiveness of this new method was evaluated by assessing the color vision of both color-deficient observers and normal controls. The results from both a small sample (N = 29, Experiment 1) and a large sample (N = 171, Experiment 2) showed that color-deficient observers could be perfectly identified within 20 s using the tracking performance. We also compared the new method with a traditional psychophysical detection task to examine the consistency of perceptual noise estimation between the two methods, and the results showed a moderate correlation (Pearson's r = .59 ~ .64). The results also demonstrated that the new method could measure individuals' contrast response functions of both red-green and blue-yellow colors (e.g., the L-M and S-(L + M) axes in DKL color space) in just a few minutes, showing much higher efficiency than traditional methods. All the findings from this study indicate that the continuous motion tracking method is a promising tool for both rapid screening of color vision deficiencies and fundamental research on color vision.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"3"},"PeriodicalIF":4.6,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142779287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Establishing construct validity for dynamic measures of behavior using naturalistic study designs.","authors":"Roberto C French, Daniel P Kennedy, Anne C Krendl","doi":"10.3758/s13428-024-02519-9","DOIUrl":"10.3758/s13428-024-02519-9","url":null,"abstract":"<p><p>There has been a recent surge of naturalistic methodology to assess complex topics in psychology and neuroscience. Such methods are lauded for their increased ecological validity, aiming to bridge a gap between highly controlled experimental design and purely observational studies. However, these measures present challenges in establishing construct validity. One domain in which this has emerged is research on theory of mind: the ability to infer others' thoughts and emotions. Traditional measures utilize rigid methodology which suffer from ceiling effects and may fail to fully capture how individuals engage theory of mind in everyday interactions. In the present study, we validate and test a novel approach utilizing a naturalistic task to assess theory of mind. Participants watched a mockumentary-style show while using a joystick to provide continuous, real-time theory of mind judgments. A baseline sample's ratings were used to establish a \"ground truth\" for the judgments. Ratings from separate young and older adult samples were compared against the ground truth to create similarity scores. This similarity score was compared against two independent tasks to assess construct validity: an explicit judgment performance-based paradigm, and a neuroimaging paradigm assessing response to a static measure of theory of mind. The similarity metric did not have ceiling effects and was significantly positively related to both the performance-based and neural measures. It also replicated age effects that other theory of mind measures demonstrate. Together, our multimodal approach provided convergent evidence that dynamic measures of behavior can yield robust and rigorous assessments of complex psychological processes.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"2"},"PeriodicalIF":4.6,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142765921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Gonzalo Martínez, Juan Diego Molero, Sandra González, Javier Conde, Marc Brysbaert, Pedro Reviriego
{"title":"Using large language models to estimate features of multi-word expressions: Concreteness, valence, arousal.","authors":"Gonzalo Martínez, Juan Diego Molero, Sandra González, Javier Conde, Marc Brysbaert, Pedro Reviriego","doi":"10.3758/s13428-024-02515-z","DOIUrl":"10.3758/s13428-024-02515-z","url":null,"abstract":"<p><p>This study investigates the potential of large language models (LLMs) to provide accurate estimates of concreteness, valence, and arousal for multi-word expressions. Unlike previous artificial intelligence (AI) methods, LLMs can capture the nuanced meanings of multi-word expressions. We systematically evaluated GPT-4o's ability to predict concreteness, valence, and arousal. In Study 1, GPT-4o showed strong correlations with human concreteness ratings (r = .8) for multi-word expressions. In Study 2, these findings were repeated for valence and arousal ratings of individual words, matching or outperforming previous AI models. Studies 3-5 extended the valence and arousal analysis to multi-word expressions and showed good validity of the LLM-generated estimates for these stimuli as well. To help researchers with stimulus selection, we provide datasets with LLM-generated norms of concreteness, valence, and arousal for 126,397 English single words and 63,680 multi-word expressions.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"5"},"PeriodicalIF":4.6,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142779297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Diederick C Niehorster, Will Whitham, Benjamin R Lake, Steven J Schapiro, Ian M Andolina, Jessica L Yorzinski
{"title":"Enhancing eye tracking for nonhuman primates and other subjects unable to follow instructions: Adaptive calibration and validation of Tobii eye trackers with the Titta toolbox.","authors":"Diederick C Niehorster, Will Whitham, Benjamin R Lake, Steven J Schapiro, Ian M Andolina, Jessica L Yorzinski","doi":"10.3758/s13428-024-02540-y","DOIUrl":"10.3758/s13428-024-02540-y","url":null,"abstract":"<p><p>Accurate eye tracking is crucial for gaze-dependent research, but calibrating eye trackers in subjects who cannot follow instructions, such as human infants and nonhuman primates, presents a challenge. Traditional calibration methods rely on verbal instructions, which are ineffective for these populations. To address this, researchers often use attention-grabbing stimuli in known locations; however, existing software for video-based calibration is often proprietary and inflexible. We introduce an extension to the open-source toolbox Titta-a software package integrating desktop Tobii eye trackers with PsychToolbox experiments-to facilitate custom video-based calibration. This toolbox extension offers a flexible platform for attracting attention, calibrating using flexible point selection, and validating the calibration. The toolbox has been refined through extensive use with chimpanzees, baboons, and macaques, demonstrating its effectiveness across species. Our adaptive calibration and validation procedures provide a standardized method for achieving more accurate gaze tracking, enhancing gaze accuracy across diverse species.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"4"},"PeriodicalIF":4.6,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11618154/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142779292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Philomène Périn, Santiago Herrera, Caroline Bogliotti
{"title":"FLexSign: A lexical database in French Sign Language (LSF).","authors":"Philomène Périn, Santiago Herrera, Caroline Bogliotti","doi":"10.3758/s13428-024-02521-1","DOIUrl":"10.3758/s13428-024-02521-1","url":null,"abstract":"<p><p>In psycholinguistics, studies are conducted to understand language processing mechanisms, whether in comprehension or in production, and independently of the language modality. To do so, researchers need accurate psycholinguistic information about the linguistic material they use. One main obstacle to this process is the lack of information available in sign language. While some lexical databases exist in multiple sign languages, to the best of our knowledge, psycholinguistic data for French Sign Language (LSF) signs are not yet available. The present study presents FLexSign, the first interactive lexical database for LSF, inspired by ASL-Lex (Caselli et al., 2017). The database includes familiarity, concreteness, and iconicity data for 546 signs of LSF. These three factors are known to influence the speed or the accuracy of lexical processing. Familiarity and concreteness are known to generate a robust facilitative effect on sign processing, while iconicity plays a complex but crucial role in the creation and organization of sign language lexicons. Therefore, having accurate information on the iconicity of LSF signs would help to better understand the role of this notion in lexical processing. To develop the database, 33 participants were recruited and asked to complete an online questionnaire. The FLexSign database will be of great use to sign language researchers, providing linguistic information that was previously unavailable and offering many opportunities at both the experimental and clinical levels. The database is also open to future contributions.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 1","pages":"1"},"PeriodicalIF":4.6,"publicationDate":"2024-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142765922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}