{"title":"The role of closure in defining the \"objects\" of object-based attention.","authors":"Alexandria C. Marino, B. Scholl","doi":"10.1167/4.8.270","DOIUrl":"https://doi.org/10.1167/4.8.270","url":null,"abstract":"Many recent studies have concluded that the underlying units of visual attention are often discrete objects whose boundaries constrain the allocation of attention. However, relatively few studies have explored the particular stimulus cues that determine what counts as an \"object\" of attention. We explore this issue in the context of the two-rectangles stimuli previously used by many investigators. We first show, using both spatial-cuing and divided-attention paradigms, that same-object advantages occur even when the ends of the two rectangles are not drawn. This is consistent with previous reports that have emphasized the importance of individual contours in guiding attention, and our study shows that such effects can occur in displays that also contain grouping cues. In our divided-attention experiment, however, this contour-driven same-object advantage was significantly weaker than that obtained with the standard stimulus, with the added cue of closure--demonstrating that contour-based processes are not the whole story. These results confirm and extend the observation that same-object advantages can be observed even without full-fledged objects. At the same time, however, these studies show that boundary closure-one of the most important cues to objecthood per se-can directly influence attention. We conclude that object-based attention is not an all-or-nothing phenomenon; object-based effects can be independently strengthened or weakened by multiple cues to objecthood.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"27 1","pages":"1140-9"},"PeriodicalIF":0.0,"publicationDate":"2004-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80621892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Contextual influences on the internal structure of phonetic categories: a distinction between lexical status and speaking rate.","authors":"J. Allen, Joanne L. Miller","doi":"10.1121/1.427642","DOIUrl":"https://doi.org/10.1121/1.427642","url":null,"abstract":"Previous research has shown that phonetic categories have a graded internal structure that is highly dependent on acoustic-phonetic contextual factors, such as speaking rate; these factors alter not only the location of phonetic category boundaries, but also the location of a category's best exemplars. The purpose of the present investigation, which focused on the voiceless category as specified by voice onset time (VOT), was to determine whether a higher order linguistic contextual factor, lexical status, which is known to alter the location of the voiced-voiceless phonetic category boundary, also alters the location of the best exemplars of the voiceless category. The results indicated that lexical status has a more limited and qualitatively different effect on the category's best exemplars than does the acoustic-phonetic factor of speaking rate. This dissociation is discussed in terms of a production-based account in which perceived best exemplars of a category track contextual variation in speech production.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"10 1","pages":"798-810"},"PeriodicalIF":0.0,"publicationDate":"1999-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84242226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The representation of auditory source characteristics: simple geometric form.","authors":"Stephen Lakatos, Stephen McAdams, René Causse","doi":"10.1121/1.416150","DOIUrl":"https://doi.org/10.1121/1.416150","url":null,"abstract":"Two experiments examined listeners' ability to discriminate the geometric shape of simple resonating bodies on the basis of their corresponding auditory attributes. In cross-modal matching tasks, subjects listened to recordings of pairs of metal bars (Experiment 1) or wooden bars (Experiment 2) struck in sequence and then selected a visual depiction of the bar cross sections that correctly represented their relative widths and heights from two opposing pairs presented on a computer screen. Multidimensional scaling solutions derived from matching scores for metal and wooden bars indicated that subjects' performance varied directly with increasing differences in the width/height (W/H) ratios of both sets of bars. Subsequent acoustic analyses revealed that the frequency components from torsional vibrational modes and the ratios of frequencies of transverse bending modes in the bars correlated strongly with both the bars' W/H ratios and bar coordinates in the multidimensional configurations. The results suggest that listeners can encode the auditory properties of sound sources by extracting certain invariant physical characteristics of their gross geometric properties from their acoustic behavior.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"29 1","pages":"1180-90"},"PeriodicalIF":0.0,"publicationDate":"1996-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80038125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Distinctiveness and serial position effects in tonal sequences","authors":"A. Surprenant","doi":"10.1121/1.415043","DOIUrl":"https://doi.org/10.1121/1.415043","url":null,"abstract":"The proportion-of-the-total-duration rule (Kidd & Watson, 1992) states that the detectability of a change in a component of a tonal sequence can be predicted by the proportional duration of the changed component relative to the length of the sequence as a whole. A similar viewpoint relies on temporal distinctiveness to account for primacy, recency, and other serial position effects in memory (Murdock, 1960; Neath, 1993a, 1993b). Such distinctiveness models predict that an item will be remembered if it is more distinctive along some dimension relative to possible competitors. Three experiments explored the relation between distinctiveness and proportional duration by examining the effects of the proportion of the total duration of a tone in a sequence, serial position, and interstimulus interval (ISI) on the detection of a change in one component of a tonal sequence. Experiment 1 replicated the basic effect with relatively untrained subjects and a fixed frequency difference. Experiment 2 showed that distinctiveness holds for tonal sequences and a same/different task. Experiment 3 combined the two to show that proportional duration, ISI, and position of the changed tone all contribute to discrimination performance. The present research combines theories that have been proposed in the psychophysics and memory fields and suggests that a comprehensive principle based on relative distinctiveness may be able to account for both perceptual and memory effects.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"97 1","pages":"737-745"},"PeriodicalIF":0.0,"publicationDate":"1996-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77958220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Perception of musical tension in short chord sequences: the influence of harmonic function, sensory dissonance, horizontal motion, and musical training.","authors":"E Bigand, R Parncutt, F Lerdahl","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>This study investigates the effect of four variables (tonal hierarchies, sensory chordal consonance, horizontal motion, and musical training) on perceived musical tension. Participants were asked to evaluate the tension created by a chord X in sequences of three chords [C major-->X-->C major] in a C major context key. The X chords could be major or minor triads major-minor seventh, or minor seventh chords built on the 12 notes of the chromatic scale. The data were compared with Krumhansl's (1990) harmonic hierarchy and with predictions of Lerdahl's (1988) cognitive theory, Hutchinson and Knopoff's (1978) and Parncutt's (1989) sensory-psychoacoustical theories, and the model of horizontal motion defined in the paper. As a main outcome, it appears that judgments of tension arose from a convergence of several cognitive and psychoacoustics influences, whose relative importance varies, depending on musical training.</p>","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"58 1","pages":"124-41"},"PeriodicalIF":0.0,"publicationDate":"1996-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"19645882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Locus equations as phonetic descriptors of consonantal place of articulation.","authors":"H. Sussman, J. Shore","doi":"10.1121/1.414302","DOIUrl":"https://doi.org/10.1121/1.414302","url":null,"abstract":"This experiment tested whether locus equation coefficients, slope, and y-intercept could serve as indices of place of articulation for obstruents sharing the same place of articulation across different manner classes. Locus equations for 22 speakers were derived from CV/t/ words with initial voiced stop /d/, voiceless aspirated stop /t/, nasal /n/, voiced fricative /z/, and voiceless fricative /s/ preceding 10 vowel contexts. Post hoc tests revealed /d/ = /z/ = /n/ for slope means. Voiced /d/ and voiceless /t/ were also equivalent when F2 transition onset measurement points were equated. Scatterplots of locus equation coefficients revealed three nonoverlapping and distinct clusters when the diverse coronal group was compared with labials and velars. A discriminant analysis using slope and y-intercept as predictors successfully categorized all five coronals into one alveolar group with 87.1% accuracy. The collective results support the contention that locus equations can serve as effective phonetic descriptors of consonant place of articulation across manner classes.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"68 1","pages":"936-46"},"PeriodicalIF":0.0,"publicationDate":"1995-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90795359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The cocktail party effect in infants.","authors":"P. Jusczyk, Rochelle S. Newman","doi":"10.1121/1.412966","DOIUrl":"https://doi.org/10.1121/1.412966","url":null,"abstract":"Most speech research with infants occurs in quiet laboratory rooms with no outside distractions. However, in the real world, speech directed to infants often occurs in the presence of other competing acoustic signals. To learn language, infants need to attend to their caregiver's speech even under less than ideal listening conditions. We examined 7.5-month-old infants' abilities to selectively attend to a female talker's voice when a male voice was talking simultaneously. In three experiments, infants heard a target voice repeating isolated words while a distractor voice spoke fluently at one of three different intensities. Subsequently, infants heard passages produced by the target voice containing either the familiar words or novel words. Infants listened longer to the familiar words when the target voice was 10 dB or 5 dB more intense than the distractor, but not when the two voices were equally intense. In a fourth experiment, the assignment of words and passages to the familiarization and testing phases was reversed so that the passages and distractors were presented simultaneously during familiarization, and the infants were tested on the familiar and unfamiliar isolated words. During familiarization, the passages were 10 dB more intense than the distractors. The results suggest that this may be at the limits of what infants at this age can do in separating two different streams of speech. In conclusion, infants have some capacity to extract information from speech even in the face of a competing acoustic voice.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"15 1","pages":"1145-56"},"PeriodicalIF":0.0,"publicationDate":"1995-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76254859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The McGurk effect in infants.","authors":"J. Johnson, L. Rosenblum, M. Schmuckler","doi":"10.1121/1.411551","DOIUrl":"https://doi.org/10.1121/1.411551","url":null,"abstract":"In the McGurk effect, perceptual identification of auditory speech syllables is influenced by simultaneous presentation of discrepant visible speech syllables. This effect has been found in subjects of different ages and with various native language backgrounds. But no McGurk tests have been conducted with prelinguistic infants. In the present series of experiments, 5-month-old English-exposed infants were tested for the McGurk effect. Infants were first gaze-habituated to an audiovisual /va/. Two different dishabituation stimuli were then presented: audio /ba/-visual /va/ (perceived by adults as /va/), and audio /da/-visual /va/ (perceived by adults as /da/). The infants showed generalization from the audiovisual /va/ to the audio /ba/-visual /va/ stimulus but not to the audio /da/-visual /va/ stimulus. Follow-up experiments revealed that these generalization differences were not due to a general preference for the audio /da/-visual /va/ stimulus or to the auditory similarity of /ba/ to /va/ relative to /da/. These results suggest that the infants were visually influenced in the same way as English-speaking adults are visually influenced.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"113 1","pages":"347-57"},"PeriodicalIF":0.0,"publicationDate":"1995-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81096944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Spectral discontinuities and the vowel length effect","authors":"A. Lotto, K. Kluender, K. Green","doi":"10.1121/1.411163","DOIUrl":"https://doi.org/10.1121/1.411163","url":null,"abstract":"Perception of voicing for stop consonants in consonant-vowel syllables can be affected by the duration of the following vowel so that longer vowels lead to more “voiced” responses. On the basis of several experiments, Green, Stevens, and Kuhl (1994) concluded that continuity of fundamental frequency (f0), but not continuity of formant structure, determined the effective length of the following vowel. In an extension of those efforts, we found here that both effects were critically dependent on particular f 0s and formant values. First, discontinuity inf0 does not necessarily preclude the vowel length effect because the effect maintains whenf0 changes from 200 to 100 Hz, and 200-Hz partials extend continuously through test syllables. Second, spectral discontinuity does preclude the vowel length effect when formant changes result in a spectral peak shifting to another harmonic. The results indicate that the effectiveness of stimulus changes for sustaining or diminishing the vowel length effect depends critically on particulars of spectral composition.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"46 1","pages":"1005-1014"},"PeriodicalIF":0.0,"publicationDate":"1994-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90439172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of a dynamical model of speech perception.","authors":"Pamela Case, B. Tuller, M. Ding, J. Kelso","doi":"10.1121/1.408981","DOIUrl":"https://doi.org/10.1121/1.408981","url":null,"abstract":"Previous work (Tuller, Case, Ding, & Kelso, 1994) has revealed signature properties of nonlinear dynamical systems in how people categorize speech sounds. The data were modeled by using a two-well potential function that deformed with stimulus properties and was sensitive to context. Here we evaluate one prediction of the model--namely, that the rate of change of the potential's slope should increase when the category is repeatedly perceived. Judged goodness of category membership was used as an index of the slope of the potential. Stimuli from a \"say\"-\"stay\" continuum were presented with gap duration changing sequentially throughout the range from 0 to 76 to 0 msec, or from 76 to 0 to 76 msec. Subjects identified each token as either \"say\" or \"stay\" and rated how good an exemplar it was of the identified category. As predicted, the same physical stimulus presented at the end of a sequence was judged a better exemplar of the category than was the identical stimulus presented at the beginning of the sequence. In contrast, stimuli presented twice near the middle of a sequence with few (or no) stimuli between them, as well as stimuli presented with an intervening random set, showed no such differences. These results confirm the hypothesis of a context-sensitive dynamical representation underlying speech.","PeriodicalId":19838,"journal":{"name":"Perception & Psychophysics","volume":"12 1","pages":"977-88"},"PeriodicalIF":0.0,"publicationDate":"1994-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78045760","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}