Abstracts 2005 - The Psychonomic Society
Abstracts 2005 - The Psychonomic Society
Abstracts 2005 - The Psychonomic Society
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Saturday Noon Posters 4008–4015<br />
sponse latencies in Experiments 3 and 4 showed entrainment to stimulus<br />
rate, and repetition priming was modulated by tonal expectedness.<br />
We discuss cognitive factors that can affect auditory repetition<br />
priming.<br />
(4008)<br />
Is Categorical Perception of Musical Intervals a Short-Term<br />
Memory Phenomenon? SINI E. MAURY, University of Helsinki, &<br />
ELISABET M. SERVICE, University of Helsinki and Dalhousie University<br />
(sponsored by Elisabet M. Service)—This study explores whether<br />
categorical perception of musical intervals can vary as a function of<br />
immediate memory load caused by interference from other sounds in<br />
a sequence. In a two-interval same–different discrimination task, musicians<br />
heard melodic intervals in isolation or embedded in four-note<br />
sequences. Half of the interval pairs were similar, and half were derived<br />
either from the same or from a different interval category. <strong>The</strong><br />
results showed that discriminability measured by d′ was significantly<br />
higher for intervals straddling the category boundary. This effect was<br />
even more pronounced when intervals formed a part of a melodic sequence.<br />
This could mean that categorical perception is a short-term<br />
memory phenomenon in which degraded auditory traces are repaired<br />
with top-down categorical information. <strong>The</strong> results also imply that the<br />
categorical information retrieved in the repair process takes the form<br />
of the prototype of the category and is not general knowledge about<br />
category membership.<br />
(4009)<br />
Musicians, Intermediate Musicians, and Nonmusicians’ Perception<br />
of Bitonality. MAYUMI HAMAMOTO, MARGARET P. MUNGER,<br />
& KYOTA KO, Davidson College—Bitonal music is characterized by<br />
a dissonant “crunch” sound that had been believed to be clearly audible<br />
by everyone (Wolpert, 2000). However, Wolpert found that nonmusicians<br />
did not identify bitonality in a free response task. <strong>The</strong> present<br />
study replicated Wolpert’s findings but also had participants rate<br />
song clips for preference, correctness and pleasantness. Monotonal<br />
music was rated higher on all dimensions, independently of the individual’s<br />
level of musical training. In addition, following a brief training<br />
session, nonmusicians (less than 1 year of musical training) identified<br />
the tonality of the final clips at equivalently high rates as the intermediate<br />
(mean, 2.4 years) and expert (mean, 9.2 years) musician groups.<br />
(4010)<br />
Cross-Modal Perception of Contour: <strong>The</strong> Role of Surface Correlation<br />
and Fourier Analysis Similarity. JON B. PRINCE & MARK A.<br />
SCHMUCKLER, University of Toronto, Scarborough—<strong>The</strong> perceived<br />
similarity of cross-modally presented contours was investigated with<br />
two experiments. <strong>The</strong> combination of surface correlation and Fourier<br />
analysis techniques allows quantitative descriptions of both global and<br />
local contour information. Experiment 1 investigated auditory–visual<br />
similarity by presenting a tonal melody followed by a line drawing and<br />
asking participants to rate the similarity between the two. Both stimuli<br />
were coded as integer series representing pitch or vertical height,<br />
respectively. Ratings were predicted by the surface correlation between<br />
the melody and the drawing (the correlation of the two integer<br />
series). Experiment 2 reversed the order of presentation by presenting<br />
the drawing first, followed by the melody. Surface correlation again<br />
predicted similarity ratings, in addition to amplitude and phase components<br />
derived from a Fourier analysis model. <strong>The</strong>se results validate<br />
the Fourier analysis model of contour cross-modally, particularly<br />
when participants must attend to the global character of visual and auditory<br />
contours.<br />
(4011)<br />
Effect of Encoding Processes on Remembering Melodies. ESRA<br />
MUNGAN & ZEHRA F. PEYNIRCIO ˇGLU, American University—<br />
In this study, both musicians and nonmusicians were asked to study a<br />
list of highly familiar melodies, using four different orienting tasks.<br />
Two were conceptually driven (continuing the melody and judging the<br />
107<br />
mood conveyed by the melody), and two were data driven (counting<br />
the number of long notes and tracing the melodic shape). <strong>The</strong> study<br />
phase was followed by an incidental free-choice recognition test.<br />
Findings showed that for nonmusicians, conceptually driven orienting<br />
tasks led to better memory performance than did data-driven orienting<br />
tasks, whereas for musicians the reverse was true. <strong>The</strong>se findings<br />
are discussed within the transfer-appropriate-processing framework.<br />
(4012)<br />
<strong>The</strong> Relationship Between Emotions Expressed and Elicited by<br />
Music and the Effect of Familiarity. OMAR ALI & ZEHRA F.<br />
PEYNIRCIO ˇGLU, American University (sponsored by Zehra F. Peynircio<br />
ˇglu)—We examined the effects of melodies on participants’ ratings<br />
of emotionality. <strong>The</strong> intensity of the ratings was higher when participants<br />
were asked to judge the emotion that was expressed by a melody<br />
(i.e., how happy/sad/calm/angry is this music?) than when they were<br />
asked to judge the emotion elicited by the same melody (i.e., how<br />
happy/sad/calm/angry does this music make you feel?). This pattern<br />
held across all four of the emotions and also even when the melodies<br />
were made familiar through repetition. In addition, positive emotions<br />
(i.e., happy and calm) were rated higher than negative emotions (i.e.,<br />
sad and angry). Finally, for both types of ratings (i.e., conveying or<br />
eliciting the emotion), the ratings in response to the repeated melodies<br />
were higher, but only for the sad and calm emotions.<br />
• EVENT COGNITION •<br />
(4013)<br />
Time Estimation and Fluency in Event Perception. MACKENZIE<br />
GLAHOLT, AVA ELAHIPANAH, ANTHONY R. MCINTOSH, &<br />
EYAL M. REINGOLD, University of Toronto, Mississauga—Intervals<br />
in which familiar stimuli (e.g., words) are presented are judged as<br />
longer than equal duration intervals in which unfamiliar stimuli are<br />
presented (e.g., nonwords). This perceptual illusion may result from<br />
the misattribution of the enhanced perceptual fluency associated with<br />
processing familiar stimuli. We investigated whether a similar phenomenon<br />
occurs in the perception of events. To manipulate event familiarity,<br />
we used 2-sec video clips of collisions between hockey players,<br />
played forward or in reverse. Reversed clips were closely matched<br />
to forward clips in terms of low-level perceptual characteristics, but<br />
they depicted events that violated physical laws and, as such, were unfamiliar.<br />
Participants judged reverse clips as having shorter duration<br />
and faster motion, as compared with forward clips. <strong>The</strong>se findings<br />
replicate and extend the findings with linguistic stimuli.<br />
(4014)<br />
From Seeing to Remembering Events in Time. SHULAN LU, Texas<br />
A&M University, Commerce—Everyday events have beginnings, ends,<br />
and intervals. <strong>The</strong>se temporal parameters have different combinations,<br />
and events have dynamic temporal trajectories. Previous research tends<br />
to assume that events follow one another and that subevents occur<br />
sometime in between. Recently, studies have begun to suggest that<br />
people may make finer grained temporal links than we previously<br />
thought. What kind of temporal properties get preserved more robustly?<br />
Participants viewed animations of fish-swimming events, where test<br />
events were embedded in a schema. For example, a group of fish chased<br />
away other fish. In Experiment 1, participants made judgments about the<br />
temporal relation of two given events immediately after they had viewed<br />
each animation. In Experiment 2, participants made judgments after<br />
viewing each animation and then drawing a maze for 25 sec. <strong>The</strong> results<br />
showed that people did not remember the time interval that occurred between<br />
two events but robustly preserved the overlap between events.<br />
(4015)<br />
Event Recognition in Free View and at an Eyeblink. REINHILD<br />
GLANEMANN, CHRISTIAN DOBEL, & PIENIE ZWITSERLOOD,<br />
Westfälische Wilhelms-Universität Münster—Recent studies demonstrated<br />
that brief visual presentation (around 20 msec) of photoreal-