29.01.2013 Views

Abstracts 2005 - The Psychonomic Society

Abstracts 2005 - The Psychonomic Society

Abstracts 2005 - The Psychonomic Society

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Saturday Noon Posters 4008–4015<br />

sponse latencies in Experiments 3 and 4 showed entrainment to stimulus<br />

rate, and repetition priming was modulated by tonal expectedness.<br />

We discuss cognitive factors that can affect auditory repetition<br />

priming.<br />

(4008)<br />

Is Categorical Perception of Musical Intervals a Short-Term<br />

Memory Phenomenon? SINI E. MAURY, University of Helsinki, &<br />

ELISABET M. SERVICE, University of Helsinki and Dalhousie University<br />

(sponsored by Elisabet M. Service)—This study explores whether<br />

categorical perception of musical intervals can vary as a function of<br />

immediate memory load caused by interference from other sounds in<br />

a sequence. In a two-interval same–different discrimination task, musicians<br />

heard melodic intervals in isolation or embedded in four-note<br />

sequences. Half of the interval pairs were similar, and half were derived<br />

either from the same or from a different interval category. <strong>The</strong><br />

results showed that discriminability measured by d′ was significantly<br />

higher for intervals straddling the category boundary. This effect was<br />

even more pronounced when intervals formed a part of a melodic sequence.<br />

This could mean that categorical perception is a short-term<br />

memory phenomenon in which degraded auditory traces are repaired<br />

with top-down categorical information. <strong>The</strong> results also imply that the<br />

categorical information retrieved in the repair process takes the form<br />

of the prototype of the category and is not general knowledge about<br />

category membership.<br />

(4009)<br />

Musicians, Intermediate Musicians, and Nonmusicians’ Perception<br />

of Bitonality. MAYUMI HAMAMOTO, MARGARET P. MUNGER,<br />

& KYOTA KO, Davidson College—Bitonal music is characterized by<br />

a dissonant “crunch” sound that had been believed to be clearly audible<br />

by everyone (Wolpert, 2000). However, Wolpert found that nonmusicians<br />

did not identify bitonality in a free response task. <strong>The</strong> present<br />

study replicated Wolpert’s findings but also had participants rate<br />

song clips for preference, correctness and pleasantness. Monotonal<br />

music was rated higher on all dimensions, independently of the individual’s<br />

level of musical training. In addition, following a brief training<br />

session, nonmusicians (less than 1 year of musical training) identified<br />

the tonality of the final clips at equivalently high rates as the intermediate<br />

(mean, 2.4 years) and expert (mean, 9.2 years) musician groups.<br />

(4010)<br />

Cross-Modal Perception of Contour: <strong>The</strong> Role of Surface Correlation<br />

and Fourier Analysis Similarity. JON B. PRINCE & MARK A.<br />

SCHMUCKLER, University of Toronto, Scarborough—<strong>The</strong> perceived<br />

similarity of cross-modally presented contours was investigated with<br />

two experiments. <strong>The</strong> combination of surface correlation and Fourier<br />

analysis techniques allows quantitative descriptions of both global and<br />

local contour information. Experiment 1 investigated auditory–visual<br />

similarity by presenting a tonal melody followed by a line drawing and<br />

asking participants to rate the similarity between the two. Both stimuli<br />

were coded as integer series representing pitch or vertical height,<br />

respectively. Ratings were predicted by the surface correlation between<br />

the melody and the drawing (the correlation of the two integer<br />

series). Experiment 2 reversed the order of presentation by presenting<br />

the drawing first, followed by the melody. Surface correlation again<br />

predicted similarity ratings, in addition to amplitude and phase components<br />

derived from a Fourier analysis model. <strong>The</strong>se results validate<br />

the Fourier analysis model of contour cross-modally, particularly<br />

when participants must attend to the global character of visual and auditory<br />

contours.<br />

(4011)<br />

Effect of Encoding Processes on Remembering Melodies. ESRA<br />

MUNGAN & ZEHRA F. PEYNIRCIO ˇGLU, American University—<br />

In this study, both musicians and nonmusicians were asked to study a<br />

list of highly familiar melodies, using four different orienting tasks.<br />

Two were conceptually driven (continuing the melody and judging the<br />

107<br />

mood conveyed by the melody), and two were data driven (counting<br />

the number of long notes and tracing the melodic shape). <strong>The</strong> study<br />

phase was followed by an incidental free-choice recognition test.<br />

Findings showed that for nonmusicians, conceptually driven orienting<br />

tasks led to better memory performance than did data-driven orienting<br />

tasks, whereas for musicians the reverse was true. <strong>The</strong>se findings<br />

are discussed within the transfer-appropriate-processing framework.<br />

(4012)<br />

<strong>The</strong> Relationship Between Emotions Expressed and Elicited by<br />

Music and the Effect of Familiarity. OMAR ALI & ZEHRA F.<br />

PEYNIRCIO ˇGLU, American University (sponsored by Zehra F. Peynircio<br />

ˇglu)—We examined the effects of melodies on participants’ ratings<br />

of emotionality. <strong>The</strong> intensity of the ratings was higher when participants<br />

were asked to judge the emotion that was expressed by a melody<br />

(i.e., how happy/sad/calm/angry is this music?) than when they were<br />

asked to judge the emotion elicited by the same melody (i.e., how<br />

happy/sad/calm/angry does this music make you feel?). This pattern<br />

held across all four of the emotions and also even when the melodies<br />

were made familiar through repetition. In addition, positive emotions<br />

(i.e., happy and calm) were rated higher than negative emotions (i.e.,<br />

sad and angry). Finally, for both types of ratings (i.e., conveying or<br />

eliciting the emotion), the ratings in response to the repeated melodies<br />

were higher, but only for the sad and calm emotions.<br />

• EVENT COGNITION •<br />

(4013)<br />

Time Estimation and Fluency in Event Perception. MACKENZIE<br />

GLAHOLT, AVA ELAHIPANAH, ANTHONY R. MCINTOSH, &<br />

EYAL M. REINGOLD, University of Toronto, Mississauga—Intervals<br />

in which familiar stimuli (e.g., words) are presented are judged as<br />

longer than equal duration intervals in which unfamiliar stimuli are<br />

presented (e.g., nonwords). This perceptual illusion may result from<br />

the misattribution of the enhanced perceptual fluency associated with<br />

processing familiar stimuli. We investigated whether a similar phenomenon<br />

occurs in the perception of events. To manipulate event familiarity,<br />

we used 2-sec video clips of collisions between hockey players,<br />

played forward or in reverse. Reversed clips were closely matched<br />

to forward clips in terms of low-level perceptual characteristics, but<br />

they depicted events that violated physical laws and, as such, were unfamiliar.<br />

Participants judged reverse clips as having shorter duration<br />

and faster motion, as compared with forward clips. <strong>The</strong>se findings<br />

replicate and extend the findings with linguistic stimuli.<br />

(4014)<br />

From Seeing to Remembering Events in Time. SHULAN LU, Texas<br />

A&M University, Commerce—Everyday events have beginnings, ends,<br />

and intervals. <strong>The</strong>se temporal parameters have different combinations,<br />

and events have dynamic temporal trajectories. Previous research tends<br />

to assume that events follow one another and that subevents occur<br />

sometime in between. Recently, studies have begun to suggest that<br />

people may make finer grained temporal links than we previously<br />

thought. What kind of temporal properties get preserved more robustly?<br />

Participants viewed animations of fish-swimming events, where test<br />

events were embedded in a schema. For example, a group of fish chased<br />

away other fish. In Experiment 1, participants made judgments about the<br />

temporal relation of two given events immediately after they had viewed<br />

each animation. In Experiment 2, participants made judgments after<br />

viewing each animation and then drawing a maze for 25 sec. <strong>The</strong> results<br />

showed that people did not remember the time interval that occurred between<br />

two events but robustly preserved the overlap between events.<br />

(4015)<br />

Event Recognition in Free View and at an Eyeblink. REINHILD<br />

GLANEMANN, CHRISTIAN DOBEL, & PIENIE ZWITSERLOOD,<br />

Westfälische Wilhelms-Universität Münster—Recent studies demonstrated<br />

that brief visual presentation (around 20 msec) of photoreal-

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!