29.01.2013 Views

S1 (FriAM 1-65) - The Psychonomic Society

S1 (FriAM 1-65) - The Psychonomic Society

S1 (FriAM 1-65) - The Psychonomic Society

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Papers 223–229 Saturday Afternoon<br />

2:50–3:05 (223)<br />

Testing the Whorfian Hypothesis: Lateralized Presentation and<br />

Color Recognition. CLARK G. OHNESORGE, Carleton College,<br />

GEORGE KACHERGIS, Indiana University, & AARON FANTA,<br />

Epic Systems—In recent years, several papers report effects consistent<br />

with the Whorfian hypothesis, the claim that one’s language exerts<br />

an influence on perceptual and cognitive tasks. Generally it is unclear<br />

whether the asserted influence of language is on early<br />

(perceptual) processes or later (decisional) processes. Using a lateralized<br />

color detection task we replicate the findings of a representative<br />

paper, (Gilbert, Regier, Kay, & Ivry, PNAS, 2006) and extend<br />

them through a modification of the basic task in a design that uses response<br />

time as the dependent variable. Our findings support their conclusion<br />

that language seems to exert an influence in this task. We also<br />

present data from a study employing a 2AFC task allowing a strong<br />

test of the claim that linguistic influence exerts itself at the perceptual<br />

level. In this task we find no evidence to support the claim of an influence<br />

of language at the perceptual level.<br />

Multisensory Integration and Auditory Perception<br />

Shoreline, Saturday Afternoon, 1:30–3:10<br />

Chaired by Adele A. Diederich, Jacobs University<br />

1:30–1:45 (224)<br />

Too Old to Integrate? A Cross-Modal Study With the Elderly.<br />

ADELE A. DIEDERICH, Jacobs University, & HANS COLONIUS,<br />

Oldenburg University—Saccadic reaction time (SRT) to visual targets<br />

tends to be faster when auditory stimuli occur in close temporal or<br />

spatial proximity even when subjects are instructed to ignore the nontarget<br />

(focused attention paradigm). Here it is investigated whether<br />

this facilitation is modulated by the subjects’ age. One group consisted<br />

of college students (age: 20–23 years), the other of elderly (age:<br />

<strong>65</strong>–75 years). Auditory nontargets (white noise) were presented 100,<br />

50, or 0 msec before or 50 msec after the visual target (LED). Mean<br />

SRT of the elderly was about twice as long as for the college students.<br />

Multisensory response enhancement (MRE), a measure of relative<br />

speed-up of responses in the bimodal versus the visual-alone condition,<br />

was larger for the elderly. <strong>The</strong> time-window-of-integration<br />

(TWIN) model for multisensory integration in SRT, developed by the<br />

authors, is utilized to disentangle the effects of peripheral sensory processing<br />

from multisensory integration efficiency.<br />

1:50–2:05 (225)<br />

Neural Correlates of Multisensory Integration of Ecologically Valid<br />

Audiovisual Events. JEAN VROOMEN & JEROEN STEKELEN-<br />

BURG, Tilburg University—A question that has emerged over recent<br />

years is whether audiovisual (AV) speech perception is a special case<br />

of multisensory perception. Electrophysiological (ERP) studies have<br />

found that auditory neural activity (N1 component of the ERP) induced<br />

by speech is suppressed and speeded up when a speech sound is accompanied<br />

by concordant lip movements. In Experiment 1, we show<br />

that this AV interaction is not speech-specific. Ecologically valid nonspeech<br />

AV events (actions performed by an actor such as handclapping)<br />

were associated with a similar speeding up and suppression of auditory<br />

N1 amplitude as AV speech (syllables). Experiment 2 demonstrated that<br />

these AV interactions were not influenced by whether A and V were<br />

congruent or incongruent. In Experiment 3 we show that the AV interaction<br />

on N1 was absent when there was no anticipatory visual motion,<br />

indicating that the AV interaction only occurred when visual anticipatory<br />

motion preceded the sound. <strong>The</strong>se results demonstrate that the visually<br />

induced speeding-up and suppression of auditory N1 amplitude<br />

reflect multisensory integrative mechanisms of AV events that crucially<br />

depend on whether vision predicts when the sound occurs.<br />

2:10–2:25 (226)<br />

Apparent Causality and Audio-Visual Synchrony Perception.<br />

ROB L. J. VAN EIJK, Eindhoven University of Technology, ARMIN<br />

35<br />

KOHLRAUSCH, Philips Research Laboratories and Eindhoven University<br />

of Technology, JAMES F. JUOLA, University of Kansas and<br />

Eindhoven University of Technology, & STEVEN VAN DE PAR, Philips<br />

Research Laboratories (read by James F. Juola)—In audio-visual synchrony<br />

perception, the proportion of “synchronous” responses is not<br />

symmetric around physical synchrony, but centers around an audio<br />

delay called the point of subjective simultaneity (PSS). We investigated<br />

whether an observer's causal interpretation influences the PSS.<br />

We used an animation of Newton’s Cradle, showing a left-to-right pendulum<br />

movement, in three conditions: (1) the entire stimulus, (2) the<br />

left half, or (3) the right half. In Conditions 1 and 2, the visual movement<br />

appeared to cause a collision sound, whereas in Condition 3 the<br />

sound appeared to cause the visual movement. Conditions 1 and 2<br />

yielded typical PSSs, whereas for Condition 3, perceptual synchrony<br />

coincided with physical synchrony. Opposite results were found using<br />

a temporal-order judgment (TOJ) paradigm. Such response shifts indicate<br />

that PSSs found in synchrony judgment experiments are influenced<br />

by apparent causal relations between auditory and visual stimuli,<br />

whereas TOJs are driven by strategies that are at best indirectly<br />

related to perceived causality.<br />

2:30–2:45 (227)<br />

Are Valid Auditory Cues Helpful? ADAM J. REEVES & BERTRAM<br />

SCHARF, Northeastern University—We asked whether a valid cue,<br />

one that indicates the frequency of an upcoming auditory signal, helps<br />

detection. Forty millisecond signals and cues were presented in continuous<br />

50- or 60-dB noise. Signals were near threshold, and cues 4<br />

to 8 dB above. A valid frequency cue presented just before a signal of<br />

uncertain frequency reduces the well-known deleterious effect of frequency<br />

uncertainty. However, when signal frequency is certain, the<br />

cue interferes with detection over the next 100 msec or so by up to 2<br />

to 3 dB. This “forward interference” effect is about the same whether<br />

the cue is ipsi- or contralateral to the signal, so is not forward masking,<br />

and it is less for invalid cues. Attending to the cue apparently not<br />

only aids focusing on the critical band containing the upcoming signal,<br />

but also interferes by absorbing processing resources that otherwise<br />

would go to the signal.<br />

2:50–3:05 (228)<br />

Hypervigilance Does Not Increase the Loudness of Unpleasant<br />

Sounds. MARK HOLLINS, DANIEL E. HARPER, SHANNON M.<br />

GALLAGHER, & WILLIAM MAIXNER, University of North Carolina,<br />

Chapel Hill—Hypervigilance, a feature of some chronic pain<br />

conditions such as fibromyalgia and temporomandibular joint disorder<br />

(TMJD), is a heightened attentiveness to, and perhaps a perceptual<br />

amplification of, painful sensations. McDermid et al. (1996) proposed,<br />

mainly on the basis of noise tolerance data, that this<br />

amplification may extend to aversive stimuli that are not painful. To<br />

test directly the idea that hypervigilance involves an increase in perceived<br />

intensity of painless but aversive stimuli, we asked individuals<br />

with TMJD (n = 10), and healthy control subjects (n = 11), to estimate<br />

the loudness of auditory stimuli (two-tone combinations) ranging<br />

from soft to unpleasantly loud (35–90 dB SPL). Although the TMJD<br />

subjects were significantly more hypervigilant than the controls based<br />

on questionnaire (PILL) responses, the loudness estimates of the two<br />

groups were equivalent across the intensity range. <strong>The</strong> results indicate<br />

that hypervigilance need not be accompanied by an increase in the<br />

subjective intensity of all aversive stimuli.<br />

Face Processing<br />

Regency ABC, Saturday Afternoon, 4:10–5:30<br />

Chaired by K. Suzanne Scherf, University of Pittsburgh<br />

4:10–4:25 (229)<br />

Developmental Trajectory of Visual Object Recognition Revealed by<br />

fMRI. K. SUZANNE SCHERF, University of Pittsburgh, MARLENE<br />

BEHRMANN, Carnegie Mellon University, KATE HUMPHREYS,

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!