S1 (FriAM 1-65) - The Psychonomic Society
S1 (FriAM 1-65) - The Psychonomic Society
S1 (FriAM 1-65) - The Psychonomic Society
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Papers 223–229 Saturday Afternoon<br />
2:50–3:05 (223)<br />
Testing the Whorfian Hypothesis: Lateralized Presentation and<br />
Color Recognition. CLARK G. OHNESORGE, Carleton College,<br />
GEORGE KACHERGIS, Indiana University, & AARON FANTA,<br />
Epic Systems—In recent years, several papers report effects consistent<br />
with the Whorfian hypothesis, the claim that one’s language exerts<br />
an influence on perceptual and cognitive tasks. Generally it is unclear<br />
whether the asserted influence of language is on early<br />
(perceptual) processes or later (decisional) processes. Using a lateralized<br />
color detection task we replicate the findings of a representative<br />
paper, (Gilbert, Regier, Kay, & Ivry, PNAS, 2006) and extend<br />
them through a modification of the basic task in a design that uses response<br />
time as the dependent variable. Our findings support their conclusion<br />
that language seems to exert an influence in this task. We also<br />
present data from a study employing a 2AFC task allowing a strong<br />
test of the claim that linguistic influence exerts itself at the perceptual<br />
level. In this task we find no evidence to support the claim of an influence<br />
of language at the perceptual level.<br />
Multisensory Integration and Auditory Perception<br />
Shoreline, Saturday Afternoon, 1:30–3:10<br />
Chaired by Adele A. Diederich, Jacobs University<br />
1:30–1:45 (224)<br />
Too Old to Integrate? A Cross-Modal Study With the Elderly.<br />
ADELE A. DIEDERICH, Jacobs University, & HANS COLONIUS,<br />
Oldenburg University—Saccadic reaction time (SRT) to visual targets<br />
tends to be faster when auditory stimuli occur in close temporal or<br />
spatial proximity even when subjects are instructed to ignore the nontarget<br />
(focused attention paradigm). Here it is investigated whether<br />
this facilitation is modulated by the subjects’ age. One group consisted<br />
of college students (age: 20–23 years), the other of elderly (age:<br />
<strong>65</strong>–75 years). Auditory nontargets (white noise) were presented 100,<br />
50, or 0 msec before or 50 msec after the visual target (LED). Mean<br />
SRT of the elderly was about twice as long as for the college students.<br />
Multisensory response enhancement (MRE), a measure of relative<br />
speed-up of responses in the bimodal versus the visual-alone condition,<br />
was larger for the elderly. <strong>The</strong> time-window-of-integration<br />
(TWIN) model for multisensory integration in SRT, developed by the<br />
authors, is utilized to disentangle the effects of peripheral sensory processing<br />
from multisensory integration efficiency.<br />
1:50–2:05 (225)<br />
Neural Correlates of Multisensory Integration of Ecologically Valid<br />
Audiovisual Events. JEAN VROOMEN & JEROEN STEKELEN-<br />
BURG, Tilburg University—A question that has emerged over recent<br />
years is whether audiovisual (AV) speech perception is a special case<br />
of multisensory perception. Electrophysiological (ERP) studies have<br />
found that auditory neural activity (N1 component of the ERP) induced<br />
by speech is suppressed and speeded up when a speech sound is accompanied<br />
by concordant lip movements. In Experiment 1, we show<br />
that this AV interaction is not speech-specific. Ecologically valid nonspeech<br />
AV events (actions performed by an actor such as handclapping)<br />
were associated with a similar speeding up and suppression of auditory<br />
N1 amplitude as AV speech (syllables). Experiment 2 demonstrated that<br />
these AV interactions were not influenced by whether A and V were<br />
congruent or incongruent. In Experiment 3 we show that the AV interaction<br />
on N1 was absent when there was no anticipatory visual motion,<br />
indicating that the AV interaction only occurred when visual anticipatory<br />
motion preceded the sound. <strong>The</strong>se results demonstrate that the visually<br />
induced speeding-up and suppression of auditory N1 amplitude<br />
reflect multisensory integrative mechanisms of AV events that crucially<br />
depend on whether vision predicts when the sound occurs.<br />
2:10–2:25 (226)<br />
Apparent Causality and Audio-Visual Synchrony Perception.<br />
ROB L. J. VAN EIJK, Eindhoven University of Technology, ARMIN<br />
35<br />
KOHLRAUSCH, Philips Research Laboratories and Eindhoven University<br />
of Technology, JAMES F. JUOLA, University of Kansas and<br />
Eindhoven University of Technology, & STEVEN VAN DE PAR, Philips<br />
Research Laboratories (read by James F. Juola)—In audio-visual synchrony<br />
perception, the proportion of “synchronous” responses is not<br />
symmetric around physical synchrony, but centers around an audio<br />
delay called the point of subjective simultaneity (PSS). We investigated<br />
whether an observer's causal interpretation influences the PSS.<br />
We used an animation of Newton’s Cradle, showing a left-to-right pendulum<br />
movement, in three conditions: (1) the entire stimulus, (2) the<br />
left half, or (3) the right half. In Conditions 1 and 2, the visual movement<br />
appeared to cause a collision sound, whereas in Condition 3 the<br />
sound appeared to cause the visual movement. Conditions 1 and 2<br />
yielded typical PSSs, whereas for Condition 3, perceptual synchrony<br />
coincided with physical synchrony. Opposite results were found using<br />
a temporal-order judgment (TOJ) paradigm. Such response shifts indicate<br />
that PSSs found in synchrony judgment experiments are influenced<br />
by apparent causal relations between auditory and visual stimuli,<br />
whereas TOJs are driven by strategies that are at best indirectly<br />
related to perceived causality.<br />
2:30–2:45 (227)<br />
Are Valid Auditory Cues Helpful? ADAM J. REEVES & BERTRAM<br />
SCHARF, Northeastern University—We asked whether a valid cue,<br />
one that indicates the frequency of an upcoming auditory signal, helps<br />
detection. Forty millisecond signals and cues were presented in continuous<br />
50- or 60-dB noise. Signals were near threshold, and cues 4<br />
to 8 dB above. A valid frequency cue presented just before a signal of<br />
uncertain frequency reduces the well-known deleterious effect of frequency<br />
uncertainty. However, when signal frequency is certain, the<br />
cue interferes with detection over the next 100 msec or so by up to 2<br />
to 3 dB. This “forward interference” effect is about the same whether<br />
the cue is ipsi- or contralateral to the signal, so is not forward masking,<br />
and it is less for invalid cues. Attending to the cue apparently not<br />
only aids focusing on the critical band containing the upcoming signal,<br />
but also interferes by absorbing processing resources that otherwise<br />
would go to the signal.<br />
2:50–3:05 (228)<br />
Hypervigilance Does Not Increase the Loudness of Unpleasant<br />
Sounds. MARK HOLLINS, DANIEL E. HARPER, SHANNON M.<br />
GALLAGHER, & WILLIAM MAIXNER, University of North Carolina,<br />
Chapel Hill—Hypervigilance, a feature of some chronic pain<br />
conditions such as fibromyalgia and temporomandibular joint disorder<br />
(TMJD), is a heightened attentiveness to, and perhaps a perceptual<br />
amplification of, painful sensations. McDermid et al. (1996) proposed,<br />
mainly on the basis of noise tolerance data, that this<br />
amplification may extend to aversive stimuli that are not painful. To<br />
test directly the idea that hypervigilance involves an increase in perceived<br />
intensity of painless but aversive stimuli, we asked individuals<br />
with TMJD (n = 10), and healthy control subjects (n = 11), to estimate<br />
the loudness of auditory stimuli (two-tone combinations) ranging<br />
from soft to unpleasantly loud (35–90 dB SPL). Although the TMJD<br />
subjects were significantly more hypervigilant than the controls based<br />
on questionnaire (PILL) responses, the loudness estimates of the two<br />
groups were equivalent across the intensity range. <strong>The</strong> results indicate<br />
that hypervigilance need not be accompanied by an increase in the<br />
subjective intensity of all aversive stimuli.<br />
Face Processing<br />
Regency ABC, Saturday Afternoon, 4:10–5:30<br />
Chaired by K. Suzanne Scherf, University of Pittsburgh<br />
4:10–4:25 (229)<br />
Developmental Trajectory of Visual Object Recognition Revealed by<br />
fMRI. K. SUZANNE SCHERF, University of Pittsburgh, MARLENE<br />
BEHRMANN, Carnegie Mellon University, KATE HUMPHREYS,