Abstracts 2005 - The Psychonomic Society
Abstracts 2005 - The Psychonomic Society
Abstracts 2005 - The Psychonomic Society
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Posters 4016–4022 Saturday Noon<br />
istic stimuli allowed for semantic categorization. We examined the uptake<br />
of visual information from highly complex scenes with two actors<br />
involved in a meaningful action (agent position was balanced).<br />
Participants’ task was to indicate the patient’s position by buttonpress.<br />
Eyetracking revealed the tendency to fixate agents and action regions<br />
first. In most cases, both actors were inspected before an overt response<br />
was made. However, brief presentation (150 msec) of stimuli<br />
also produced highly accurate answers (about 95% correct). Under<br />
this condition, correct naming of actions was achieved only where<br />
body posture allowed few alternative actions. Apparently, visual event<br />
recognition is characterized by a rapid analysis that allows for role<br />
identification and a subsequent period of attention shifts that are necessary<br />
for more complex processes, such as verb retrieval.<br />
(4016)<br />
Event Clustering and Event Chaining: Associations in Autobiographical<br />
Memory. JENNIFER M. TALARICO, Duke University<br />
(sponsored by David C. Rubin)—Previous research has identified<br />
“event-clusters” as groups of autobiographical memories that cue one<br />
another, that share narrative elements, and that are thought to inform<br />
our understanding of autobiographical memory organization. However,<br />
the way in which these memories are recalled (event cuing) may<br />
determine a specific retrieval strategy, and it may be this strategic retrieval<br />
that produces the similarities between memories, not overall<br />
memory organization. A similar paradigm, “memory chaining,” has<br />
been developed that relies on associatively cued memories that is believed<br />
to be a more effective procedure for revealing the structure of<br />
autobiographical memory.<br />
(4017)<br />
Do Observers Encode Goal–Subgoal Structure Through Embodied<br />
Simulation? BRIDGETTE MARTIN, SANDRA C. LOZANO, &<br />
BARBARA TVERSKY, Stanford University—Hierarchical encoding<br />
of activities—that is, segmenting them into discrete actions organized<br />
as goals and subgoals—enhances learning (Martin, Lozano, & Tversky,<br />
<strong>2005</strong>). Surprisingly, segmenting and describing actions from the<br />
actor’s perspective promotes hierarchical encoding and learning more<br />
than does segmenting and describing them from one’s own perspective<br />
(Lozano, Martin, & Tversky, <strong>2005</strong>). Are enhanced hierarchical<br />
encoding and learning achieved through embodied simulation of observed<br />
actions? To test this, participants segmented an assembly task<br />
that they later performed, without warning. Participants segmented<br />
and described the activity from an actor- or self-perspective while attending<br />
to (1) movements of objects to different table sides, (2) actions<br />
completed on different table sides, or (3) actions completed by different<br />
hands. In accord with embodied simulation, hierarchical encoding<br />
and learning were best for actor-perspective describers who attended<br />
to action by hand, then to action by table side and object<br />
movement. For self-perspective participants, the order was reversed.<br />
(4018)<br />
<strong>The</strong> Effects of Elaboration on Confidence Inflation for Counterfactual<br />
Events. DIANNE M. LEARNED & MARIA S. ZARAGOZA,<br />
Kent State University (sponsored by F. Robert Treichler)—Imagination<br />
Inflation is the finding that imagining counterfactual autobiographical<br />
events from childhood increases participants’ confidence that the<br />
events transpired. Participants who imagine events are usually asked<br />
to elaborate on different aspects of the false events. However, it has<br />
been shown that participants are more likely to misattribute false<br />
events to a witnessed crime if they are asked to elaborate on the false<br />
events (Drivdahl & Zaragoza, 2001). <strong>The</strong> present study attempts to explore<br />
the effects of elaboration on confidence inflation for imagined<br />
counterfactual events. Three groups of participants were asked to<br />
imagine counterfactual events. <strong>The</strong> number of elaboration questions<br />
asked for each event varied between groups: six, three, or zero. <strong>The</strong><br />
results of this study indicate that participants are more likely to increase<br />
in confidence and to endorse memories for counterfactual life<br />
events the more they are asked to elaborate on the events.<br />
108<br />
• IMAGE PROCESSING •<br />
(4019)<br />
Using Visual Masking to Explore the Nature of Scene Gist.<br />
LESTER C. LOSCHKY, Kansas State University, AMIT SETHI &<br />
DANIEL J. SIMONS, University of Illinois, Urbana-Champaign, &<br />
DANIEL OCHS, JEREMY CORBIELLE, & KATIE GIBB, Kansas<br />
State University—We used visual masking to explore scene gist acquisition<br />
and to determine whether the meaning of a mask or its spatial<br />
frequency amplitude spectrum contributes more to visual masking of<br />
scene photographs. We systematically varied the recognizability of<br />
scenes, as both targets and masks, by parametrically randomizing their<br />
image phase, while maintaining their spatial frequency amplitude<br />
spectra. Increasing phase randomization reduced scene recognition<br />
and induced less masking of other scenes. Several types of meaningless<br />
masks that varied in spatial frequency amplitude spectra were<br />
equally ineffective—all were substantially less effective masks than<br />
were recognizable scenes. Together, the results of four experiments<br />
suggest that scene meaning plays a greater role in masking visual<br />
scenes than does the match in the amount of energy at various spatial<br />
frequencies. This further supports the psychological construct of conceptual<br />
masking and calls into question the value of spatial frequency<br />
amplitude information for acquiring scene gist.<br />
(4020)<br />
Correlating Adjacent Local Texture Elements to Recognize Natural<br />
Scenes. HIROYUKI WASHINO, Kyoto University, & JUN SAIKI,<br />
Kyoto University & JST—Our visual system can rapidly recognize complex<br />
natural scenes in the visual world. An important issue is how much<br />
bottom-up processing alone can account for scene categorization. Some<br />
previous works proposed that statistics of spatial frequencies or local texture<br />
elements are useful for rapid scene categorization (e.g., natural/<br />
artificial or forest/other). In this study, we proposed a computational<br />
model of basic-level scene categorization (e.g., forest, kitchen, etc.) that<br />
used the correlation of adjacent local texture elements. To examine this<br />
correlation’s efficiency in scene recognition, we investigated how people<br />
categorize a visual scene having only limited exposure time. Human<br />
accuracy data as a function of exposure time were well accounted for by<br />
the model using the number of sampled texture elements. <strong>The</strong> confusion<br />
matrices of humans and the model did not significantly differ. <strong>The</strong>se results<br />
suggest that the correlation of local texture statistics can account<br />
for our scene categorization performance.<br />
(4021)<br />
Effects of Zoom and Field of View on Relative Distance Judgments.<br />
JOCELYN M. KEILLOR, PAYAL AGARWAL, & JENNIFER<br />
JEON, Defence R&D Canada, & MICHAEL E. PERLIN, CMC Electronics—When<br />
a nonorthogonal sensor image is projected onto a rectangular<br />
display, the observer is not typically shown the viewing frustum<br />
and may, therefore, be unable to compensate for the ensuing<br />
distortion. <strong>The</strong> present experiment was designed to evaluate whether<br />
this type of distortion affects relative distance judgments. In a synthetic<br />
environment, three different fields of view were selected by adjusting<br />
zoom such that the size of a central object remained the same<br />
in all three conditions but the surrounding objects were subject to differential<br />
foreshortening and size distortion as a function of condition.<br />
In addition, a moving sensor was simulated to determine whether motion<br />
parallax allowed viewers to overcome any effects of the different<br />
fields of view and zoom settings. Despite the presence of multiple<br />
cues to depth in the scene, relative depth judgments were affected by<br />
the distortion introduced by the projection, and the addition of motion<br />
parallax did not reduce this effect.<br />
(4022)<br />
“Lightning Bolts” Reveal Categorical Coding of Relative Orientation.<br />
ERIC E. COOPER, GLENN E. CASNER, & ALEXANDER M.<br />
O’BRIEN, Iowa State University—<strong>The</strong> research tested whether orientation<br />
is categorically coded during object perception for T-vertex,