05.01.2013 Views

Perceptual Coherence : Hearing and Seeing

Perceptual Coherence : Hearing and Seeing

Perceptual Coherence : Hearing and Seeing

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

filter inferred from the background change (this is similar to the strategy<br />

proposed in chapter 7 for asymmetric color matching).<br />

Schulz <strong>and</strong> Sanocki (2003) have demonstrated that proximal (retinal)<br />

<strong>and</strong> constancy grouping coexist but have different time courses. In fact,<br />

Rock (1997) proposed that initially there is lower-level literal perception<br />

corresponding to the retinal stimulation. If that is inadequate for some reason,<br />

the literal perception is replaced by what he termed a higher-order<br />

world or constancy mode. Usually, the transition from the literal to the<br />

world mode is irresistible, although the literal mode may bias the worldmode<br />

judgment <strong>and</strong> although it is still possible to make literal-mode judgments.<br />

To make this concrete, although observers grouped by reflectance in<br />

the experiment described above (Rock et al., 1992), it is very likely that<br />

they would be able to group the columns by luminance if asked.<br />

In the later experiment (Schulz & Sanocki, 2003), the exposure time of<br />

a grid similar to that used by Rock et al. (1992) was varied from 200 to<br />

2,000 ms. At the shortest exposure time, nearly all choices were based on<br />

luminance matching (88%). In contrast, at the longest exposure time nearly<br />

all matches were based on reflectance (83%). Schulz <strong>and</strong> Sanocki proposed<br />

that retinal matching is performed in the earliest cortical regions (V1 or<br />

V2) <strong>and</strong> that reflectance matching is carried out in V4.<br />

These sorts of results parallel those discussed in chapter 1 with respect<br />

to the perceived motion of spatial gratings through an aperture. Recall that<br />

the initial percept is roughly toward the average of the motion of the two<br />

gratings (i.e., the additive vector direction), what we would call the proximal<br />

percept. This percept is replaced by a perception that is based on the<br />

intersection of the lines of constraint, what we would call constancy or<br />

world mode. The vector average still biased the intersection of constraint<br />

judgments toward the additive vector direction.<br />

Audition<br />

Auditory <strong>and</strong> Visual Segmentation 403<br />

The issue of when grouping occurs has not been explicitly discussed with<br />

respect to stream segregation. As described above, the default percept is<br />

one of coherence <strong>and</strong> that it takes several repetitions before streams arise if<br />

several sounds are recycled continuously (Bregman, 1990). Thus, there is a<br />

shift from proximal to distal perceiving.<br />

It is also likely that the overall context determines segregation. We can<br />

imagine the following experiment. We start with a stationary tone with a<br />

spoken-vowel spectrum of harmonic components <strong>and</strong> add energy at one<br />

frequency. (A natural vowel would not work because it changes its pitch<br />

<strong>and</strong> spectral content over time.) If the resulting sound were presented for<br />

a short duration, we would then expect the listener to report a funnysounding<br />

sound. This outcome would be analogous to retinal or proximal

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!