Welcome To Website IAS

Hot news
Achievement

Independence Award

- First Rank - Second Rank - Third Rank

Labour Award

- First Rank - Second Rank -Third Rank

National Award

 - Study on food stuff for animal(2005)

 - Study on rice breeding for export and domestic consumption(2005)

VIFOTEC Award

- Hybrid Maize by Single Cross V2002 (2003)

- Tomato Grafting to Manage Ralstonia Disease(2005)

- Cassava variety KM140(2010)

Centres
Website links
Vietnamese calendar
Library
Visitors summary
 Curently online :  7
 Total visitors :  7684370

Birds tune in to sequential information when categorizing their songs

Imagine a friend telling you how she recently bought a goat. Why would your friend possibly buy a goat, you wonder—or had she actually said she bought a coat? Usually, listeners spend little time deciding whether someone had talked about a goat or a coat, or a pet or a bet; instead, we instantly assign the phonemes that we are hearing to one or the other category, even though the main distinguishing feature between the phonemes /ba/ and /pa/, or /ga/ and /ka/, shows a continuous, albeit bimodal distribution.

Julia Fischer

Cognitive Ethology Laboratory, German Primate Center, 37077 Göttingen, Germany

 

Imagine a friend telling you how she recently bought a goat. Why would your friend possibly buy a goat, you wonder—or had she actually said she bought a coat? Usually, listeners spend little time deciding whether someone had talked about a goat or a coat, or a pet or a bet; instead, we instantly assign the phonemes that we are hearing to one or the other category, even though the main distinguishing feature between the phonemes /ba/ and /pa/, or /ga/ and /ka/, shows a continuous, albeit bimodal distribution. A key feature is the amount of time that passes between the plosive burst and the onset of voicing, the so-called “voice-onset-time” (VOT). For phonemes beginning with voiced stop consonants such as /ba/ and /da/, the voice sets in immediately or sometimes even shortly before the plosive sound, whereas for unvoiced stop consonants such as /pa/ and /ta/, between 20 and 100 ms may pass until voice onset. In experimental settings, English-speaking listeners assign all phonemes with a VOT up to about 35 ms to one category (e.g., /da/), whereas beyond this categorical boundary, the perception flips, and listeners assign the phoneme to the other category (in this case /ta/). This phenomenon is known as categorical perception (CP) (1) and is a crucial prerequisite for online processing and meaning attribution of the continuous speech stream. Categorical boundaries are not static, however, and depend on the linguistic community, but also the immediate linguistic information in which the phoneme in question is embedded: for instance, whether a phoneme occurs at the beginning, in the middle, or at the end of a word (2). In PNAS, Lachlan and Nowicki (3) use detailed acoustic analyses and playback experiments to show that swamp sparrows, Melospiza georgiana, categorize one specific note type depending on its position within a song syllable. With their study, they provide strong evidence that categorical perception in birds is influenced by the acoustic context in a similar fashion as in humans.

 

Initially, CP in the auditory domain was believed to be restricted to the perception of speech sounds by humans (4). Following a conservative definition, categorical perception could be diagnosed when the following criteria were fulfilled: (i) distinct labeling of stimulus categories (“this is a /da/”), (ii) failure to discriminate within categories, e.g., different tokens of /da/, (iii) high sensitivity to differences at the category boundary between /da/ and /ta/, for instance, and (iv) a close agreement between labeling and discrimination functions (5). More loosely, categorical perception can be conceived as a compression of within-category and/or a separation of between-category differences (6). Kuhl and Miller were the first to challenge the assumption that categorical perception in the auditory domain was restricted to human speech, by training chinchillas to discriminate between different speech tokens (7). The animals were rewarded for distinguishing between the end points of the voiced-voiceless continuum distinguishing /da/ and /ta/. In the test trials, the animals placed the phonetic boundary more or less in the middle between the two end points. However, because the animals were trained, it remained unclear whether the observed categorization was simply a result of the training and had therefore little to do with their natural categorization of sounds.

 

Lachlan and Nowicki built their experiments on a classic study by Nelson and Marler (8), who had investigated categorical perception in a natural learned communication system: the song of the swamp sparrow. Nelson and Marler studied the birds’ responses to variation in note duration, a feature characteristic for different populations of this species, applying the habituation-dishabituation paradigm initally used with human infants (9). With this technique, a series of stimuli is presented until the subject ceases to respond. Subsequently, a putatively distinct stimulus is broadcast. A recovery in response suggests that this stimulus is placed in a different category than those used for habituation, whereas a failure to respond to this test stimulus suggests that it is placed in the same category (10). In the original experiments, swamp sparrows from one population in New York only showed renewed responses when the note duration was switched to a length of the other category, whereas they failed to do so when the same absolute variation fell within a given category. Interestingly, in a different population of swamp sparrows in Pennsylvania, the perceptual boundary between two categories of notes differed from that found in the New York population, in accordance with the difference in the overall distribution of the different note lengths (11).

 

Categorical perception is not restricted to learned communication systems such as speech or bird song: by now, it has been found in such diverse taxa as crickets (12), frogs (13), and monkeys (14). Different populations of macaques differed with regard to the categorization of sounds (14), lending further support for the view that experience with the stimuli can influence the location of category boundaries, much as in human speech (15). Nonlinear responses to continuous variation in sound features thus appear to be common in a range of species. Neurobiological studies of awake swamp sparrows furthermore revealed that such categorical responses may be underpinned by single neurons, which also exhibit categorical responses in relation to changes in note duration. In these experiments, the neural response accurately mapped the learned categorical boundary typical for the local population (11). These studies suggest that birds categorically represent continuous variation in some stimulus features.

 

Lachlan and Nowicki now add an intriguing twist to such categorization processes by investigating whether additional information, such as the position in the syllable, may modulate the categorization of the note into one class or another (3). They first measured the characteristics of a large number of notes. The notes varied in length between about 5 and 50 ms, with three particularly frequently occurring durations, namely 9.5 (short), 17 (intermediate), and 30 ms (long). Lachlan and Nowicki then went on and tested whether the birds would distinguish between these different note lengths, using the habituation-dishabituation paradigm mentioned above. First, they presented syllables with short notes and then switched to intermediate notes (or vice versa), or they first presented intermediate notes and then switched to long ones. The birds’ responses depended on both, the switch type and the position in the syllable, in such a way that the birds ignored switches from short to intermediate notes when they occurred at the beginning of the syllable, whereas they responded strongly to the same switch when it occurred in the final position. Conversely, they responded strongly to switches from intermediate to long notes when they occurred at the beginning of the syllable, but not when they occurred in the final position. Thus, intermediate notes were grouped with short notes when in the initial position, and grouped with long notes at the final position (Fig. 1).

 

Fig. 1.

  1. Swamp sparrow, M. georgiana. Photo courtesy of Rob Lachlan (Queen Mary University of London, London, UK). (B) The birds group intermediate notes with short notes and distinguish them from long notes when they occur at the beginning of the syllable, whereas they group them with long notes when they occur at the end of the syllable. Swamp sparrow songs typically consist of multiple repetitions of a given syllable.

See: http://www.pnas.org/content/112/6/1658.full

PNAS February 10, 2015 vol. 112 no. 6  1658–1659

Trở lại      In      Số lần xem: 775

[ Tin tức liên quan ]___________________________________________________

 

Designed & Powered by WEBSO CO.,LTD