A key characteristic of human language efficiency is that more frequently used words tend to be shorter in length—the ‘law of brevity’. To date, no test of this relationship between frequency of use and length has been carried out on non-human animal vocal communication. We show here that the vocal repertoire of the Formosan macaque (Macaca cyclopis) conforms to the pattern predicted by the law of brevity, with an inverse relationship found between call duration and rate of utterance. This finding provides evidence for coding efficiency in the vocal communication system of this species, and indicates commonality in the basic structure of the coding system between human language and vocal communication in this non-human primate.
Formosan macaque; communication; language; coding; primate
All non-human primates communicate with conspecifics using vocalizations, a system involving both the production and perception of species-specific vocal signals. Much of the work on the neural basis of primate vocal communication in cortex has focused on the sensory processing of vocalizations, while relatively little data are available for vocal production. Earlier physiological studies in squirrel monkeys had shed doubts on the involvement of primate cortex in vocal behaviors. The aim of the present study was to identify areas of common marmoset (Callithrix jacchus) cortex that are potentially involved in vocal communication. In this study, we quantified cFos expression in three areas of marmoset cortex – frontal, temporal (auditory), and medial temporal – under various vocal conditions. Specifically, we examined cFos expression in these cortical areas during the sensory, motor (vocal production), and sensory–motor components of vocal communication. Our results showed an increase in cFos expression in ventrolateral prefrontal cortex as well as the medial and lateral belt areas of auditory cortex in the vocal perception condition. In contrast, subjects in the vocal production condition resulted in increased cFos expression only in dorsal premotor cortex. During the sensory–motor condition (antiphonal calling), subjects exhibited cFos expression in each of the above areas, as well as increased expression in perirhinal cortex. Overall, these results suggest that various cortical areas outside primary auditory cortex are involved in primate vocal communication. These findings pave the way for further physiological studies of the neural basis of primate vocal communication.
immediate early gene expression; common marmoset; vocal communication; frontal cortex; auditory cortex; medial temporal cortex
Non-human primate communication is thought to be fundamentally different from human speech, mainly due to vast differences in vocal control. The lack of these abilities in non-human primates is especially striking if compared to some marine mammals and bird species, which has generated somewhat of an evolutionary conundrum. What are the biological roots and underlying evolutionary pressures of the human ability to voluntarily control sound production and learn the vocal utterances of others? One hypothesis is that this capacity has evolved gradually in humans from an ancestral stage that resembled the vocal behavior of modern primates. Support for this has come from studies that have documented limited vocal flexibility and convergence in different primate species, typically in calls used during social interactions. The mechanisms underlying these patterns, however, are currently unknown. Specifically, it has been difficult to rule out explanations based on genetic relatedness, suggesting that such vocal flexibility may not be the result of social learning.
To address this point, we compared the degree of acoustic similarity of contact calls in free-ranging Campbell's monkeys as a function of their social bonds and genetic relatedness. We calculated three different indices to compare the similarities between the calls' frequency contours, the duration of grooming interactions and the microsatellite-based genetic relatedness between partners. We found a significantly positive relation between bond strength and acoustic similarity that was independent of genetic relatedness.
Genetic factors determine the general species-specific call repertoire of a primate species, while social factors can influence the fine structure of some the call types. The finding is in line with the more general hypothesis that human speech has evolved gradually from earlier primate-like vocal communication.
Male Rocky Mountain elk (Cervus elaphus nelsoni) produce loud and high fundamental frequency bugles during the mating season, in contrast to the male European Red Deer (Cervus elaphus scoticus) who produces loud and low fundamental frequency roaring calls. A critical step in understanding vocal communication is to relate sound complexity to anatomy and physiology in a causal manner. Experimentation at the sound source, often difficult in vivo in mammals, is simulated here by a finite element model of the larynx and a wave propagation model of the vocal tract, both based on the morphology and biomechanics of the elk. The model can produce a wide range of fundamental frequencies. Low fundamental frequencies require low vocal fold strain, but large lung pressure and large glottal flow if sound intensity level is to exceed 70 dB at 10 m distance. A high-frequency bugle requires both large muscular effort (to strain the vocal ligament) and high lung pressure (to overcome phonation threshold pressure), but at least 10 dB more intensity level can be achieved. Glottal efficiency, the ration of radiated sound power to aerodynamic power at the glottis, is higher in elk, suggesting an advantage of high-pitched signaling. This advantage is based on two aspects; first, the lower airflow required for aerodynamic power and, second, an acoustic radiation advantage at higher frequencies. Both signal types are used by the respective males during the mating season and probably serve as honest signals. The two signal types relate differently to physical qualities of the sender. The low-frequency sound (Red Deer call) relates to overall body size via a strong relationship between acoustic parameters and the size of vocal organs and body size. The high-frequency bugle may signal muscular strength and endurance, via a ‘vocalizing at the edge’ mechanism, for which efficiency is critical.
More than 5,000 species of mammals share a basic larynx design. Many of them use the larynx to produce an enormous variability of sounds, but only in a handful of species has the physiology of sound production been studied. It is impracticable in most species because observation requires invasive techniques. Furthermore, many mammals do not spontaneously vocalize if they are manipulated or handled. We have constructed a finite element model of vocal fold tissue vibration on the basis of morphological and biomechanical features of the Rocky Mountain elk vocal organs. Operating within reasonable physiological parameter ranges, it allows the investigation of sound production efficiency as well as selective forces. The model can produce sounds with fundamental frequencies ranging between 60 and 1,200 Hz, covering not only some of the natural vocal repertoire of the elk's high-pitched bugle calls but also those of its close relative, the European Red Deer, who produces low-pitched roaring sounds with a similar anatomy. The approach is of broader interest, first because techniques can be adapted to other mammal species using only landmark anatomical and biomechanical features, and second, because simulations can serve as playbacks for perception studies investigating the role of vocalizations in communication.
Noisy acoustic environments present several challenges for the evolution of acoustic communication systems. Among the most significant is the need to limit degradation of spectro-temporal signal structure in order to maintain communicative efficacy. This can be achieved by selecting for several potentially complementary processes. Selection can act on behavioral mechanisms permitting signalers to control the timing and occurrence of signal production to avoid acoustic interference. Likewise, the signal itself may be the target of selection, biasing the evolution of its structure to comprise acoustic features that avoid interference from ambient noise or degrade minimally in the habitat. Here, we address the latter topic for common marmoset (Callithrix jacchus) long-distance contact vocalizations, known as phee calls. Our aim was to test whether this vocalization is specifically adapted for transmission in a species-typical forest habitat, the Atlantic forests of northeastern Brazil. We combined seasonal analyses of ambient habitat acoustics with experiments in which pure tones, clicks, and vocalizations were broadcast and rerecorded at different distances to characterize signal degradation in the habitat. Ambient sound was analyzed from intervals throughout the day and over rainy and dry seasons, showing temporal regularities across varied timescales. Broadcast experiment results indicated that the tone and click stimuli showed the typically inverse relationship between frequency and signaling efficacy. Although marmoset phee calls degraded over distance with marked predictability compared with artificial sounds, they did not otherwise appear to be specially designed for increased transmission efficacy or minimal interference in this habitat. We discuss these data in the context of other similar studies and evidence of potential behavioral mechanisms for avoiding acoustic interference in order to maintain effective vocal communication in common marmosets.
Callithrix jacchus; vocal communication; behavioral ecology; sound broadcasts; sound window
To understand the evolution of acoustic communication in animals, it is important to distinguish between the structure and the usage of vocal signals, since both aspects are subject to different constraints. In terrestrial mammals, the structure of calls is largely innate, while individuals have a greater ability to actively initiate or withhold calls. In closely related taxa, one would therefore predict a higher flexibility in call usage compared to call structure. In the present study, we investigated the vocal repertoire of free living Guinea baboons (Papio papio) and examined the structure and usage of the animals’ vocal signals. Guinea baboons live in a complex multi-level social organization and exhibit a largely tolerant and affiliative social style, contrary to most other baboon taxa. To classify the vocal repertoire of male and female Guinea baboons, cluster analyses were used and focal observations were conducted to assess the usage of vocal signals in the particular contexts.
In general, the vocal repertoire of Guinea baboons largely corresponded to the vocal repertoire other baboon taxa. The usage of calls, however, differed considerably from other baboon taxa and corresponded with the specific characteristics of the Guinea baboons’ social behaviour. While Guinea baboons showed a diminished usage of contest and display vocalizations (a common pattern observed in chacma baboons), they frequently used vocal signals during affiliative and greeting interactions.
Our study shows that the call structure of primates is largely unaffected by the species’ social system (including grouping patterns and social interactions), while the usage of calls can be more flexibly adjusted, reflecting the quality of social interactions of the individuals. Our results support the view that the primary function of social signals is to regulate social interactions, and therefore the degree of competition and cooperation may be more important to explain variation in call usage than grouping patterns or group size.
Evolution; Vocal communication; Call structure; Call usage; Guinea baboon; Social complexity; Competition
The ability to record well-isolated action potentials from individual neurons in naturally behaving animals is crucial for understanding neural mechanisms underlying natural behaviors. Traditional neurophysiology techniques, however, require the animal to be restrained which often restricts natural behavior. An example is the common marmoset (Callithrix jacchus), a highly vocal New World primate species, used in our laboratory to study the neural correlates of vocal production and sensory feedback. When restrained by traditional neurophysiological techniques marmoset vocal behavior is severely inhibited. Tethered recording systems, while proven effective in rodents pose limitations in arboreal animals such as the marmoset that typically roam in a three-dimensional environment. To overcome these obstacles, we have developed a wireless neural recording technique that is capable of collecting single-unit data from chronically implanted multi-electrodes in freely moving marmosets. A lightweight, low power and low noise wireless transmitter (headstage) is attached to a multi-electrode array placed in the premotor cortex of the marmoset. The wireless headstage is capable of transmitting 15 channels of neural data with signal-to-noise ratio (SNR) comparable to a tethered system. To minimize radio-frequency (RF) and electro-magnetic interference (EMI), the experiments were conducted within a custom designed RF/EMI and acoustically shielded chamber. The individual electrodes of the multi-electrode array were periodically advanced to densely sample the cortical layers. We recorded single-unit data over a period of several months from the frontal cortex of two marmosets. These recordings demonstrate the feasibility of using our wireless recording method to study single neuron activity in freely roaming primates.
Action potential; free-roaming; marmoset; multi-channel; multi-electrode array; natural behavior; neurophysiology; neural telemetry; primate; single-unit; vocalization; wireless
Spoken language and learned song are complex communication behaviors found in only a few species, including humans and three groups of distantly related birds – songbirds, parrots, and hummingbirds. Despite their large phylogenetic distances, these vocal learners show convergent behaviors and associated brain pathways for vocal communication. However, it is not clear whether this behavioral and anatomical convergence is associated with molecular convergence. Here we used oligo microarrays to screen for genes differentially regulated in brain nuclei necessary for producing learned vocalizations relative to adjacent brain areas that control other behaviors in avian vocal learners versus vocal non-learners. A top candidate gene in our screen was a calcium-binding protein, parvalbumin (PV). In situ hybridization verification revealed that PV was expressed significantly higher throughout the song motor pathway, including brainstem vocal motor neurons relative to the surrounding brain regions of all distantly related avian vocal learners. This differential expression was specific to PV and vocal learners, as it was not found in avian vocal non-learners nor for control genes in learners and non-learners. Similar to the vocal learning birds, higher PV up-regulation was found in the brainstem tongue motor neurons used for speech production in humans relative to a non-human primate, macaques. These results suggest repeated convergent evolution of differential PV up-regulation in the brains of vocal learners separated by more than 65–300 million years from a common ancestor and that the specialized behaviors of learned song and speech may require extra calcium buffering and signaling.
Plasticity studies suggest that behavioral relevance can change the cortical processing of trained or conditioned sensory stimuli. However, whether this occurs in the context of natural communication, where stimulus significance is acquired through social interaction, has not been well investigated, perhaps because neural responses to species-specific vocalizations can be difficult to interpret within a systematic framework. The ultrasonic communication system between isolated mouse pups and adult females that either do or do not recognize the calls' significance provides an opportunity to explore this issue. We applied an information-based analysis to multi- and single unit data collected from anesthetized mothers and pup-naïve females to quantify how the communicative significance of pup calls affects their encoding in the auditory cortex. The timing and magnitude of information that cortical responses convey (at a 2-ms resolution) for pup call detection and discrimination was significantly improved in mothers compared to naïve females, most likely because of changes in call frequency encoding. This was not the case for a non-natural sound ensemble outside the mouse vocalization repertoire. The results demonstrate that a sensory cortical change in the timing code for communication sounds is correlated with the vocalizations' behavioral relevance, potentially enhancing functional processing by improving its signal to noise ratio.
Like a student in a foreign country immersed in an unfamiliar language or a young mother trying to decipher her baby's cries, we all encounter initially meaningless sounds that in fact carry meaning. As these sounds gain significance, we become better at detecting and discriminating between them. How does this occur? What happens in our brain to facilitate this improvement? We explored these questions in a mouse model by measuring how neurons in the auditory cortex of female mice respond when the ultrasonic calls of mouse pups are played back to the animals. Earlier studies demonstrated that mothers, but not virgin females, recognize these calls as behaviorally significant. Our results indicate that the timing and magnitude of the auditory cortical responses to these communicative sounds differ between these two groups of female mice and that this difference may provide the auditory system in mothers with the capacity for detecting and discriminating pup calls. The results demonstrate that behavioral significance can be correlated with quantifiable functional improvements in the sensory cortical representation of a communication sound.
Pup calls produce quicker and larger neural responses, which convey more information for pup call detection and discrimination, in the auditory cortex of mother mice compared with virgin female mice.
Primates are intensely social and exhibit extreme variation in social structure, making them particularly well suited for uncovering evolutionary connections between sociality and vocal complexity. Although comparative studies find a correlation between social and vocal complexity, the function of large vocal repertoires in more complex societies remains unclear. We compared the vocal complexity found in primates to both mammals in general and human language in particular and found that non-human primates are not unusual in the complexity of their vocal repertoires. To better understand the function of vocal complexity within primates, we compared two closely related primates (chacma baboons and geladas) that differ in their ecology and social structures. A key difference is that gelada males form long-term bonds with the 2–12 females in their harem-like reproductive unit, while chacma males primarily form temporary consortships with females. We identified homologous and non-homologous calls and related the use of the derived non-homologous calls to specific social situations. We found that the socially complex (but ecologically simple) geladas have larger vocal repertoires. Derived vocalizations of geladas were primarily used by leader males in affiliative interactions with ‘their’ females. The derived calls were frequently used following fights within the unit suggesting that maintaining cross-sex bonds within a reproductive unit contributed to this instance of evolved vocal complexity. Thus, our comparison highlights the utility of using closely related species to better understand the function of vocal complexity.
derived vocalizations; group size; homologous vocalizations; social complexity; vocal complexity; vocal repertoire
All animals are anatomically constrained in the number of discrete call types they can produce. Recent studies suggest that by combining existing calls into meaningful sequences, animals can increase the information content of their vocal repertoire despite these constraints. Additionally, signalers can use vocal signatures or cues correlated to other individual traits or contexts to increase the information encoded in their vocalizations. However, encoding multiple vocal signatures or cues using the same components of vocalizations usually reduces the signals' reliability. Segregation of information could effectively circumvent this trade-off. In this study we investigate how banded mongooses (Mungos mungo) encode multiple vocal signatures or cues in their frequently emitted graded single syllable close calls.
The data for this study were collected on a wild, but habituated, population of banded mongooses. Using behavioral observations and acoustical analysis we found that close calls contain two acoustically different segments. The first being stable and individually distinct, and the second being graded and correlating with the current behavior of the individual, whether it is digging, searching or moving. This provides evidence of Marler's hypothesis on temporal segregation of information within a single syllable call type. Additionally, our work represents an example of an identity cue integrated as a discrete segment within a single call that is independent from context. This likely functions to avoid ambiguity between individuals or receivers having to keep track of several context-specific identity cues.
Our study provides the first evidence of segmental concatenation of information within a single syllable in non-human vocalizations. By reviewing descriptions of call structures in the literature, we suggest a general application of this mechanism. Our study indicates that temporal segregation and segmental concatenation of vocal signatures or cues is likely a common, but so far neglected, dimension of information coding in animal vocal communication. We argue that temporal segregation of vocal signatures and cues evolves in species where communication of multiple unambiguous signals is crucial, but is limited by the number of call types produced.
vocal signature; vocal cue; syllable; close call; segregation of information; graded calls; banded mongoose; segmental concatenation
Comparative analyses used to reconstruct the evolution of traits associated with the human language faculty, including its socio-cognitive underpinnings, highlight the importance of evolutionary constraints limiting vocal learning in non-human primates. After a brief overview of this field of research and the neural basis of primate vocalizations, we review studies that have addressed the genetic basis of usage and structure of ultrasonic communication in mice, with a focus on the gene FOXP2 involved in specific language impairments and neuroligin genes (NL-3 and NL-4) involved in autism spectrum disorders. Knockout of FoxP2 leads to reduced vocal behavior and eventually premature death. Introducing the human variant of FoxP2 protein into mice, in contrast, results in shifts in frequency and modulation of pup ultrasonic vocalizations. Knockout of NL-3 and NL-4 in mice diminishes social behavior and vocalizations. Although such studies may provide insights into the molecular and neural basis of social and communicative behavior, the structure of mouse vocalizations is largely innate, limiting the suitability of the mouse model to study human speech, a learned mode of production. Although knockout or replacement of single genes has perceptible effects on behavior, these genes are part of larger networks whose functions remain poorly understood. In humans, for instance, deficiencies in NL-4 can lead to a broad spectrum of disorders, suggesting that further factors (experiential and/or genetic) contribute to the variation in clinical symptoms. The precise nature as well as the interaction of these factors is yet to be determined.
Autism; communication; evolution; FOXP2; mice; neuroligin; speech; ultrasound; vocalization
Researchers have described multilevel societies with one-male, multifemale units (OMUs) forming within a larger group in several catarrhine species, but not in platyrhines. OMUs in multilevel societies are associated with extremely large group sizes, often with >100 individuals, and the only platyrhine genus that forms groups of this size is Cacajao. We review available evidence for multilevel organization and the formation of OMUs in groups of Cacajao, and test predictions for the frequency distribution patterns of male–male and male–female interindividual distances within groups of red-faced uakaris (Cacajao calvus ucayalii), comparing year-round data with those collected at the peak of the breeding season, when group cohesion may be more pronounced. Groups of Cacajao fission and fuse, forming subgroup sizes at frequencies consistent with an OMU organization. In Cacajao calvus ucayalii and Cacajao calvus calvus, bachelor groups are also observed, a characteristic of several catarrhine species that form OMUs. However, researchers have observed both multimale–multifemale groups and groups with a single male and multiple females in Cacajao calvus. The frequency distributions of interindividual distances for male–male and male–female dyads are consistent with an OMU-based organization, but alternative interpretations of these data are possible. The distribution of interindividual distances collected during the peak breeding season differed from those collected year-round, indicating seasonal changes in the spatial organization of Cacajao calvus ucayalii. We suggest a high degree of flexibility may characterize the social organization of Cacajao calvus ucayalii, which may form OMUs under certain conditions. Further studies with identifiable individuals, thus far not possible in Cacajao, are required to confirm the social organization.
Breeding system; Mating system; One-male unit; Pitheciine; Social structure
Response properties of primary auditory cortical neurons in the adult common marmoset monkey (Callithrix jacchus) were modified by extensive exposure to altered vocalizations that were self-generated and rehearsed frequently. A laryngeal apparatus modification procedure permanently lowered the frequency content of the native twitter call, a complex communication vocalization consisting of a series of frequency modulation (FM) sweeps. Monkeys vocalized shortly after this procedure and maintained voicing efforts until physiological evaluation 5–15 months later. The altered twitter calls improved overtime, with FM sweeps approaching but never reaching the normal spectral range. Neurons with characteristic frequencies <4.3 kHz that had been weakly activated by native twitter calls were recruited to encode self-uttered altered twitter vocalizations. These neurons showed a decrease in response magnitude and an increase in temporal dispersion of response timing to twitter call and parametric FM stimuli but a normal response profile to pure tone stimuli. Tonotopic maps in voice-modified monkeys were not distorted. These findings suggest a previously unrecognized form of cortical plasticity that is specific to higher-order processes involved in the discrimination of more complex sounds, such as species-specific vocalizations.
auditory cortex; plasticity; primate; vocalization; learning; twitter call
High background noise is an important obstacle in successful signal detection and perception of an intended acoustic signal. To overcome this problem, many animals modify their acoustic signal by increasing the repetition rate, duration, amplitude or frequency range of the signal. An alternative method to ensure successful signal reception, yet to be tested in animals, involves the use of two different types of signal, where one signal type may enhance the other in periods of high background noise. Humpback whale communication signals comprise two different types: vocal signals, and surface-generated signals such as ‘breaching’ or ‘pectoral slapping’. We found that humpback whales gradually switched from primarily vocal to primarily surface-generated communication in increasing wind speeds and background noise levels, though kept both signal types in their repertoire. Vocal signals have the advantage of having higher information content but may have the disadvantage of loosing this information in a noisy environment. Surface-generated sounds have energy distributed over a greater frequency range and may be less likely to become confused in periods of high wind-generated noise but have less information content when compared with vocal sounds. Therefore, surface-generated sounds may improve detection or enhance the perception of vocal signals in a noisy environment.
acoustic communication; humpback whales; background noise; acoustic behaviour; communication strategy
Neuronal activity in single prefrontal neurons has been correlated with behavioral responses, rules, task variables and stimulus features. In the non-human primate, neurons recorded in ventrolateral prefrontal cortex (VLPFC) have been found to respond to species-specific vocalizations. Previous studies have found multisensory neurons which respond to simultaneously presented faces and vocalizations in this region. Behavioral data suggests that face and vocal information are inextricably linked in animals and humans and therefore may also be tightly linked in the coding of communication calls in prefrontal neurons. In this study we therefore examined the role of VLPFC in encoding vocalization call type information. Specifically, we examined previously recorded single unit responses from the VLPFC in awake, behaving rhesus macaques in response to 3 types of species-specific vocalizations made by 3 individual callers. Analysis of responses by vocalization call type and caller identity showed that ∽ 19 % of cells had a main effect of call type with fewer cells encoding caller. Classification performance of VLPFC neurons was ∽ 42% averaged across the population. When assessed at discrete time bins, classification performance reached 70 percent for coos in the first 300 ms and remained above chance for the duration of the response period, though performance was lower for other call types. In light of the sub-optimal classification performance of the majority of VLPFC neurons when only vocal information is present, and the recent evidence that most VLPFC neurons are multisensory, the potential enhancement of classification with the addition of accompanying face information is discussed and additional studies recommended. Behavioral and neuronal evidence has shown a considerable benefit in recognition and memory performance when faces and voices are presented simultaneously. In the natural environment both facial and vocalization information is present simultaneously and neural systems no doubt evolved to integrate multisensory stimuli during recognition.
Auditory; Communication; Frontal Lobe; Macaque; Monkey; Language
Bats are among the most gregarious and vocal mammals, with some species demonstrating a diverse repertoire of syllables under a variety of behavioral contexts. Despite extensive characterization of big brown bat (Eptesicus fuscus) biosonar signals, there have been no detailed studies of adult social vocalizations. We recorded and analyzed social vocalizations and associated behaviors of captive big brown bats under four behavioral contexts: low aggression, medium aggression, high aggression, and appeasement. Even limited to these contexts, big brown bats possess a rich repertoire of social vocalizations, with 18 distinct syllable types automatically classified using a spectrogram cross-correlation procedure. For each behavioral context, we describe vocalizations in terms of syllable acoustics, temporal emission patterns, and typical syllable sequences. Emotion-related acoustic cues are evident within the call structure by context-specific syllable types or variations in the temporal emission pattern. We designed a paradigm that could evoke aggressive vocalizations while monitoring heart rate as an objective measure of internal physiological state. Changes in the magnitude and duration of elevated heart rate scaled to the level of evoked aggression, confirming the behavioral state classifications assessed by vocalizations and behavioral displays. These results reveal a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a caller.
The evolution of the autonomic nervous system provides an organizing principle to interpret the adaptive significance of physiological systems in promoting social behavior and responding to social challenges. This phylogenetic shift in neural regulation of the autonomic nervous system in mammals has produced a neuroanatomically integrated social engagement system, including neural mechanisms that regulate both cardiac vagal tone and muscles involved in vocalization. Mammalian vocalizations are part of a conspecific social communication system, with several mammalian species modulating acoustic features of vocalizations to signal affective state. Prosody, defined by variations in rhythm and pitch, is a feature of mammalian vocalizations that communicate emotion and affective state. While the covariation between physiological state and the acoustic frequencies of vocalizations is neurophysiologically based, few studies have investigated the covariation between vocal prosody and autonomic state. In response to this paucity of scientific evidence, the current study explored the utility of vocal prosody as a sensitive index of autonomic activity in human infants during the Still Face challenge. Overall, significant correlations were observed between several acoustic features of the infant vocalizations and autonomic state, demonstrating an association between shorter heart period and reductions in heart period and respiratory sinus arrhythmia following the challenge with the dampening of the modulation of acoustic features (fundamental frequency, variance, 50% bandwidth, and duration) that are perceived as prosody.
Infant vocalizations; Prosody; Polyvagal Theory; Autonomic nervous system; Heart rate; Respiratory sinus arrhythmia
Across mammals many vocal sounds are produced by airflow induced vocal fold oscillation. We tested the hypothesis that stress-strain and stress-relaxation behavior of rat vocal folds can be used to predict the fundamental frequency range of the species’ vocal repertoire. In a first approximation vocal fold oscillation has been modeled by the string model but it is not known whether this concept equally applies to large and small species. The shorter the vocal fold, the more the ideal string law may underestimate normal mode frequencies. To accommodate the very small size of the tissue specimen, a custom-built miniaturized tensile test apparatus was developed. Tissue properties of 6 male rat vocal folds were measured. Rat vocal folds demonstrated the typical linear stress-strain behavior in the low strain region and an exponential stress response at strains larger than about 40%. Approximating the rat’s vocal fold oscillation with the string model suggests that fundamental frequencies up to about 6 kHz can be produced, which agrees with frequencies reported for audible rat vocalization. Individual differences and time-dependent changes in the tissue properties parallel findings in other species, and are interpreted as universal features of the laryngeal sound source.
larynx; viscoelastic properties; bioacoustics; anisotropy
Human-generated noise pollution now permeates natural habitats worldwide, presenting evolutionarily novel acoustic conditions unprecedented to most landscapes. These acoustics not only harm humans, but threaten wildlife, and especially birds, via changes to species densities, foraging behavior, reproductive success, and predator-prey interactions. Explanations for negative effects of noise on birds include disruption of acoustic communication through energetic masking, potentially forcing species that rely upon acoustic communication to abandon otherwise suitable areas. However, this hypothesis has not been adequately tested because confounding stimuli often co-vary with noise and are difficult to separate from noise exposure.
Using a natural experiment that controls for confounding stimuli, we evaluate whether species vocal features or urban-tolerance classifications explain their responses to noise measured through habitat use. Two data sets representing nesting and abundance responses reveal that noise filters bird communities nonrandomly. Signal duration and urban tolerance failed to explain species-specific responses, but birds with low-frequency signals that are more susceptible to masking from noise avoided noisy areas and birds with higher frequency vocalizations remained. Signal frequency was also negatively correlated with body mass, suggesting that larger birds may be more sensitive to noise due to the link between body size and vocal frequency.
Our findings suggest that acoustic masking by noise may be a strong selective force shaping the ecology of birds worldwide. Larger birds with lower frequency signals may be excluded from noisy areas, whereas smaller species persist via transmission of higher frequency signals. We discuss our findings as they relate to interspecific relationships among body size, vocal amplitude and frequency and suggest that they are immediately relevant to the global problem of increases in noise by providing critical insight as to which species traits influence tolerance of these novel acoustics.
Spoken language is a result of the human capacity to assemble simple vocal units into more complex utterances, the basic carriers of semantic information. Not much is known about the evolutionary origins of this behaviour. The vocal abilities of non-human primates are relatively unimpressive in comparison, with gibbon songs being a rare exception. These apes assemble a repertoire of call notes into elaborate songs, which function to repel conspecific intruders, advertise pair bonds, and attract mates. We conducted a series of field experiments with white-handed gibbons at Khao Yai National Park, Thailand, which showed that this ape species uses songs also to protect themselves against predation. We compared the acoustic structure of predatory-induced songs with regular songs that were given as part of their daily routine. Predator-induced songs were identical to normal songs in the call note repertoire, but we found consistent differences in how the notes were assembled into songs. The responses of out-of-sight receivers demonstrated that these syntactic differences were meaningful to conspecifics. Our study provides the first evidence of referential signalling in a free-ranging ape species, based on a communication system that utilises combinatorial rules.
Mice produce ultrasonic vocalizations in various inter-individual encounters and with high call rates. However, it is so far virtually unknown how these vocal patterns are generated. On the one hand, these vocal patterns could be embedded into the normal respiratory cycle, as happens in bats and other mammals that produce similar call rates and frequencies. On the other, mice could possess distinct vocal pattern generating systems that are capable of modulating the respiratory cycle, which is what happens in non-human and human primates. In the present study, we investigated the temporal call patterns of two different mammalian species, bats and mice, in order to differentiate between these two possibilities for mouse vocalizations. Our primary focus was on comparing the mechanisms for the production of rapid, successive ultrasound calls of comparable frequency ranges in the two species.
We analyzed the temporal call pattern characteristics of mice, and we compared these characteristics to those of ultrasonic echolocation calls produced by horseshoe bats. We measured the distributions of call durations, call intervals, and inter-call intervals in the two species. In the bat, and consistent with previous studies, we found that call duration was independent of corresponding call intervals, and that it was negatively correlated with the corresponding inter-call interval. This indicates that echolocation call production mechanisms in the bat are highly correlated with the respiratory cycle. In contrast, call intervals in the mouse were directly correlated with call duration. Importantly, call duration was not, or was only slightly, correlated with inter-call intervals, consistent with the idea that vocal production in the mouse is largely independent of the respiratory cycle.
Our findings suggest that ultrasonic vocalizations in mice are produced by call-pattern generating mechanisms that seem to be similar to those that have been found in primates. This is in contrast to the production mechanisms of ultrasonic echolocation calls in horseshoe bats. These results are particularly interesting, especially since mouse vocalizations have recently attracted increased attention as potential indicators for the degree of progression of several disease patterns in mouse models for neurodegenerative and neurodevelopmental disorders of humans.
Acoustic communication; Bat; Mammal; Mice; Mouse model; Vocal pattern generation
The structure and function of primate communication have attracted much attention, and vocal signals, in particular, have been studied in detail. As a general rule, larger social groups emit more types of vocal signals, including those conveying the presence of specific types of predators. The adaptive advantages of receiving and responding to alarm calls are expected to exert a selective pressure on the auditory system. Yet, the comparative biology of primate hearing is limited to select species, and little attention has been paid to the effects of social and vocal complexity on hearing. Here, we use the auditory brainstem response method to generate the largest number of standardized audiograms available for any primate radiation. We compared the auditory sensitivities of 11 strepsirrhine species with and without independent contrasts and show that social complexity explains a significant amount of variation in two audiometric parameters—overall sensitivity and high-frequency limit. We verified the generality of this latter result by augmenting our analysis with published data from nine species spanning the primate order. To account for these findings, we develop and test a model of social drive. We hypothesize that social complexity has favoured enhanced hearing sensitivities, especially at higher frequencies.
Strepsirrhini; hearing; auditory brainstem response; evoked potential audiogram; sensory ecology; acoustic communication
The ubiquity of social vocalization among animals provides the opportunity to identify conserved mechanisms of auditory processing that subserve vocal communication. Identifying auditory coding properties that are shared across vocal communicators will provide insight into how human auditory processing leads to speech perception. Here, we compare auditory response properties and neural coding of social vocalizations in auditory midbrain neurons of mammalian and avian vocal communicators. The auditory midbrain is a nexus of auditory processing because it receives and integrates information from multiple parallel pathways and provides the ascending auditory input to the thalamus. The auditory midbrain is also the first region in the ascending auditory system where neurons show complex tuning properties that are correlated with the acoustics of social vocalizations. Single unit studies in mice, bats and zebra finches reveal shared principles of auditory coding including tonotopy, excitatory and inhibitory interactions that shape responses to vocal signals, nonlinear response properties that are important for auditory coding of social vocalizations and modulation tuning. Additionally, single neuron responses in the mouse and songbird midbrain are reliable, selective for specific syllables, and rely on spike timing for neural discrimination of distinct vocalizations. We propose that future research on auditory coding of vocalizations in mouse and songbird midbrain neurons adopt similar experimental and analytical approaches so that conserved principles of vocalization coding may be distinguished from those that are specialized for each species.
vocal communication; subcortical; auditory tuning; inferior colliculus; birdsong; comparative
Among mammals living in social groups, individuals form communication networks where they signal their identity and social status, facilitating social interaction. In spite of its importance for understanding of mammalian societies, the coding of individual-related information in the vocal signals of non-primate mammals has been relatively neglected. The present study focuses on the spotted hyena Crocuta crocuta, a social carnivore known for its complex female-dominated society. We investigate if and how the well-known hyena's laugh, also known as the giggle call, encodes information about the emitter.
By analyzing acoustic structure in both temporal and frequency domains, we show that the hyena's laugh can encode information about age, individual identity and dominant/subordinate status, providing cues to receivers that could enable assessment of the social position of an emitting individual.
The range of messages encoded in the hyena's laugh is likely to play a role during social interactions. This call, together with other vocalizations and other sensory channels, should ensure an array of communication signals that support the complex social system of the spotted hyena. Experimental studies are now needed to decipher precisely the communication network of this species.