SciELO - Scientific Electronic Library Online

 
vol.16 número1Matemática cotidiana e matemática científicaPiscar atencional e as limitações do processamento de informações visuais conforme a frequência temporal índice de autoresíndice de materiabúsqueda de artículos
Home Pagelista alfabética de revistas  

Ciências & Cognição

versión On-line ISSN 1806-5821

Ciênc. cogn. vol.16 no.1 Rio de Janeiro dic. 2011

 

Revisão

 

Brain and the music: A window to the comprehension of the interactive brain functioning

 

Música e cérebro: Uma janela para o entendimento do funcionamento interativo do cérebro

 

 

Paulo Estêvão AndradeI and Elisabete Castelon KonkiewitzII

IDepartamento de Estudos Pedagógicos do Ensino Fundamental, "Colégio Criativo", Marilia, São Paulo, Brazil;
IIFaculdade de Ciências da Saúde, Universidade Federal da Grande Dourados (UFGD), Dourados, Mato Grosso do Sul, Brazil

Endereço para correspondência

 

 


Abstract

Here we review the most important psychological aspects of music, its neural substrates, its universality and likely biological origins and, finally, how the study of neurocognition and emotions of music can provide one of the most important windows to the comprehension of the higher brain functioning and human mind. We begin with the main aspects of the theory of modularity, its behavioral and neuropsychological evidences. Then, we discuss basic psychology and neuropsychology of music and show how music and language are cognitively and neurofunctionally related. Subsequently we briefly present the evidences against the view of a high degree of specification and encapsulation of the putative language module, and how the ethnomusicological, pscychological and neurocognitive studies on music help to shed light on the issues of modularity and evolution, and appear to give further support for a cross-modal, interactive view of neurocognitive processes. Finally, we will argue that the notion of large modules do not adequately describe the organization of complex brain functions such as language, math or music, and propose a less radical view of modularity, in which the modular systems are specified not at the level of culturally determined cognitive domains but more at the level of perceptual and sensorimotor representations. © Cien. Cogn. 2011; Vol. 16 (1): 137-164.

Keywords: music; language; modularity; brain; evolution; cognition.


Resumo

Aqui revisamos os aspectos psicológicos mais importantes da música, seus substratos neurais, sua universalidade e prováveis origens biológicas e, finalmente, como o estudo da neurocognição e das emoções associadas à música pode fornecer uma das mais importantes janelas para a compreensão das funções cerebrais superiores e da mente humana. Iniciamos com os principais aspectos da teoria da modularidade, suas evidências comportamentais e neuropsicológicas. Então discutimos a psicologia e a neuropsicologia básicas da música e mostramos como a música e a linguagem estão cognitiva e neurofuncionalmente interrelacionadas. Em seguida, apresentamos brevemente as evidências contrárias à visão de um alto grau de especificação e encapsulamento do módulo putativo da linguagem e como os estudos etnomusicológicos, psicológicos e neurocognitivos sobre música auxiliam na elucidação dos temas modularidade e evolução e parecem fornecer subsídios adicionais para uma visão interativa e intermodal dos processos neurocognitivos. Finalmente iremos argumentar que a noção de amplos módulos não descreve adequadamente a organização de funções cerebrais complexas, como a linguagem, a matemática, ou a música. Propomos uma visão menos radical de modularidade, na qual os sistemas modulares são especificados não ao nível de domínios cognitivos culturalmente determinados, e sim mais ao nível de representações perceptuais e sensoriomotoras. © Cien. Cogn. 2011; Vol. 16 (1): 137-164.

Palavras-chave: música; linguagem; modularidade; cérebro; evolução; cognição.


 

 

Introduction

Nowadays, the principles of natural and sexual selection are being used to guide theoretical and empirical research in the behavioral and social sciences with increasing frequency, and nearly all of this research has focused on social behavior, cognitive mechanisms, and other phenomenon that are thought to be evolved as universal adaptations, such as features of human behavior that are evident in all cultures as, for instance, language(Geary, 2001). Here, the core idea is that some current behaviors are universally present in the human cultures because it would have been evolutionary efficacious. From the perspective of evolutionary psychology, natural selection shapes species to their environment not only in physical and physiological traits but also in behavioral traits, and any trait that has evolved with time is likely to be distributed universally, to be found in immature members of the species, and to be processed with some degree of automation.

Consider language as an example. All languages sum approximately 150 different phonemes, but each human languages contain about 25 to 40 phonetic units, varying from 11 in Polynesian to 141 in the language of the Bushmen(Pinker, 1994), and for any given language all speakers distinguish between grammatical and ungrammatical sentences without any formal instruction, suggesting universal linguistic principles defining sentence structure. This led Chomsky to conclude that children have implicit knowledge of both phonemes and abstract grammatical rules, the latter assumed to be a very general kind allowing them to learn whichever language they are exposed to, rather than just learning by imitating the language they hear (Chomsky, 1957). Thus, language was one of the primary examples of what Fodor (1983) called a mental module in his theory of modularity of mind, where a module is described as domain-specific (only activated by language-specific information), informational encapsulated (do not interfere with other modules), and innate.

In addition to the anthropological (universalities) and behavioral/psychological evidences, the modular approach of mind is further supported by neuropsychological cases. Brain lesions causing impairments in one domain while sparing others, and selective congenital deficits in one domain in the presence of otherwise completely normal cognitive functions have been taken as a strong support for domain-specific neural networks of some cognitive domains, particularly language, and more recently also number processing (Dehaene, Dehaene-Lambertz & Cohen, 1998) and music(Dalla Bella & Peretz, 1999; Peretz & Coltheart, 2003). Evolutionary psychologists have incorporated with much enthusiasm the hypothesis that some cognitive mechanisms and its respective neural substrates can be viewed as individual traits that have been subject to selection and perfected in recent human evolutionary history. In their view, some cognitive mechanisms involved in language, or mathematics, would be a highly complex and specific adaptation evolved with admirable effectiveness, have a genetic component and, finally, have no parallel in nonhuman animals(Geary & Huffman, 2002). In this perspective, although homologous mechanisms may exist in other animals, the human versions have been modified by natural selection to the extent that they can be reasonably seen as constituting novel traits (e.g., social intelligence, tool-making)(Bickerton, 1990; Dunbar, 1996; Kimura, 1993; Lieberman, 1984).

However, a strong modular view, predominant in the evolutionary psychology, is not a consensus in the field of cognitive neuroscience. Both lesion(Bookheimer, 2002; Koelsch, et al., 2002; Koelsch et al., 2004), and developmental(Bishop, 1999) studies do not unequivocally support the idea that language is an autonomous function relying on its own encapsulated structural and functional architecture. In addition, to make the claim that a trait evolved uniquely in humans for the function of language processing, it must be demonstrated that no other animal has this particular trait, because any trait present in nonhuman animals did not evolve specifically for human language, although it may be part of the language faculty and play an intimate role in language processing. In this respect, recent studies have shown that the sensory-motor and conceptual intentional systems believed to be language-specific are also present in non-human animals itative (Gil-da-Costa et al., 2004; Hauser, Chomsky & Fitch, 2002).

An alternative view is the interactive approach where cognitive outcomes arise from mutual, bidirectional interactions among neuronal populations representing different types of information, and many primary brain areas (e.g., primary visual or auditory cortices) appear to participate in the representation of the global structure of a stimulus or response situation, and these brain areas that are treated as modality specific can be modulated by influences from other modalities (McClelland, 2001). The discovery of various cross-modal interactions in recent years suggests that these interactions are the rule rather than exception in human perceptual processing (Macaluso & Driver, 2005), and even the visual modality, which has long been viewed as the dominant modality, has been shown to be affected by signals of other sensory modalities (Foxe & Schroeder, 2005; Murray et al., 2005; Saint-Amour, Saron, Schroeder & Foxe, 2005; Schroeder & Foxe, 2005; Shams, Iwaki, Chawla & Bhattacharya, 2005; Shimojo & Shams, 2001). Thus, in place of the notion that complex brain functions such as language, math or music, are organized in specific and encapsulated large modules, we present a less radical view of modularity, in which the modular systems are domain-specific not in terms of encapsulation, but in terms of functional specializations for their crucial relevance for a particular domain (Barret & Kurzban, 2006).

Since music is a stimulus involving complex auditory and motor patterns and representations, and also has an enormous power in generating complex emotions and abstract moods, we believe that its study offers a suitable window to understand the higher cognitive functioning of the human brain and to observe and to evaluate the extent up to which the brain has a modular organization.

 

Basic sensory-perceptive and cognitive aspects of music and language

Music, like verbal language, is based on the intentionally structured patterns of pitch, duration and intensity. Both domains are composed of sequences of basic auditory events, namely, phonemes (vowels and consonants) in language and pitches (melody and chords) and percussive sounds in music.

The 'pitch' in language (determined mainly by the trajectory of fundamental frequency (fo) of the vowels) originates the intonation of speech that contributes to mark the boundaries of structural units, distinguishes pragmatic categories of utterance and conveying intentional and affective cue (Patel, Peretz, Tramo & Labreque, 1998). However, perception of the rapid acoustic transitions, i.e. noise bursts or formant transitions between consonants and vowels and vice-versa, within the range of tens of milliseconds have been empirically shown to be the crucial aspect for the discrimination of phonemes in language. For instance, when the spectral information is greatly reduced in artificially degraded speech, but with temporal variations preserved, adult listeners and 10 to 12 years old children are still able to recognize speech, although younger children of 5 to 7 years old required significantly more spectral resolution. Thus, spectrally degraded speech with as few as two spectral channels still allows relatively good speech comprehension, indicating that the perception of temporal changes is more relevant for speech than the spectral cues (Eisenberg, Shannon, Martinez, Wygonski & Boothroyd, 2000; Shannon, Zeng, Kamath, Wygonski & Ekelid, 1995).

Conversely, the primary acoustic cue in music is the trajectory of fundamental frequency (fo) of pitches of the musical notes. Despite the importance of rapid temporal succession of tones to musical phrasing and expression, the gist of musical content relies on the succession of pitches, and the time intervals between them in music are normally greater than in language. However, it is difficult to precisely determine the time intervals common to music, as it greatly varies between musical styles and cultures. Zatorre, Belin e Penhune (2002) have noted that musical notes are typically much longer in duration than the range characteristic of consonant phonemes, what is definitely true, and that melodies with note durations shorter than about 160 ms are very difficult to identify. In sum, the gist is that the interval times between phonemes in verbal language are still much shorter, in the range of 10-30 ms, than between the notes in the musical rhythms as well the suprasegmental rhythm in language (Phillips & Farmer, 1990).

Cross-culturally, music involves not only patterned sound, but also overt and/or covert action, and even 'passive' listening to music can involve activation of brain regions concerned with movement (Janata & Grafton, 2003): it is not a coincidence that semantically speaking music is, unlike language, both embodied (involving not only sound but also action) and polysemic (mutable in its specific significances or meanings) (Cross, 2001; Cross, 2003b). Developmental precursors of music in infancy, and through early childhood, occur in the form of universal proto-musical behaviors (Trevarthen, 2000) which are exploratory and kinesthetically embedded and closely bound to vocal play and to whole body movement, and are universal. This leads some cultures to employ terms to define music that are far more inclusive than the western notion of music, like the word nkwa that for the Igbo people of Nigeria denotes "singing, playing instruments and dancing" (Cross, 2001). Music functions in many different ways across cultures, from a medium for communication (the Kaluli of Papua New Guinea used music to communicate with dead members), for restructuring social relations (the Venda tribes use music for the domba initiation), to constitute a path to healing and establishing sexual relationships (Tribes in northwest China play the 'flower songs', hua'er) (Cross, 2003b). Thus, music is a property not only of individual cognitions and behaviors but also of inter-individual interactions. Developmental studies give further support for the notion that music is embodied.

Based on these arguments, we consider a re-definition of music in terms of both sound and of movement and their relationships between them (Cross, 2001, 2003a, 2003b). Here, music is defined as a sound-based, embodied, non-referential and polysemic form of communication whose content is essentially emotional, built on the organization of sounds in two main dimensions, pitch and time.

 

Music, evolutionary psychology, and modularity

Some theorists (Pinker, 1994; Pinker, 1999) see no adaptive role for human musicality in evolution because it has no evident and immediate efficacy or fixed consensual reference (but see (Cross, 2001)). Unlike language, music is not referential: musical sounds do not refer to another object external to it, be it concrete or abstract, what led to the precipitous conclusion that music does not have an obvious adaptive value. However, this notion contrasts with the fact that neuropsychological, developmental and anthropological evidences are as strong for modularity of music as that for modularity of language, including the isolation of dedicated neural circuits in lesions and congenital musical deficits (Andrade & Bhattacharya, 2003; Peretz & Hebert, 2000; Trehub, 2003). Briefly, it includes double dissociation between music and language as well as selective impairments of musical subfunctions, congenital deficits specific to musical ability called congenital amusia (in which the most basic musical abilities are lacking, such as recognition of pitch changes and discrimination of one melody from another, despite intact intellect and language skills), universality of common features across different music styles and cultures, in the principles underlying melody, consonant/dissonant intervals, pitch in musical scales and metre and, finally, infants' musical abilities are similar to that of adults in the processing of melody, consonance and rhythms.

 

Music cognition and modularity

Cognitively speaking, the pitch and temporal dimensions in music can be separated. The pitch dimension itself can be viewed under three distinct, although related, aspects: (i) the pitch direction that contribute to global pitch contour of the melody, (ii) the exact pitch distance between the notes that define the local musical intervals, and, (iii) the simultaneous combination of pitches that forms the chords. We can recognize melodies just by the global perception of its pitch contour, being the local interval analysis dispensable in this global task. The local interval analysis, in turn, is crucial to discriminate short melodies (motifs) with same contours, differing only by one note, in the construction and discrimination of scales (a particular subset of unidirectional sequences of pitches- equal contours), and, finally, in the appreciation and analysis of chords, once they are "vertical" structures that do not have a contour, and differentiate from each other in its interval structure.

On the pitch dimension we focus specifically on melody, defined as successive combination of tones varying in pitch and duration, because it is a universal human phenomenon that features prominently in most kinds of music around the world and can be traced back to prehistoric times (Eerola, 2004).

An early source of evidence for the modular view of language and music has been provided by the clinical cases of auditory agnosia. Amusias consist in an intriguing subtype of auditory agnosia selective for music, preserving perception and recognition of verbal and environmental sounds. Amusias can be divided in apperceptive (recognition is impaired by deficits in perception) and associative (recognition is impaired despite of intact perception) musical agnosia (or amusia), according to the terminology offered by Lissauer (1889). In short, we can say that music perception and recognition systems, in contrast to language, which is predominantly processed in left perisylvian cortex (frontotemporoparietal areas around Sylvian fissure), but almost mirroring it, rely mainly in right frontotemporal regions whose lesions are associated with selective impairments in music perception and/or recognition sparing language (Dalla Bella & Peretz, 1999; Peretz & Coltheart, 2003).

In the temporal organization, researchers have agreed on a conceptual division between the global meter and the local rhythm in music (Drake, 1998). Meter is the global perception of a temporal invariance of recurrent pulses, providing durational units by which we recognize a march (major accent on the first of two beats) or a waltz (major accent in the firs of three beats), and functioning as a "cognitive clock" that allow us to evaluate the exact duration of auditory events (Wilson, Pressing & Wales, 2002). Rhythm is the local perception of the temporal proximity, or clustering of adjacent auditory events that form larger units. This temporal grouping that leads to rhythm perception can occur solely on the basis of the estimation of note duration and, hence, does not require the exact evaluation of the note duration through the meter (Penhune, Zatorre & Feindel, 1999; Wilson et al., 2002).

The existence of distinct local and global cognitive strategies in music perception, either in pitch and temporal dimensions, are supported by behavioral (Gordon, 1978; Peretz & Babai, 1992; Tasaki, 1982), neuropsychological (Liegeois-Chauvel, Peretz, Babai, Laguitton & Chauvel, 1998; Peretz, 1990; Schuppert, Munte, Wieringa & Altenmuller, 2000) and functional neuroimaging (Mazziotta, Phelps, Carson & Kuhl, 1982; Platel et al., 1997) studies. According to the lesion studies, right hemisphere structures, mainly in the right superior temporal lobe (STG), are crucially involved in the fine perception of pitch variations in melody, while contralateral structures in the left hemispheres are crucially involved in the perception of rapid acoustic transitions like that occurring in the perception of consonant/vowel combinations (Zatorre et al., 2002).

Evidences against modularity I: Language and music meet in pitch processing

The infant's incredible ability to virtually discriminate every phonetic contrast of all languages, initially taken as evidence for "innate phonetic feature detectors specifically evolved for speech", was also found when the stimuli are non-speech sounds that mimicked speech features, a categorical perception of phonemes also exhibited by several animal species, and monkeys, including perception of the prosodic cues of speech as well (Kuhl, Tsao, Liu, Zhang & De Boer, 2001). In addition, contrary to the modularity view which argues that linguistic experience produces either maintenance or loss of phonetic detectors, new evidences show that infants exhibit genuine developmental change, not merely maintenance of an initial ability. At 7 months, american, japanese and taiwanese infants performed equally well in discriminating between native and non-native language sounds, whereas at 11 months of age they showed a significant increase in native-language phonetic perception and a decrease in the perception of foreign-language phonemes (Kuhl et al., 2001). An emerging view suggests that infants engage in a new kind of learning in which language input is mapped in detail by the infant brain. The central notion is that by simply listening to language infants can acquire sophisticated information and perceptually ''map'' critical aspects of ambient language in the first year of life, even before they can speak, and this perceptual abilities are universal, but not domain specific or species specific. Further, contrary to the notion that development is based on selection and maintenance of a subset of those units triggered by language input with subsequent loss of those that are not stimulated, infants exhibit a genuine developmental change; furthermore, adults, submitted to techniques that minimize the effects of memory and extensive training, can increase their performance on non-native language phonemes indicating that there is not an immutable loss of phonetic abilities for non-native units (Kuhl, 2000).

Recent neuropsychological evidences show that inferior parietal and temporal regions comprising the Wernicke's area (BA22/39/40), traditionally considered as a specific language area for the processing of phonetics and semantics, was even more important for processing non-verbal than verbal sounds, suggesting that the semantic processing of auditory events, either verbal or non-verbal sounds, relies on the shared neural resources (Saygin, Dick, Wilson, Dronkers & Bates, 2003).

Finally, pitch contour and interval are also used in language (Zatorre, Evans, Meyer & Gjedde, 1992), and when impaired the deficits extend to its processing in both domains (Nicholson et al., 2003; Patel et al., 1998; Steinke, Cuddy & Jakobson, 2001). In fact, prosodic deficits in language previously verified in right brain-lesioned patients with amusia are also found in individuals diagnosed with congenital amusia (Ayotte, Peretz & Hyde, 2002; Hyde & Peretz, 2004). Taken together, behavior, developmental and neuropsychological data clearly confirm that brain specificity for music or language is more relative than absolute. The relative dissociation between music and language relies on the fact that speech intonation contours normally use variations in pitch that are larger than half an octave to convey relevant information, whereas mostly of melodies use small pitch intervals, of the order of 1/12 (the pitch difference between two adjacent keys on the piano) or 1/6 (the pitch difference between two keys separated by only one key on the piano) of an octave. Therefore, a degraded pitch perception system may compromise music perception but leave basic aspects of speech prosody practically unaffected. Congenital amusia are elucidative here.

In the pitch discrimination tasks, congenital amusic subjects are unable to detect a small pitch deviation in either isolated tones or monotonic and isochronous sequences, such as pitch changes lesser than two semitones (normal acuity lies in the order of a quarter of a semitone), a performance that could not be improved even under practice (Peretz et al., 2002), but they perform like controls in detecting a slight time deviation in the same contexts (Ayotte et al., 2002). It indicates that their deficits appear to reside in the pitch dimension. In contrast, pitch variations in speech are well perceived by congenital amusic individuals because they are very coarse compared to those used in music. The final pitch rise that is indicative of a question is normally larger than half an octave (the pitch distance between two notes on the piano separated by twelve keys) (Hyde & Peretz, 2004). These facts are consistent with the notion that, rather than a deficit specific to musical domain, congenital amusia reflect a more elemental problem in the low level processing of auditory information, resulting in deficits on fine-grained discrimination of pitch much in the same way as many language-processing difficulties arise from deficiencies in auditory temporal resolution (Zatorre et al., 2002). Thus, although there are claims that language impairments arise from failures specific to language processing (Studdert-Kennedy & Mody, 1995), language deficiencies may also result from a more elemental problem on the perception of fine acoustic temporal changes (Tallal, Miller & Fitch, 1993). Indeed, difficulties for processing time in its different dimensions is an amazingly universal characteristic in dyslexia, a congenital language deficit present in 5-10% of school-age children who fail to learn to read in spite of normal intelligence, adequate environment and educational opportunities, and which likely relies on the left hemisphere difficulty to integrate rapidly changing stimuli (Habib, 2000).

However it is worthy emphasizing that, jointly with congenital amusia, the existence of associative music agnosias in which patients cannot recognize previously familiar melodies neither detect out-of-key notes purposely inserted on melodies (tonal encoding) (Peretz, 1996; Peretz et al., 1994) despite completely preserved both music perception (pith and melodic discrimination) and general linguistic abilities (including intonational prosody), both represent spectacular cases of isolation of music components in the brain whose common deficit is an impairment of the tonal encoding of pitch due to alterations in an anterior frontotemporal network (Peretz, 2006), perhaps more due to insensitivity to dissonance than to to consonance (Peretz, Blood, Penhune & Zatorre, 2001; Tramo, Cariani, Delgutte & Braida, 2001). Congenital amusia is characterized by little sensitivity to dissonance (a sensitivity that is already present in infants) and, differently from acquired associative amusias, impaired fine-grained pitch perception, both deficits that are devastating for the development of tonal encoding and long-term musical representations, but not for the development of linguistic abilities (Hyde & Peretz, 2004). Although evidence points to shared cognitive mechanisms and neural substrates involved in the fine-grained pitch perception between music and language, there is an increasing consensus that ability is of crucial relevance only for the development of a putative tonal module but not for linguistic abilities (Peretz, 2006).

Taken together, these findings support the notion that at least one cognitive component of music, the tonal encoding, could be an evolved tonal module in a broader concept of modularity, according to which domain-specificity is better characterized in terms of a functional specialization; i.e. computational mechanisms that evolved to handle biologically relevant information in specialized ways according to specific input criteria such as fine-grained pitch perception, but "No mechanism is either encapsulated or unencapsulated in an absolute sense" (Barret & Kurzban, 2006, p. 631; see also Peretz, 2006). Therefore, although music and language appear to be clearly dissociated in functional terms at the level of fine-grained pitch perception both domains can interact in the pitch dimension.

Evidences against modularity I: Language and music can meet in rhythm

Consistently with our contention that is in the time domain where there is the most conspicuous and greater amount of sharing between language and music, the temporal deficits of dyslexic children are not confined to language but also extend to the musical domain, and in an interesting way: they are confined to temporal dimension (meter and rhythm) of music. This link between language and musical temporal deficits in dyslexic children led to a testable hypothesis of musical therapeutics based on group music lessons consisting of singing and rhythm games which, although apparently did not help reading, had a positive effect on both phonologic and spelling skills (Overy, 2003; Overy, Nicolson, Fawcett & Clarke, 2003). Children aged between 4-5 years old show that music have a significantly positive effect on both phonological awareness and reading development, and music perception skills can reliably predict reading ability (Anvari, Trainor, Woodside & Levy, 2002). Thus, in relation to the temporal structures of music, the dissociations between language and music are much less consistent then in the pitch dimension.

Another important temporal feature is rhythm which has been found to be impaired more often in cases of lesions in the left hemisphere which, in turn, are also more activated in the imaging studies. If rhythm is preferentially processed in the left hemisphere, the meter as well as rhythm tasks that require the extraction of exact durations of auditory events, in contrast, are often disturbed by lesions in the right temporal lobe (Penhune et al., 1999; Wilson & Davey, 2002) and/or right fronto-parietal lesions (Parsons, 2001; Sakai et al., 1999). In fact, left hemisphere lesions which induce to Broca's aphasia (Efron, 1963; Mavlov, 1980; Swisher & Hirsh, 1972), or conduction aphasia (Brust, 1980; Eustache Lechevalier, Viader & Lambert, 1990), or anomic aphasia (Pötzl & Uiberall, 1937) are also likely to induce to musical disturbances, mainly in the temporal domain. Normally, the deficits in discrimination of musical rhythms or temporal ordering are due to left hemisphere lesions, mainly left anterior lesions affecting prefrontal cortices, and have been associated with deficits in verbal working memory (Erdonmez & Morley, 1981), deficit in the phonological short-term memory (Peretz, Belleville & Fontaine, 1997), and Broca's aphasia (Efron, 1963; Swisher & Hirsh, 1972), although left prefrontal damage also cause deficits in active listening to pitch variations, such as pitch contour and intervals (Ayotte, Peretz, Rousseau, Bard & Bojanowski, 2000) or to judge if a comparison tone pair was higher or lower than a standard tone pair (Harrington et al., 1998). Imaging studies on health individuals provide additional evidence that rhythm is processed in left hemisphere, more specifically in left inferior frontal areas such as Broca's area.

Interestingly, the relations between music and language on the temporal dimension are consistent with a long held notion among musicologists that have recently received empirical support: the rhythm of a nation's music reflects the rhythm of its language. New methods have allowed to directly compared rhythm and melody in speech and music, and between different languages such as British English and French (Patel, 2003b; Patel & Daniele, 2003), and reveal that the greater durational contrast between successive vowels in English vs. French speech was reflected in greater durational contrast between successive notes in English vs. French instrumental music. But languages differ not only in their rhythmic prosody (temporal and accentual patterning) but also in the way voice pitch rises and falls across phrases and sentences ("intonation," which refers to speech melody rather than speech rhythm), and, recently, it was found that pitch intervals in the music of a given culture reflect the intonation of its language. In sum, instrumental music does indeed reflect the prosody of the national language, thus showing that the language we speak may influence the music our culture produces (Patel, Iversen & Rosenberg, 2006).

Evidences against modularity II: syntax in music and language

Music, like language, is an acoustically based form of communication with a set of rules for combining limited number of perceptual discrete acoustic elements (pitches in music, and phonemes in language) in an infinite number of ways (Hauser, Chomsky & Fitch, 2002). The set of principles governing the combination of discrete structural elements (such as phonemes and words or musical tones) into sequences is called syntax. Brain responses reflecting the processing of musical chord-sequences are similar, although not identical, to the brain activity elicited during the perception of language, in both musicians and nonmusicians (Koelsch et al., 2002; Patel et al., 1998).

The sharing between music and language of the cognitive mechanisms involved on abstraction of rules present in syntax is striking. Recent studies reveal brain responses related with music-syntactic processing that were localized in Broca's area and its right homologue (Dapretto & Bookheimer, 1999; Friederici, Meyer & von Cramon, 2000), thus suggesting that the mechanisms underlying syntactic processing are shared between music and language (Levitin & Menon, 2003; Patel, 2003a). Detection of musical deviations lead to bilateral activation not only in frontal regions comprising anterior-superior insular cortex and inferior fronto-lateral areas involved on language syntax but also in temporal regions including Wernicke's area (BA22, 22p) and PT (BA42) (Koelsch et al., 2002), areas known to support auditory language processing. These indicate that sequential auditory information is processed by brain structures which are less domain-specific than previously believed. Interestingly, earlier imaging results are consistent with recent neuropsychological data on language processing. Recent studies investigating cognitive deficits in agrammatic aphasia (Dominey & Lelekov, 2000; Lelekov et al., 2000; Keurs, Brown, Hagoort & Stegeman., 1999) in conjunction with healthy subject data (Lelekov-Boissard & Dominey, 2002) also suggest that the role of this inferior frontal region in the language syntax are more related to the processing of supramodal, linguistic and non-linguistic, abstract rules. For instance, agrammatic patients also display non-linguistic deficits such as the processing of sequence of letters such as 'ABCBAC' and 'DEFEDF', which have different serial order or serial structure but share the same abstract structure '123213'; Broca's aphasic patients with agrammatism did not show a specific deficit in grammatical, closed-class words but, instead, have a working memory related deficit reflected by delayed and/or incomplete availability of general word-class information (Keurs et al., 1999).

These data are not fully surprising if we take into account that linguistic and musical sequences are created by combinatorial principles operating at multiple levels, such as in the formation of words, phrases and sentences in language, and of chords, chord progressions and keys in music (i.e., 'harmonic structure), making us to perceive in terms of hierarchical relations that convey organized patterns of meaning. In language, one meaning supported by syntax is 'who did what to whom', that is the conceptual structure of reference and predication in sentences. In music, one meaning supported by syntax is the pattern of tension and resolution experienced as the music unfolds in time (Patel, 2003a).

Andrade, Andrade, Gaab, Gardiner & Zuk (2010) aimed at investigating the relations between music and language at the level of "patterned sequence processing" in a novel music task called Musical Sequence Transcription Task (MSTT). To control for fine-grained pitch processing we asked to forty-three seven years old brazilian students to listen to four-sound sequences based on the combination of only two different sound types: one in the low register (thick sound), corresponding to a perfect fifth with fundamental frequencies 110Hz (A) and 165 Hz (E), and the other in the high register of (thin sound), corresponding to a perfect fourth, with 330Hz (E) and 440Hz (A). Children were required to mark the thin sound with a vertical line '|' and the thick sound with an 'O', but were never told that the sequences only consist of four sounds. Performances on MSTT task were positively correlated with phonological processing and literacy skills.

These observations are also in line with recent findings in both human newborns and non-human primates (tamarins) which are capable to discriminate non-native languages that pertain to different classes of rhythmic prosody, but do not discriminate languages from the same rhythmic class (Ramus et al., 2000; Tincoff et al., 2005). These lead to a more general proposal that infants build grammatical knowledge of the ambient language by means of 'prosodic bootstrapping', with rhythm playing one of the central roles; it is consistent with the frequent co-occurrence of agrammatism and rhythm deficits in aphasic patients with left prefrontal damage (Efron, 1963; Mavlov, 1980; Swisher & Hirsh, 1972), and provides further support for the link between music and language in the temporal and sequencing domain, hence, for shared cognitive mechanisms on syntactical processing.

 

Music cognition and cross-modal processing

While discussing the sharing of neurocognitive mechanisms between music and language, jointly with the fact that music is embodied and polysemic, we automatically adopt a cross-modal approach of music cognition. As we have already pointed out that music is, in fact, as much about action as it is about perception. Simply, music 'moves' us. When one plays music, or tap, dance, or sing along with music, the sensory experience of musical patterns is intimately coupled with action.

Regarding the neuroimage data to infer eventual overlaps between musical processing and other domains it is important to keep in mind that similar brain activation patterns alone are not necessarily accepted as overlapping because distinct but intermingled neural pathways very close to each other can be seen as the same activations, reflecting the resolution limits in standard 3 X 3 X 4 mm voxels of the fMRI method. Therefore, our comparisons of neuroimaging data used to argue for overlapping between music and other domains were based mainly, but not only, on the criteria adopted in the meta-analysis on auditory spatial processing conducted by Arnott, Binns, Grady & Alain (2004). The authors analyzed data from peer-reviewed articles of auditory studies, including 27 studies on nonspatial sounds and 11 on spatial sounds, that incorporated either PET or fMRI recordings with normal, healthy adults. We then compared activations of many neuroimaging studies on the pitch pattern perception in music with the reported activations under spatial pitch processing in the study by Arnott et al. (2004) often within ~ 4 mm cubic. This same criterion was applied in the comparisons between music and language. Furthermore, when we put together both imaging and lesion data, i.e. showing the same overlap from either brain activations in the neuroimaging of healthy subjects and the patterns of deficit and spared cognitive function in lesion data, we interpret this evidence as a strong support for crossmodal processing.

There has been a significant overlap between neural substrates for the generation of visuo-spatial tasks and music perception. For example, occipital and frontoparietal cortical areas, traditionally involved in visuo-spatial tasks, including the precuneus in medial parietal lobes called the "mind's eye" for being crucially involved in generation of visuo-spatial imagery (Mellet et al., 2002; Mellet et al., 1996), are among the most frequently and significantly activated regions in perception of music either in naïve listeners or in musicians (Halpern & Zatorre, 1999). These imaging data are consistent with three well documented cases of amusia associated to deficits in spatial-temporal tasks in both visual and auditory modalities (Griffiths et al., 1997; Steinke et al., 2001; Wilson & Pressing, 1999), as well as associated deficits in pith perception and visuo-spatial tasks in left prefrontal damage patients (Harrington et al., 1998). Bilateral frontal and inferior frontal activations, mainly premotor frontal areas BA6, dorsolateral prefrontal areas (BA8/9), as well as inferior frontal areas as Broca (BA44/45), insula, and more anterior middle and inferior frontal cortices (46/47), are frequently observed in either non-musicians (Platel et al., 1997; Zatorre et al., 1994) or in musicians (Ohnishi et al., 2001; Parsons, 2001; Zatorre et al., 1998). Activations of premotor-posterior parietal network in pitch tasks are not surprising if we take into account that. functional neuroimaging studies in humans suggest premotor-posterior parietal networks involvement not only in motor control but also for cognitive processes such as mental object-construction task (Mellet et al., 1996) and other visuo-spatial tasks involving spatial working memory either implicitly (Haxby et al., 1994) or explicitly (Smith, Jonides & Koeppe, 1996).

When non-musicians were asked to tell if the second of two melodies presents a change or not, all activation foci were in the left hemisphere with the most significant activations were in the left precuneus/cuneus (BA 18/19/31), a region traditionally involved in visual imagery tasks (Platel et al., 1997). It could be argued that the analytic pitch tasks tap into visual imagery in terms of 'high' and 'low' in relation to a notational mental base line or a mental stave, what was suggested by the subjects themselves during debriefing. Left superior parietal lobes activations, particularly in the parieto-occipital junction, was activated when professional pianists were reading musical scores without listening or performing (Sergent, Zuck, Terriah & Macdonald, 1992), and this was interpreted as the participation of the dorsal visual system in spatial processing for reading scores, once musical notation is based on the spatial distribution and the relative height separation of the notes on the scores, forming 'pitch intervals'. Other imaging studies have frequently reported superior parietal lobes (Zatorre et al., 1998) and precuneus activation also in perceptual tasks (Satoh et al., 2001). Halpern & Zatorre (1999) found that musicians when asked to mentally reproduce a just heard unfamiliar melody yield significant activations in several left inferior prefrontal (BA 44, 47, 10) and middle/superior frontal gyrus (BA 8, 6), as well as in premotor (BA6), visual association cortex (BA19), and parietal cortex and precuneus (BA40/7), but no activations in temporal lobes were found. Musicians with both absolute pitch (AP) and relative pitch (RP) both showed bilateral activations of parietal cortices in judging a musical interval, but with stronger left parietal activations in RP musicians whose cognitive strategies rely more on the maintaneance of pitch information on auditory working memory for pitch comparison on a mental stave (Zatorre et al., 1998), whereas AP musicians rely more on long-term memory (Zatorre et al., 1998). Finally, an important revelation of imaging studies in music perception and production are the significant activations of the cerebellum (Parsons, 2001; Zatorre et al., 1994) in either pitch or rhythm tasks.

As pointed by Janata & Grafton (2003), three research domains in psychology and neuroscience are particularly pertinent to understanding the neural basis for sequencing behaviors in music: they are timing, attention and sequence learning. In fact, music can be thought of as a sequence of events that are patterned in time and a 'feature space' that is multidimensional and consists of both motor and sensory information. Motor patterns determine how we position effectors in space, such as fingers on piano keys, whereas sensory patterns reflect the organization of auditory objects, such as notes or sets of notes in time when we hear melodies and chords, respectively. Temporal and spatial sequence production in humans yield activations in many brain regions including the cerebellum, SMA, PMC, basal ganglia and parietal cortex, which are richly interconnected. There possibly lies a core circuit underlying sequenced behaviors in which different levels of complexity show considerable overlap in some regions such as the SMC and SMA, whereas higher levels of complexity are related with overlap and greater activations in the cerebellum, basal ganglia, thalamus, PMC (premotor cortex), VLPFC (ventrolateral prefrontal cortex), IPS (intraparietal sulcus) and precuneus. All these regions have also been related with music processing (Janata & Grafton, 2003).

Given that music inherently consists of sequential auditory and motor information, it is quite expected to find shared cognitive mechanisms and neural networks between music and sequencing behavior.

 

Universalities in music

Infant research and cross-cultural comparisons are the two main approaches to studying the contribution of universal and cultural cues in music perception. Infant research helps to clarify the culture-transcendent predispositions for processing music and to describe how the culture-specific properties develop (for a review see (Trehub, 2003)). Cross-cultural investigations compare the responses of listeners from culturally separate musical backgrounds searching for culture-transcendent and culture-specific factors in music. The topics in cross-cultural studies of music have ranged from the emotional responses to music (Balkwill, Thompson & Matsunaga, 2004), to melodic expectancies (Castellano, Bharucha & Krumhansl, 1984; Krumhansl et al., 2000), to pulse-finding (Toiviainen & Eerola, 2003), and to interval discrimination (Burns, 1999).

There is evidence of common features, across different music styles, in the principles underlying melody (Krumhansl & Toiviainen, 2001) and in response to features such as consonant/ dissonant intervals, pitch in musical scales and meter (Drake, 1998; Drake & Bertrand, 2001). The way an infant processes musical patterns is similar to that of adults (Trainor & Trehub, 1992); infants respond better to melodic as well as to harmonic consonant patterns, and to complex metric rhythms (Trainor, McDonald & Alain, 2002a; Zentner & Kagan, 1996). There is reason to believe that infants possess absolute pitch early in life but change to relative pitch later (Saffran & Griepentrog, 2001), and that they have long-term musical memory (Saffran, Loman & Robertson, 2000). The widespread use of scales of unequal steps, comprised of 5 to 7 notes based around the octave (the most consonant interval corresponding to 12 semitones and formed by the first and the eighth final note of the scale), is another universality of music which is preferred even by infants when compared to scales of equal steps between notes (Burns, 1999; Trehub, Schellenberg & Kamenetsky, 1999; Trehub, Thorpe & Trainor, 1990).

Expectations formed by the listeners when listening to music are also universal across cultures. Expectation is a basic component of music perception, operating on a variety of levels, including melodic, harmonic, metrical, and rhythmic, and it addresses the question "what" and "when", that is, what tones or chords are expected to occur and when, in a given musical sequence. It is not only presumed to play an important role in how listeners group the sounded events into coherent patterns, but also to appreciate patterns of tension and relaxation contributing to music's emotional effects. Both cognitive and emotional responses to music depend on whether, when, and how the expectations are fulfilled. The main method used in cross-cultural comparisons of musical expectations is the probe tone task, first developed by Krumhansl and Shepard (1979) for quantifying the perceived hierarchy of stability of tones.

In experiments on melodic expectations a melody is presented to the listeners many times, but followed on each occasion by a single probe-tone with varying degree of fitness. Using a rating scale, the listeners assess the degree to which the probe-tone fits their expectations about how the melody might continue (Krumhansl et al., 2000). Cross-cultural studies comparing the fitness ratings given by Indian and Western listeners to North Indian ragas (Castellano et al., 1984), and by Western and Balinese listeners to Balinese music (Kessler, Hansen & Shepard, 1984), as well as native Chinese and American listeners' responses to Chinese and British folk songs (Krumhansl, 1995) found strong agreement between listeners from these different musical cultures. A remarkable consistency is found between (South Asian) Indian and Western listeners (Castellano et al., 1984) in giving higher ratings to the tonic and the fifth degree of the scale, tones that were also sounded most frequently and for the longest total duration, which, theoretically, are predicted to be important structural tones in Indian ragas. Interestingly, the western listeners' responses did not correspond to tonal hierarchies of major and minor keys, according to the Theory of Harmony in Western music (Krumhansl, 1990), but rather to theoretical predictions for Indian music. Additionally, it was shown that all listeners can appreciate some aspects of different musical cultures by attending to statistical properties of the music, such as the number of times that different tones and tone combinations appear in the presented musical contexts, although there were some residual effects of expertise depending on the listeners' familiarity with the particular musical culture (Castellano et al., 1984). Recent cross-cultural works on melodic expectancies, with Korean music (Nam, 1998) and music of indigenous people of the Scandinavian Peninsula (Eerola, 2004; Krumhansl et al., 2000), have provided additional evidences for the universal reliance on the hierarchical arrangement of pitches, indicating that music draws on common psychological principles of expectation even if musical cultures have a distinct effect in these principles, although the exact way it is accomplished varies with culture.

Listeners ranging from experts to completely unfamiliar, namely Sami yoikers (experts), Finnish music students (semi-experts) and Central European music students (Krumhansl et al., 2000) as well as traditional healers from South Africa (non-experts) (Eerola, 2004) rated the fitness of probe-tones as continuations of North Sami yoik excerpts, a musical style from indigenous people (Sami people sometimes referred to as Lapps) of the Scandinavian Peninsula that is quite distinct from Western tonal music (Krumhansl et al., 2000; Eerola, 2004). All groups, consistently with previous studies, showed strong agreement (Castellano et al., 1984; Kessler et al., 1984; Krumhansl, 1995). Further support was also provided by the event-frequency models (statistical properties of tone distributions) according to which listeners seem to extract the frequently occurring events in the melody which act as cognitive reference points that are established by repetition (Eerola, 2004). In fact, sensitivity to the statistical properties of music also influences listeners's expectations (Oram & Cuddy, 1995; Krumhansl et al., 2000), and is apparently a universal process, exhibited by even 8-month-old infants (Saffran, Johnson, Aslin & Newport, 1999).

Today is generally agreed that in addition to cultural cues, listeners' possess innate general psychological principles in auditory perception such as sensitivity to consonance, interval proximity, and, finally, the statistical properties, which have been extensively shown to influence listeners' expectations (Oram & Cuddy, 1995; Krumhansl et al., 2000). Early on development, consonance is preferred even by 2-4 months old infants in relation to dissonance (Trainor, Tsang & Cheung, 2002b) and sensitivity to statistical properties of music are apparently a universal process exhibited by even 8-month-old infants (Saffran et al., 1999, Trainor et al., 2002a).

Finally, in the temporal organization of auditory sequences two universal processes have been identified (Drake, 1998): (i) segmentation sequence into groups of events, accomplished in terms of changes in pitch durations and salience, and already present in early infancy (Krumhansl & Juscsyk, 1990), and (ii) the extraction of an underlying pulse through the perception of regular intervals (Drake & Baruch, 1995), which allows to infer the steady pulse (meter) from patterns of temporal grouping (rhythm) and categorize unique rhythmic and melodic patterns, a cross-cultural ability present in adults (Toiviainen & Eerola, 2003), and also in infants (Hannon & Johnson, 2005).

 

Music and emotions

As we have pointed above, the emotional response to music depends on whether, when and how the expectations are fulfilled (for a review see (Eerola, 2004)), and music's extraordinary ability to evoke powerful emotions is likely the main reason why we listen to music and why music is generally referred to as the "language of emotions".

Consistent with results on musical expectancies (Eerola, 2004), recent empirical findings suggest that emotional meaning in music is conveyed by universal acoustic cues, such as complexity, loudness, tempo, whose interpretation does not require familiarity with culture-specific musical conventions (Balkwill et al., 2004). Cross-culturally, increases in perceived complexity appear to be consistently associated with the negative emotions of anger and sadness, while decreases are associated with joy and happiness. Further, increases in perceived tempo are associated with joy, whereas decreases in tempo are associated with sadness. Finally, increases in loudness are frequently associated with anger (Balkwill et al., 2004), and create displeasure even in nonhuman primates (McDermott & Hauser, 2004). Consonance and dissonance certainly play a role in the musical emotions related to positive and negative affective values, respectively (Peretz et al., 2001; Blood & Zatorre, 2001; Trainor et al., 2002b).

Music is capable of inducing strong emotions with both positive and negative emotional valence consistently across subjects (Krumhansl, 1997) and cultures (Balkwill et al., 2004; Eerola, 2004). For example, Jaak Panksepp (1995) asked several hundred young men and women why they felt music to be important in their lives and 70% of both sexes said it was "because it elicits emotions and feelings", in the second place came "To alleviate boredom".

Today, we already know that music elicits emotions rather than merely express an emotion that the listener recognizes. Most people experience a particularly intense, euphoric response to music, frequently accompaniment by an autonomic or psychophysiological component, described as ''shivers-down-the-spine'' or ''chills''. Listening to music automatically elicits physiological changes in blood circulation, respiration, skin conductivity and body temperature, which are autonomic responses of the sympathetic nervous system (Krumhansl, 1997). Khalfa, Isabelle, Jean-Pierre & Manon (2002) demonstrated that SCRs (electrodermal activity measured by Skin Conductance Response to measure the stimulus arousal) can be evoked and modulated by musical emotional arousal. Similarities in the event-related measures of musical and visual emotions suggests that the processing of musical emotions may not be domain-specific, that is, it may not be different in nature from the emotions elicited by other events such as words and pictures (Khalfa et al., 2002).

Imaging and lesion studies reveal the deep subcortical foundations of emotional musical experiences in many brain areas that are homologous between humans and all of the other mammals (Blood & Zatorre, 2001; Brown, Martinez & Parsons, 2004; Gosselin et al., 2005). Pleasant/consonant music directly affects paralimbic areas such as insula, orbitofrontal cortex and ventromedial prefrontal cortex associated with reward/motivation, emotion, and arousal, as well as the mesolimbic system, the most important reward pathway (Blood & Zatorre, 2001; Blood, Zatorre, Bermudez & Evans, 1999; Menon & Levitin, 2005). In contrast, umpleasent/dissonant music directly affects the parahippocampal gyrus (Blood et al., 1999), a paralimbic structure activated during unpleasant emotional states evoked by pictures with negative emotional valence jointly with amygdale (Lane et al., 1997), which, in turn, is a key structure in fear processing and with which parahippocampal gyrus has strong reciprocal connections (Bechara, Damasio & Damasio, 2003; Mesulam, 1998). Recently it was demonstrated that unilateral damage to amygdale selectively impairs the perception of emotional expression of fear in scary music (minor chords on the third and sixth degrees and fast tempos), while recognition of happiness (major mode and fast tempo) was normal, and recognition of peacefulness (major mode and intermediate tempo played with pedal and arpeggio accompaniment) and sadness (minor mode at an average slow tempo) in music was less clearly affected by the medial temporal lobe resection (Gosselin et al., 2005).

There are also evidences for the double dissociation between musical cognition and emotional processing. Bilateral damage to the auditory cortex caused severe, irreversible deficits in musical expression, as well as perception on both pitch and temporal dimensions, such as tasks requiring `same-different' discrimination for musical sequences, including consonant and dissonant musical material (Peretz et al., 2001), yet the patients could appreciate and enjoy the affective/emotional meaning (Peretz et al., 1997) and classify melodies as 'happy' and 'sad' as normal controls (Peretz & Gagnon, 1999). Conversely, loss of emotional response to music in presence of completely preserved musical perception abilities is reported in a patient with infarction in the left insula, extending anteriorly into the left frontal lobe and inferiorly into the left amygdale (Griffiths et al., 2004), all areas with a close correspondence with left hemisphere areas activated during emotional response to music in normal subjects (Blood & Zatorre, 2001). Both studies offer an important double dissociation between musical cognition and emotional processing, i.e., between the perceptual and emotional components of music processing, the first involving the superior temporal and inferior frontal and parietal areas, and the second engaging a distributed brain network that is also recruited by other powerful emotional stimuli that produce autonomic arousal and includes bilateral medial limbic structures, insula, ventral striatum, thalamus, midbrain, and widespread neocortical regions.

In sum, there are increasing evidences that music derives its affective charge directly from dynamic aspects of brain systems that normally control real emotions and which are distinct from, albeit highly interactive with, cognitive processes. The ability of music to stir our emotions rather directly is a compelling line of evidence for the idea that cognitive attributions in humans are not exclusively needed for arousing emotional processes within the brain/ mind. In other words, cognitive processing to decode cultural conventions is not absolutely essential to extract the emotional meaning in music, a notion supported by cognitive, cross-cultural studies, as well as by lesion and neuroimaging studies.

The link between music and the deep subcortical structures of emotion in the human brain is reinforced by the relation between music emotions and the neurochemistries of specific emotions within the mammalian brain. Panksepp predicted that opiate receptor antagonists such as naloxone and naltrexone can diminish our emotional appreciation of music (Panksepp, 1995). Indeed, preliminary data provided by Goldstein (1980) have long indicated that such manipulations can reduce the chills or thrills that many people experience when they listen to especially 'moving' music.

 

Discussion: role of music in development

Contrary to the belief that music is a mere cultural byproduct and its appreciation necessarily involves the appropriation of cultural conventions, the growing body of scientific research reveals that, in addition to the observed universal cognitive processes involved in music perception, the music could not be a meaningful and desired experience without the ancestral emotional systems of our brains. Besides, although the existence of some crucial and particularly music-relevant neural networks in the right hemispheres, the neurocognitive processes involved in music and language do not appear to be music or language specific, neither encapsulated. Music, particularly, has been proved to be a supramodal cognitive domain involving cortical areas traditionally involved in language, spatial and motor processing, as well as subcortical areas involved in basic emotional processing.

Emotionally speaking, unlike many other stimuli, music can often evoke emotion spontaneously, in the absence of external associations and our ability to decipher the neural underpinnings of aesthetic responses may turn out to be easier for music than for the visual arts (Ramachandran & Hirstein, 1999), and, like biologically relevant visual stimuli, activates primitive structures in the limbic and paralimbic areas related to fear and reward, including waking arousal control systems such as those based on norepinephrine (NE) and serotonin that regulate emotional responses, as well as generalized incentive seeking system centered on mesolimbic and mesocortical dopamine (DA) circuits, which are important also in the intra-cerebral estimation of the passage of time and likely related to rhythmic movements of the body which may be ancestral pre-adaptation for the emotional components of music (Panksepp & Bernatzky, 2002). In this sense, well constructed music is uniquely efficacious in resonating with our basic emotional systems, bringing to life many affective proclivities that may be encoded, as birthrights, within ancient neural circuits constructed by our genes, many of which we share homologously with other mammals (Panksepp & Bernatzky, 2002). Therefore, ultimately, our love of music possibly reflects the ancestral ability of our mammalian brain to transmit and receive basic emotional sounds that can arouse affective feelings which are implicit indicators of evolutionary fitness.

Conversely, cognitively speaking, music, unlike many other stimuli, is abysmally supramodal, involving auditory, motor, spatial, numerical and emotional and linguistic neural networks.

The central question here is "Why"? The data reviewed here appear to have two important implications. First, that music may be based on the existence of the intrinsic emotional sounds we make (the animalian prosodic elements of our utterances) and on the rhythmic movements of our instinctual/emotional motor apparatus, both evolutionarily designed to index whether certain states of being were likely to promote or hinder our well-being (Panksepp & Bernatzky, 2002). Second, that these emotional and supramodal characteristics of music are also evolutionary important to promote stimulation and development of several cognitive capabilities (Cross, 2001, 2003a).

We have seen that crucial aspects of pitch perception in music are likely shaped by relatively basic auditory processing mechanisms that are not music specific and can be studied in experimental animals (Hauser & McDermott, 2003). In social creatures like ourselves, whose ancestors lived in arboreal environments where sound was one of the most effective ways to coordinate cohesive group activities, reinforce social bonds, resolve animosities, and to establish stable hierarchies of submission and dominance, there should have been a premium on being able to communicate shades of emotional meaning by the melodic character (prosody) of emitted sounds (Panksepp & Bernatzky, 2002). Therefore, sound is an excellent way to help synchronize and regulate emotions so as to sustain social harmony, and we can think that the capacity to generate and decode emotional sounds should be excellent tools for survival. Our brains' ability to resonate emotionally with music may have a deep multidimensional evolutionary history, including issues related to the emergence of intersubjective communication, mate selection strategies, and the emergence of regulatory processes for other social dynamics such as those entailed in group cohesion and activity coordination.

The principles of evolutionary psychology proposes that the brains of humans and other species are adapted to process and respond to three general classes of information: social, biological, and physical. The higher the species in the evolutionary hierarchy, the more important are the psychological traits related to social behavior. Theory of mind is especially salient in humans and represents the ability to make inferences about the intentions, beliefs, emotional states, and likely future behavior of other individuals (Meltzoff, 2002), and plays a central role on all socio-cognitive competences. From a comparative perspective, language and theory of mind are the most highly developed sociocognitive competencies in humans (Pinker, 1999) and are hypothesized to be the evolutionary result of the complexity of human social dynamics. We propose that the links between music and imitation are striking. Let's begin with a better notion of the principles underlying imitation.

Striking convergences among the cognitive and neuroscientific findings indicate that imitation is innate in humans and precedes mentalizing and theory of mind (in development and evolution), and, finally that behavioral imitation and its neural substrate provide the mechanism by which theory of mind and empathy develop in humans. It is argued that in ontogeny, infant's imitation appears to be the seed of the adult's theory of mind. Newborns as young as 42 minutes old are capable of successful facial imitation. Further, small infants with 12-21 day-old could imitate four different adult gestures: lip protrusion, mouth opening, tongue protrusion and finger movement and confused neither actions nor body parts (Meltzoff, 2002). Infants can link self and other through what was termed a 'supramodal' representation of the observed body act. Early imitation appears to provide the first instance of infants making a connection between the world of others and the infants' own internal states, the way they feel themselves to be. Infants recognize when another acts 'like me' and the high emotional value of this experience. Further, infants look longer at the adult who was imitating them; smiled more at this adult; and most significantly, directed testing behavior at that adult (Meltzoff & Decety, 2003).

Music comes into the scene when we take into account that, due to its supramodal nature, including emotion, music perception appears to be an innate behavior that stimulates, among others things, the supramodal representation underlying imitation. A common observation in the social-developmental literature is that parent-infant games are often reciprocally imitative in nature. When parents shake a rattle or vocalize, infants shake or vocalize back. A turn-taking aspect of these games is the "rhythmic dance" between mother and child. Imitation is often bidirectional because parents mimic their infants as well as infants imitate parents (Meltzoff, 1999). Adults across cultures play reciprocal imitative games with their children that embody the temporal turn-taking (Trevarthen, 1979; Trevarthen, 1999). The links between protomusical behaviors in infants and imitation are becoming increasingly evident. We know that imitation games with music and dance are universal, and the tribal dances itself can be seen as one of the most frequent forms of imitation game, used to develop the in-group sense, the feeling of both "being like the other" and "the other is like me", and thus pertaining to a group. Perhaps not coincidentally, the developmental precursors of music in infancy, and through early childhood, occur in the form of proto-musical behaviours which are exploratory and kinesthetically embedded, being closely bound to vocal play and to whole body movement.

Human infants interact with their caregivers producing and responding to patterns of sound and action. These temporally-controlled interactions involve synchrony and turntaking, and are employed in the modulation and regulation of affective state (Dissanayake, 2000) and in the achievement and control of joint attention also referred as to 'primary intersubjectivity' (Trevarthen, 1999). This rhythmicity is also a manifestation of a fundamental musical competence, and "...musicality is part of a natural drive in human sociocultural learning which begins in infancy" (Trevarthen, 1999, p. 194) and also allows infants to follow and respond in kind to temporal regularities in vocalization, movement, and time, to initiate temporally regular sets of vocalisations and movements (Trevarthen, 1999). It makes music inextricably involved in the acquisition of language. Protomusical behaviors are so intimately intertwined with protoverbal behavior that "preverbal origins of musical skills cannot easily be differentiated from the prelinguistic stages of speech acquisition and from the basic alphabet of emotional communication" (Papousek, 1996a). In the pitch dimension, infant-caregiver interactions tend to exhibit the same range of characteristics such as exaggerated pitch contours on the caregiver's part ('motherese'), and melodic modulation and primitive articulation on the infant's part, all in the context of an interactive and kinaesthetic rhythmicity. On the part of the infant, these activities develop into exploratory vocal play (between 4 and 6 months) which gives rise to repetitive babbling (from 7 to 11 months) from which emerges both variegated babbling and early words (between 9 and 13 months) (Kuhl et al., 2001; Papousek, 1996a, 1996b). Linguistic behavior begins to differentiate from infant proto-musical/proto-linguistic capacities as the infant develops, when the parent/infant interactions increasingly make use of vocalizations and gestures to communicate affect in the exchange of 'requests' further supporting the development of referential gestures and vocalizations, orienting the attention of both participants to objects, events and persons outside the parent-infant dyad (Cross, 2001).

The biological role of musical behavior appears to fit not only with the development of socio-affective capacities and language, but also with the cognitive flexibility of human infants. Music can be viewed as a kind of multimodal or multimedia activity of temporally patterned movements. The polysemic nature of music and its characteristic of being a consequence-free means of exploring social interaction and a "play-space" for rehearsing processes appear to provide an ideal way of achieve cognitive flexibility, beyond its role in the socialization of the child. The role for music in the development of the child's individual cognitive capacities appears to be well supported by one of the main characteristic of neural processing in music, cross-modality. In fact, to listen and to make music appear to be a perfect means of forming connections and inter-relations between different domains of infant and childhood competence such as the social, biological and mechanical.

 

Conclusion

Now we can say that music is not only polysemic by also polymodal. Music is universal among human cultures, extant or extinct, and "binds us in a collective identity as members of nations, religions, and other groups" (Tramo et al., 2001).

Although at least one component of music, the tonal encoding of pitch, appears indeed to be domain-specific it results from fine-grained pitch perception, a mechanism that is music- relevant but not domain-specific in a strict fodorian sense. The evolved behaviors such as music and language, spatial processing, indeed share many common neurocognitive modules, whereas the selectivity observed in lesion studies appears to be to a result of the differential profile of relevance of these set of shared neural modules for each domain. In fact, the deficits are not completely selective, the apparently selective deficits in pitch processing in music, for instance, can be accompanied by deficits in spatial processing and language, they just appear to be more detrimental to music compared to other cognitive domains. On the other hand, the rhythm processing in music is clearly linked with language processing as shown by lesion and imaging studies.

In fact the universality of music jointly with its main characteristic, cross-modality, suggest that different cultural manifestations of music are cultural particularizations of the human capacity to form multiply intentional representations through integrating information across different functional domains of temporally extended or sequenced human experience and behaviour, generally expressed in sound. We differentiate from our relatives by an immense leap in cognitive flexibility and in our capacity to enter into and sustain a wide range of social relationships and interactions. The polysemic nature and cross-modality of musical behaviour can provide for the child a space within which she can explore the possible bindings of these multidomain representations.

The scientific research on music help us not only to understand the marvelous and mysterious functioning of our brain, but also shed light on the essence of human behavior, and hence the humankind.

 

Acknowledgements

We are thankful to Prof. Joydeep Bhattacharya who helped to improve this text considerably by providing valuable comments on an earlier version of this article.

 

References

Andrade, P.E. & Bhattacharya, J. (2003). Brain tuned to music. J. Royal Soc. Med., 96 (8), 284-287.         [ Links ]

Andrade, P.E.; Andrade, O.V.C.A.; Gaab, N.; Gardiner M. F. & Zuk J. (2010) Investigando a Relação entre Percepção Musica e Habilidades Cognitivo-Linguísticas em Leitores Iniciantes da Língua Portuguesa. Resumos/Abstracts, XXXIV Congresso Anual da Sociedade Brasileira de Neurociências e Comportamento. Retirado em 17 de dezembro de http://www.sbnec.org.br/congresso2010/resumos/R0364-1.html.

Anvari, S.H.; Trainor, L.J.; Woodside, J. & Levy, B. A. (2002). Relations among musical skills, phonological processing, and early reading ability in preschool children. J. Exp. Child Psychol., 83 (2), 111-130.

Arnott, S.R.; Binns, M.A.; Grady, C.L. & Alain, C. (2004). Assessing the auditory dual-pathway model in humans. Neuroimage, 22 (1), 401-408.         [ Links ]

Ayotte, J.; Peretz, I. & Hyde, K. (2002). Congenital amusia - A group study of adults afflicted with a music-specific disorder. Brain, 125 (2), 238-251.         [ Links ]

Ayotte, J.; Peretz, I.; Rousseau, I.; Bard, C. & Bojanowski, M. (2000). Patterns of music agnosia associated with middle cerebral artery infarcts. Brain, 123 (9), 1926-1938.         [ Links ]

Balkwill, L-L.; Thompson, W.F. & Matsunaga, R. (2004). Recognition of emotion in Japanese, Western, and Hindustani music by Japanese listeners. Jap. Psychological Res., 46 (4), 337-349.         [ Links ]

Barrett, H.C. & Kurzban, R. (2006). Modularity in cognition: framing the debate. Psychol. Rev, 113 (3):628-47.

Bechara, A.; Damasio, H. & Damasio, A.R. (2003). Role of the amygdala in decision-making. Annual New York Academy of Science, 985, 356-69.         [ Links ]

Bickerton, D. (1990). Species and Language. Chicago: Univ. of Chicago Press.         [ Links ]

Bishop, D.V. (1999). Perspectives: cognition. An innate basis for language? Science, 286 (5448), 2283-2284.         [ Links ]

Blood, A.J. & Zatorre, R.J. (2001). Intensely pleasurable responses to music correlate with activity in brain regions implicated in reward and emotion. Proceedings of the National Academy of Sciences of the United States of America, 98 (20), 11818-11823.         [ Links ]

Blood, A.J.; Zatorre, R.J.; Bermudez, P. & Evans, A. C. (1999). Emotional responses to pleasant and unpleasant music correlate with activity in paralimbic brain regions. Nature Neuroscience 2 (4), 382-387.         [ Links ]

Bookheimer, S. (2002). Functional MRI of language: new approaches to understanding the cortical organization of semantic processing. Annual Review of Neuroscience, 25, 151-88.         [ Links ]

Brown, S.; Martinez, M.J. & Parsons, L.M. (2004). Passive music listening spontaneously engages limbic and paralimbic systems. Neuroreport, 15 (13), 2033-7.         [ Links ]

Brust, J. C. (1980). Music and language: musical alexia and agraphia. Brain, 103 (2), 367-92.         [ Links ]

Burns, E.M. (1999). Intervals, scales, and tuning. In: Deutsch, D. (Ed.). In The Psychology of Music (pp. 215-264). San Diego: Academic Press.         [ Links ]

Castellano, M.A.; Bharucha, J.J. & Krumhansl, C.L. (1984). Tonal hierarchies in the music of north India. J. Exp. Psychol. Gen., 113 (3), 394-412.         [ Links ]

Chomsky, N. (1957). Syntactic Structures. The Hague, Netherlands: Mouton & Co.

Cross, I. (2001). Music, cognition, culture, and evolution. Annual New York Academy of Science, 930, 28-42.         [ Links ]

Cross, I. (2003a). Music and evolution: causes and consequences. Contemporary Music Review, 22, 79-89.

Cross, I. (2003b). Music as a biocultural phenomenon. Annual New York Academy of Science, 999, 106-111.

Dalla Bella, S. & Peretz, I. (1999). Music agnosias: Selective impairments of music recognition after brain damage. Journal of New Music Research, 28 (3), 209-216.         [ Links ]

Dapretto, M. & Bookheimer, S.Y. (1999). Form and content: dissociating syntax and semantics in sentence comprehension. Neuron, 24 (2), 427-32.         [ Links ]

Dehaene, S.; Dehaene-Lambertz, G. & Cohen, L. (1998). Abstract representations of numbers in the animal and human brain. Trends Neurosci., 21 (8), 355-61.         [ Links ]

Dissanayake, E.I.E. (2000). Antecedents of the temporal arts in early mother-infant interactions. In: Wallin, N.; Merker, B. & Brown, S. (Ed). The origins of music (pp. 389-410). Cambridge, MA: MIT Press.

Dominey, P.F. & Lelekov, T. (2000). Non-linguistic transformation processing in agrammatic aphasia. Comment on Grodzinsky: The neurology of syntax: Language use with and without Broca's area. Behavioral and Brain Sciences, 23 (1), 34.         [ Links ]

Drake, C. (1998). Psychological processes involved in the temporal organization of complex auditory sequences: Universal and acquired processes. Music Perception, 16, 11-26.         [ Links ]

Drake, C. & Baruch, C. (1995). From temporal sensitivity do models of temporal organization: Hypotheses and data on the acquisition of the temporal abilities. Année Psychologique, 95, 555-569.         [ Links ]

Drake, C. & Bertrand, D. (2001). The quest for universals in temporal processing in music. Annual New York Academy of Science, 17-27.

Dunbar, R. (1996). Grooming, Gossip and the Evolution of Language. Cambridge, MA: Harvard University Press.

Eerola, T. (2004). Data-driven influences on melodic expectancy: Continuations in North Sami Yoiks rated by South African traditional healers. In: Lipscomb, S.D.; Ashley, R.; Gjerdingen, R.O. & Webster, P. (Eds.), Proceedings ofthe Eighth International Conference of Music Perception andCognition (pp. 83-87). Adelaide, Australia: Causal Productions.         [ Links ]

Efron, R. (1963). Temporal Perception, Aphasia And D'ej'a Vu. Brain, 86, 403-24.         [ Links ]

Eisenberg, L.S.; Shannon, R.V.; Martinez, A.S.; Wygonski, J. & Boothroyd, A. (2000). Speech recognition with reduced spectral cues as a function of age. J. Acoust. Soc. Am., 107, 2704-2710.         [ Links ]

Erdonmez, D. & Morley, J.B. (1981). Preservation of acquired music performance functions with a dominant hemisphere lesion: a case report. Clin. Exp. Neurol., 18, 102-108.         [ Links ]

Eustache, F.; Lechevalier, B.; Viader, F. & Lambert, J. (1990). Identification and discrimination disorders in auditory perception: a report on two cases. Neuropsychologia, 28 (3), 257-270.         [ Links ]

Fodor, J.A. (1983). The modularity of mind. Cambridge, MA: The MIT Press.

Foxe, J.J. & Schroeder, C.E. (2005). The case for feedforward multisensory convergence during early cortical processing. Neuroreport, 16 (5), 419-23.         [ Links ]

Friederici, A.D.; Meyer, M. & Cramon, D.Y. von (2000). Auditory language comprehension: an event-related fMRI study on the processing of syntactic and lexical information. Brain Lang, 75 (3), 289-300.         [ Links ]

Geary, D.C. (2001). A Darwinian perspective on Mathematics and Instruction. In: Loveless, T. (Ed.). The great curriculum debate: How should we teach reading in math? (pp. 85-107). Washington, DC: Brookings Institute.

Geary, D.C. & Huffman, K. J. (2002). Brain and cognitive evolution: Forms of modularity and functions of mind. Psychological Bulletin, 128, 667-698.         [ Links ]

Gil-da-Costa, R.; Braun, A.; Lopes, M.; Hauser, M.D.; Carson, R.E.; Herscovitch, P. & Martin, A. (2004). Toward an evolutionary perspective on conceptual representation: species-specific calls activate visual and affective processing systems in the macaque. Proceedings of the National Academy of Sciences of the United States of America, 101 (50), 17516-17521.         [ Links ]

Gordon, H.W. (1978). Hemispheric asymmetry for dichotically-presented chords in musicians and non-musicians, males and females. Acta Psychol. (Amst), 42 (5), 383-395.         [ Links ]

Gosselin, N.; Peretz, I.; Noulhiane, M.; Hasboun, D.; Beckett, C.; Baulac, M. & Samson, S. (2005). Impaired recognition of scary music following unilateral temporal lobe excision. Brain, 128 (3), 628-640.         [ Links ]

Griffiths, T.D.; Rees, A.; Witton, C.; Cross, P.M.; Shakir, R.A. & Green, G.G. (1997). Spatial and temporal auditory processing deficits following right hemisphere infarction. A psychophysical study. Brain, 120 (5), 785-794.         [ Links ]

Griffiths, T.D.; Warren, J.D.; Dean, J.L. & Howard, D. (2004). When the feeling's gone: a selective loss of musical emotion. J. Neurol. Neurosurg. Psychiatry, 75 (2), 344-345.         [ Links ]

Habib, M. (2000). The neurological basis of developmental dyslexia: an overview and working hypothesis. Brain, 123 (12), 2373-2399.         [ Links ]

Halpern, A.R. & Zatorre, R.J. (1999). When that tune runs through your head: A PET investigation of auditory imagery for familiar melodies. Cerebral Cortex, 9 (7), 697-704.         [ Links ]

Hannon, E.E. & Johnson, S.P. (2005). Infants use meter to categorize rhythms and melodies: Implications for musical structure learning. Cognitive Psychology, 50 (4), 354-377.         [ Links ]

Harrington, D.L.; Haaland, K.Y. & Knight, R.T. (1998). Cortical networks underlying mechanisms of time perception. J. Neurosci., 18 (3), 1085-95.         [ Links ]

Hauser, M.D.; Chomsky, N. & Fitch, W.T. (2002). The faculty of language: what is it, who has it, and how did it evolve? Science, 298 (5598), 1569-1579.         [ Links ]

Hauser, M.D. & McDermott, J. (2003). The evolution of the music faculty: a comparative perspective. Nat. Neurosci., 6 (7), 663-668.         [ Links ]

Haxby, J.V.; Horwitz, B.; Ungerleider, L.G.; Maisog, J.M.; Pietrini, P. & Grady, C.L. (1994). The functional organization of human extrastriate cortex: a PET-rCBF study of selective attention to faces and locations. J. Neurosci, 14 (11), 6336-6353.         [ Links ]

Hyde, K. & Peretz, I. (2004). Brains that are out of tune but in time. Psychological Science. 15 (5). 356-360        [ Links ]

Janata, P. & Grafton, S.T. (2003). Swinging in the brain: shared neural substrates for behaviors related to sequencing and music. Nat. Neurosci., 6 (7), 682-687.         [ Links ]

Kessler, E.J.; Hansen, C. & Shepard, R.N. (1984). Tonal schemata in the perception of music in Bali and the West. Music Perception, 2 (2), 131-165.         [ Links ]

Keurs, M.; Brown, C.M.; Hagoort, P. & Stegeman, D.F. (1999). Electrophysiological manifestations of open- and closed-class words in patients with Broca's aphasia with agrammatic comprehension. An event-related brain potential study. Brain, 122 (5), 839-854.         [ Links ]

Khalfa, S.; Isabelle, P.; Jean-Pierre, B. & Manon, R. (2002). Event-related skin conductance responses to musical emotions in humans. Neurosci. Lett., 328 (2), 145-149.         [ Links ]

Kimura, D. (1993). Neuromotor Mechanisms in Human Communication. Oxford: Oxford University Press.         [ Links ]

Koelsch, S.; Gunter, T.C.; von Cramon, D.Y.; Zysset, S.; Lohmann, G. & Friederici, A.D. (2002). Bach speaks: A cortical "language-network" serves the processing of music. Neuroimage, 17 (2), 956-966.         [ Links ]

Koelsch, S.; Kasper, E.; Sammler, D.; Schulze, K.; Gunter, T. & Friederici, A.D. (2004). Music, language and meaning: brain signatures of semantic processing. Nat. Neurosci., 7 (3), 302-307.         [ Links ]

Krumhansl, C.L. (1990). Cognitive Foundations of Musical Pitch. New York: Oxford University Press.         [ Links ]

Krumhansl, C.L. (1995). Music psychology and music theory: problems and prospects. Music Theory Spectrum, 17, 53-90.         [ Links ]

Krumhansl C.L. (1997). An exploratory study of musical emotions and psychophysiology. Can. J. Exp. Psychol., 51 (4), 336-353.         [ Links ]

Krumhansl C.L. & Juscsyk, P.W. (1990). Infant's perception of phrase structure in music. Psychological Science, 1 (1), 70-73.         [ Links ]

Krumhansl, C.L. & Shepard, R.N. (1979). Quantification of the hierarchy of tonal functions within a diatonic context. J. Exp. Psychol. Hum. Percept. Perform.,5(4), 579-594.         [ Links ]

Krumhansl, C.L.; Toivanen, P.; Eerola, T.; Toiviainen, P.; Jarvinen, T. & Louhivuori, J. (2000). Cross-cultural music cognition: cognitive methodology applied to North Sami yoiks. Cognition, 76 (1), 13-58.         [ Links ]

Krumhansl, C.L. & Toiviainen, P. (2001). Tonal cognition. Annual New York Academy of Science, 930, 77-91.         [ Links ]

Kuhl, P.K. (2000). A new view of language acquisition. Proceedings of the National Academy of Sciences of the United States of America, 97 (22), 11850-11857.         [ Links ]

Kuhl, P.K.; Tsao, F.M.; Liu, H.M.; Zhang, Y. & Boer, B. De (2001). Language/culture/mind/ brain. Progress at the margins between disciplines. Annual New York Academy of Science, 935, 136-174.         [ Links ]

Lane, R.D.; Reiman, E.M.; Bradley, M.M.; Lang, P.J.; Ahern, G.L.; Davidson, R.J. & Schwartz, G.E. (1997). Neuroanatomical correlates of pleasant and unpleasant emotion. Neuropsychologia, 35 (11), 1437-1444.         [ Links ]

Lelekov-Boissard, T. & Dominey, P.F. (2002). Human brain potentials reveal similar processing of non-linguistic abstract structure and linguistic syntactic structure. Neurophysiol. Clin., 32 (1), 72-84.         [ Links ]

Lelekov, T.; Dominey, P.F.; ChossonTiraboschi, C.; Ventre-Dominey, J.; Labourel, D.; Michel, F. & Croisile, B. (2000). Selective Deficits in Processing Syntactic and Abstract Sequential Structure in Agrammatic Comprehension. Eur. J. Neurosci., 12 (11), 168.         [ Links ]

Lerdahl, F. & Jakendorff, R.A. (1983). A Generative Theory of Tonal Music. Cambridge, MA: MIT Press.

Levitin, D.J. & Menon, V. (2003). Musical structure is processed in "language" areas of the brain: a possible role for Brodmann Area 47 in temporal coherence. Neuroimage, 20 (4), 2142-2152.         [ Links ]

Lieberman, P. (1984). The Biology and Evolution of Language. Cambridge, MA: Harvard Univ. Press.

Liegeois-Chauvel, C.; Peretz, I.; Babai, M.; Laguitton, V. & Chauvel, P. (1998). Contribution of different cortical areas in the temporal lobes to music processing. Brain, 121 (10), 1853-1867.         [ Links ]

Lissauer, M. (1889). A case of visual agnosia with a contribution to theory. CognitiveNeuropsychology, 5, 157-192 (trans. from the German by M. Jackson, orig. publ. in Archiv für Psychiatrie und Nervenkrankheiten, 21, 222-270.         [ Links ]

Macaluso, E. & Driver, J. (2005). Multisensory spatial interactions: a window onto functional integration in the human brain. Trends Neurosci., 28 (5), 264-271.         [ Links ]

Maess, B.; Koelsch, S.; Gunter, T.C. & Friederici, A.D. (2001). Musical syntax is processed in Broca's area: an MEG study. Nature Neuroscience, 4 (5), 540-545.         [ Links ]

Mavlov, L. (1980). Amusia due to rhythm agnosia in a musician with left hemisphere damage: a non-auditory supramodal defect. Cortex, 16 (2), 331-338.         [ Links ]

Mazziotta, J.C.; Phelps, M.E.; Carson, R.E. & Kuhl, D.E. (1982). Tomographic Mapping of Human Cerebral Metabolism - Auditory- Stimulation. Neurology, 32 (9), 921-937.         [ Links ]

McClelland, J.L. (2001). Cognitive neuroscience. In: Smelser, N.J. and Balte, P.B.(Ed.). International Encyclopedia of the Social & Behavioral Sciences (pp. 2133-2139). Oxford: Pergamon Press.         [ Links ]

McDermott, J. & Hauser, M. (2004). Are consonant intervals music to their ears? Spontaneous acoustic preferences in a nonhuman primate. Cognition, 94 (2), B11-B21.         [ Links ]

Mellet, E.; Bricogne, S.; Crivello, F.; Mazoyer, B.; Denis, M. & Tzourio-Mazoyer, N. (2002). Neural basis of mental scanning of a topographic representation built from a text. Cereb. Cortex, 12 (12), 1322-1330.         [ Links ]

Mellet, E.; Tzourio, N.; Crivello, F.; Joliot, M.; Denis, M. & Mazoyer, B. (1996). Functional anatomy of spatial mental imagery generated from verbal instructions. J. Neurosci., 16 (20), 6504-6512.         [ Links ]

Meltzoff, A.N. (1999). Origins of theory of mind, cognition and communication. J. Commun. Disord., 32 (4), 251-269.         [ Links ]

Meltzoff, A.N. (2002). Imitation as a mechanism of social cognition: Origins of empathy, theory of mind, and the representation of action. In: Goswami, U. (Ed.). Handbook of Child Cognitive Development. Oxford: Blackwell Publishers.         [ Links ]

Meltzoff, A.N. & Decety, J. (2003). What imitation tells us about social cognition: a rapprochement between developmental psychology and cognitive neuroscience. Philos. Trans. R. Soc. Lond. B. Biol. Sci., 358 (1431), 491-500.         [ Links ]

Menon, V. & Levitin, D.J. (2005). The rewards of music listening: response and physiological connectivity of the mesolimbic system. Neuroimage, 28 (1), 175-184.         [ Links ]

Mesulam, M.M. (1998). From sensation to cognition. Brain, 121(6), 1013-1052.         [ Links ]

Murray, M.M.; Molholm, S.; Michel, C.M.; Heslenfeld, D.J.; Ritter, W.; Javitt, D.C.; Schroeder, C.E. & Foxe, J.J. (2005). Grabbing your ear: rapid auditory-somatosensory multisensory interactions in low-level sensory cortices are not constrained by stimulus alignment. Cereb. Cortex, 15 (7), 963-974.         [ Links ]

Nakamura, S.; Sadato, N.; Oohashi, T.; Nishina, E.; Fuwamoto, Y. & Yonekura, Y. (1999). Analysis of music-brain interaction with simultaneous measurement of regional cerebral blood flow and electroencephalogram beta rhythm in human subjects. Neuroscience Letters, 275 (3), 222-226.         [ Links ]

Nam, U. (1998). Pitch distribution in Korean court music. Evidence consistent with tonal hierarchies. Music Perception, 16 (2), 243-247.         [ Links ]

Nicholson, K.G.; Baum, S.; Kilgour, A.; Koh, C.K.; Munhall, K. G. & Cuddy, L.L. (2003). Impaired processing of prosodic and musical patterns after right hemisphere damage. Brain Cogn., 52 (3), 382-389.         [ Links ]

Ohnishi, T.; Matsuda, H.; Asada, T.; Aruga, M.; Hirakata, M.; Nishikawa, M.; Katoh, A. & Imabayashi, E. (2001). Functional anatomy of musical perception in musicians. Cerebral Cortex, 11 (8), 754-760.         [ Links ]

Oram, N. & Cuddy, L.L. (1995). Responsiveness of Western adults to pitch-distributional information in melodic sequences. Psychological Research , 57, 103-118.         [ Links ]

Overy, K. (2003). Dyslexia and music. From timing deficits to musical intervention. Annual New York Academy of Science, 999, 497-505.         [ Links ]

Overy, K.; Nicolson, R.I.; Fawcett, A.J. & Clarke, E.F. (2003). Dyslexia and music: measuring musical timing skills. Dyslexia, 9 (1), 18-36.         [ Links ]

Panksepp J. (1995). Hypothalamic regulation of energy balance and feeding behavior. Nutrition, 11 (4), 401-403.         [ Links ]

Panksepp, J. & Bernatzky, G. (2002). Emotional sounds and the brain: the neuro-affective foundations of musical appreciation. Behavioural Processes, 60 (2) 133-155.         [ Links ]

Papousek, H. (1996a). Musicality in infancy research: Biological and cultural origins of early musicality. In: Deliège, I. & Sloboda, J. (Eds.), Musical beginnings (pp. 37-55). Oxford: Oxford University Press.

Papousek, M. (1996b). Intuitive parenting: a hidden source of musical stimulation in infancy. In: Deliège, I. & Sloboda, J. (Eds.). Musical beginnings. (pp. 88-112) Oxford: Oxford University Press.

Parsons L.M. (2001). Exploring the functional neuroanatomy of music performance, perception, and comprehension. Annual New York Academy of Science, 930, 211-231.         [ Links ]

Patel, A.D.; Peretz, I.; Tramo, M. & Labreque, R. (1998). Processing prosodic and musical patterns: A neuropsychological investigation. Brain and Language, 61(1), 123-144.         [ Links ]

Patel, A.D. (2003a). Language, music, syntax and the brain. Nat. Neurosci., 6 (7), 674-681.

Patel, A.D. (2003b). Rhythm in language and music: parallels and differences. Annual New York Academy of Science, 999, 140-143.

Patel, A.D. & Balaban, E. (2004). Human auditory cortical dynamics during perception of long acoustic sequences: phase tracking of carrier frequency by the auditory steady-state response. Cerebral Cortex, 14 (1), 35-46.         [ Links ]

Patel, A.D. & Daniele, J.R. (2003). An empirical comparison of rhythm in language and music. Cognition, 87, B35-B45.         [ Links ]

Patel, A.D.; Iversen, J.R. & Rosenberg, J.C. (2006). Comparing the rhythm and melody of nspeech and music: The case of British English and French. Journal of the Acoustical Society ofAmerica, 119, 3034-3047.         [ Links ]

Patel, A.D.; Peretz, I.; Tramo, M. & Labreque, R. (1998). Processing prosodic and musical patterns: A neuropsychological investigation. Brain and Language, 61 (1), 123-144.         [ Links ]

Penhune, V.B.; Zatorre, R.J. & Feindel, W.H. (1999). The role of auditory cortex in retention of rhythmic patterns as studied in patients with temporal lobe removals including Heschl's gyrus. Neuropsychologia, 37 (3), 315-331.         [ Links ]

Peretz, I. (1990). Processing of Local and Global Musical Information by Unilateral Brain-Damaged Patients. Brain, 113 (4), 1185-1205.         [ Links ]

Peretz, I. (2006). The nature of music from a biological perspective. Cognition, 100,. 1-32        [ Links ]

Peretz, I. & Babai, M. (1992). The role of contour and intervals in the recognition of melody parts: evidence from cerebral asymmetries in musicians. Neuropsychologia, 30 (3), 277-292.         [ Links ]

Peretz, I.; Kolinsky, R.; Tramo, M.; Labrecque, R.; Hublet, C.; Demeurisse, G. & Belleville, S. (1994). Functional dissociations following bilateral lesions of auditory cortex. Brain, 117 (Pt 6), 1283-1301.         [ Links ]

Peretz, I. & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience, 6 (7), 688-691.         [ Links ]

Peretz, I. (1996). Can we loose memories for music? The case of music agnosia in a nonmusician. Journal of Cognitive Neuroscience, 8(6), 481-496.         [ Links ]

Peretz, I. & Gagnon, L. (1999). Dissociation between recognition and emotional judgements for melodies. Neurocase, 5, 21-30.         [ Links ]

Peretz, I. & Hebert, S. (2000). Toward a biological account of music experience. Brain and Cognition, 42 (1), 131-134.         [ Links ]

Peretz, I.; Ayotte, J.; Zatorre, R. J.; Mehler, J.; Ahad, P.; Penhune, V.B. & Jutras, B. (2002). Congenital amusia: A disorder of fine-grained pitch discrimination. Neuron, 33 (2), 185-191.         [ Links ]

Peretz, I.; Belleville, S. & Fontaine, S. (1997). Dissociations between music and language functions after cerebral resection: A new case of amusia without aphasia. Can. J. Exp. Psychol., 51 (4), 354-68.         [ Links ]

Peretz, I.; Blood, A.J.; Penhune, V. & Zatorre, R. (2001). Cortical deafness to dissonance. Brain, 124 (5), 928-940.         [ Links ]

Phillips, D.P. & Farmer, M.E. (1990). Acquired word deafness, and the temporal grain of sound representation in the primary auditory cortex. Behav. Brain Res., 40 (2), 85-94.         [ Links ]

Pinker, S. (1994). The Language Instinct. New York: Harper Perennial.         [ Links ]

Pinker, S. (1999). How The Mind Works. New York: W.W. Norton & Company.         [ Links ]

Platel, H.; Price, C.; Baron, J.C.; Wise, R.; Lambert, J.; Frackowiak, R.S.J.; Lechevalier, B. & Eustache, J. (1997). The structural components of music perception - A functional anatomical study. Brain, 120 (2), 229-243.         [ Links ]

Pötzl, O. & Uiberall, H. (1937). Zur Patholgie der Amusie. Wien Klin. Wochenschr, 50, 770-775.         [ Links ]

Ramachandran, V.S. & Hirstein, W. (1999). The science of art. Journal of Consciousness Studies, 6, 15-41.         [ Links ]

Ramus F.; Hauser, M.D.; Miller, C.; Morris, D. & Mehler, J. (2000). Language discrimination by human newborns and by cotton-top tamarin monkeys. Science, 288 (5464), 349-351.         [ Links ]

Saffran J.R.; Johnson E.K.; Aslin R.N. & Newport E.L. (1999). Statistical learning of tone sequences by human infants and adults. Cognition, 70 (1), 27-52.         [ Links ]

Saffran, J.R. & Griepentrog, G.J. (2001). Absolute pitch in infant auditory learning: Evidence for developmental reorganization. Developmental Psychology, 37 (1), 74-85.         [ Links ]

Saffran, J.R.; Loman, M.M. & Robertson, R.R.W. (2000). Infant memory for musical experiences. Cognition, 77 (1), B15-B23.         [ Links ]

Saint-Amour, D.; Saron, C.D.; Schroeder, C.E. & Foxe, J.J. (2005). Can whole brain nerve conduction velocity be derived from surface-recorded visual evoked potentials? A re-examination of Reed, Vernon, and Johnson (2004). Neuropsychologia, 43 (12), 1838-1844.         [ Links ]

Sakai, K.; Hikosaka, O.; Miyauchi, S.; Takino, R.; Tamada, T.; Iwata, N. K. & Nielsen, M. (1999). Neural representation of a rhythm depends on its interval ratio. Journal of Neuroscience, 19 (22), 10074-10081.         [ Links ]

Satoh, M.; Takeda, K.; Nagata, K.; Hatazawa, J. & Kuzuhara, S. (2001). Activated brain regions in musicians during an ensemble: a PET study. Cognitive Brain Research, 12, 101-108.         [ Links ]

Satoh, M.; Takeda, K.; Nagata, K.; Hatazawa, J. & Kuzuhara, S. (2003). The anterior portion of the bilateral temporal lobes participates in music perception: a positron emission tomography study. Am. J. Neuroradiol., 24 (9), 1843-1848.         [ Links ]

Saygin, A.P.; Dick, F.; Wilson, S.M.; Dronkers, N.F. & Bates, E. (2003). Neural resources for processing language and environmental sounds: evidence from aphasia. Brain, 126 (4), 928-945.         [ Links ]

Schroeder, C.E. & Foxe, J. (2005). Multisensory contributions to low-level, 'unisensory' processing. Curr. Opin. Neurobiol., 15 (4), 454-458.         [ Links ]

Schuppert, M.; Munte, T.F.; Wieringa, B.M. & Altenmuller, E. (2000). Receptive amusia: evidence for cross-hemispheric neural networks underlying music processing strategies. Brain, 123 (3), 546-559.         [ Links ]

Sergent, J.; Zuck, E.; Terriah, S. & Macdonald, B. (1992). Distributed Neural Network Underlying Musical Sight-Reading and Keyboard Performance. Science, 257 (5066), 106-109.         [ Links ]

Shams, L.; Iwaki, S.; Chawla, A. & Bhattacharya, J. (2005). Early modulation of visual cortex by sound: an MEG study. Neurosci. Lett., 378 (2), 76-81.         [ Links ]

Shannon, R.V.; Zeng, F.G.; Kamath, V.; Wygonski, J. & Ekelid, M. (1995). Speech recognition with primarily temporal cues. Science, 270 (5234), 303-304.         [ Links ]

Shimojo, S. & Shams, L. (2001). Sensory modalities are not separate modalities: plasticity and interactions. Curr. Opin. Neurobiol., 11 (4), 505-509.         [ Links ]

Smith, E.E.; Jonides, J.; & Koeppe, R.A. (1996). Dissociating verbal and spatial working memory using PET. Cerebral Cortex, 6 (1), 11-20.         [ Links ]

Steinke, W.R.; Cuddy, L.L. & Jakobson, L.S. (2001). Dissociations among functional subsystems governing melody recognition after right-hemisphere damage. Cognitive Neuropsychology, 18 (5), 411-437.         [ Links ]

Studdert-Kennedy, M. & Mody, M. (1995). Auditory temporal perception deficits in the reading-impaired: a critical review of the evidence. Psychon. Bull. Rev., 2, 508-514.         [ Links ]

Swisher, L. & Hirsh, I.J. (1972). Brain damage and the ordering of two temporally successive stimuli. Neuropsychologia, 10 (2), 137-152.         [ Links ]

Tallal, P.; Miller, S. & Fitch, R.H. (1993). Neurobiological basis of speech: a case for the preeminence of temporal processing. Annual New York Academy of Science, 682, 27-47.         [ Links ]

Tasaki, H. (1982). Perception of music stimuli by the Dichotic Listening Test--studies on college students making a specialty of music. No To Shinkei, 34, 1051-1057.         [ Links ]

Tincoff, R.; Hauser, M.; Tsao, F.; Spaepen, G.; Ramus, F. & Mehler, J. (2005). The role of speech rhythm in language discrimination: further tests with a non-human primate. Dev. Sci., 8 (1), 26-35.         [ Links ]

Toiviainen, P. & Eerola, T. (2003). Where is the Beat?: Comparison of Finnish and South African Listeners. In: Kopiez, R.; Lehmann, A.C.; Wolther, I. & Wolf, C. (Eds.), Proceedings of the 5th Triennial ESCOM Conference (pp. 501-504). Hanover, Germany: Hanover University of Music and Drama.         [ Links ]

Trainor, L.J.; McDonald, K.L. & Alain, C. (2002a). Automatic and controlled processing of melodic contour and interval information measured by electrical brain activity. Journal of Cognitive Neuroscience, 14 (3), 430-442.

Trainor, L.J. & Trehub, S.E. (1992). A Comparison of Infants and Adults Sensitivity to Western Musical Structure. Journal of Experimental Psychology-Human Perception and Performance, 18 (2), 394-402.         [ Links ]

Trainor, L.J.; Tsang, C.D. & Cheung, V.H.W. (2002b). Preference for consonance in 2- and 4-month-old infants. Music Perception, 20 (2), 187-194.

Tramo, M.J.; Cariani, P.A.; Delgutte, B. & Braida, L.D. (2001). Neurobiological foundations for the theory of harmony in western tonal music. Annual New York Academy of Science, 930, 92-116.         [ Links ]

Trehub, S.E. (2003). The developmental origins of musicality. Nat Neurosci, 6 (7), 669-73.         [ Links ]

Trehub, S.E.; Schellenberg, E.G. & Kamenetsky, S.B. (1999). Infants' and adults' perception of scale structure. Journal of Experimental Psychology-Human Perception and Performance, 25 (4), 965-975.         [ Links ]

Trehub, S.E.; Thorpe, L.A. & Trainor, L.J. (1990). Infants' perception of good and bad melodies. Psychomusicology, 9, 5-19.         [ Links ]

Trevarthen, C. (1979). Communication and cooperation in early infancy: a description of primary intersubjectivity. In: Bullowa, M. (Ed.). Before speech: the beginning of interpersonal communication (pp.321-347). New York: Cambridge University Press.         [ Links ]

Trevarthen, C. (1999). Musicalty and the intrinsic motive pulse: evidence from human psychobiology and infant communication. Musicae Scientiae, Special Issue, 155-215.         [ Links ]

Trevarthen, C. (2000). Autism as a neurodevelopmental disorder affecting communication and learning in early childhood: prenatal origins, post natal course and effective educational support. Prostaglandins Leukotrienes and Essential Fatty Acids, 63 (1-2), 41-46.         [ Links ]

Wilson, S.J. & Pressing, J. (1999). Neuropsychological assessment and the modeling of musical deficits. In: Pratt, R.R. & Erdonmez Grocke, D.E. (Eds). Music Medicine and Music Therapy: Expanding Horizons (pp. 47-74). Melbourne: University of Melbourne Press.         [ Links ]

Wilson, S.J.; Pressing, J.L. & Wales, R.J. (2002). Modelling rhythmic function in a musician post-stroke. Neuropsychologia, 40 (8), 1494-1505.         [ Links ]

Zatorre, R.J.; Belin, P. & Penhune, V.B. (2002). Structure and function of auditory cortex: music and speech. Trends in Cognitive Sciences, 6 (1), 37-46.         [ Links ]

Zatorre, R.J.; Evans, A.C. & Meyer, E. (1994). Neural Mechanisms Underlying Melodic Perception and Memory for Pitch. Journal of Neuroscience, 14 (4) 1908-1919.         [ Links ]

Zatorre, R.J.; Evans, A.C.; Meyer, E. & Gjedde, A. (1992). Lateralization of Phonetic and Pitch Discrimination in Speech Processing. Science, 256 (5058), 846-849.         [ Links ]

Zatorre, R.J.; Perry, D.W.; Beckett, C.A.; Westbury, C.F. & Evans, A.C. (1998). Functional anatomy of musical processing in listeners with absolute pitch and relative pitch. Proceedings of the National Academy of Sciences of the United States of America, 95 (6), 3172-3177.         [ Links ]

Zentner, M.R. & Kagan, J. (1996). Perception of music by infants. Nature, 383 (6595), 29.         [ Links ]

 

Endereço para correspondência
P. E. Andrade
Endereço para correspondência: Rua Sperêndio Cabrini, 231, Jd. Maria Isabel II. Marília, SP 17.516-300.
E-mail para correspondência: paulo_sustain@hotmail.com;

E. C. Konkiewitz
Endereço para correspondência: Alameda das Hortências, 120, Portal de Dourados, Dourados, MS 79.826-290.
E-mail para correspondência: ecastelon@hotmail.com.