EBookClubs

Read Books & Download eBooks Full Online

EBookClubs

Read Books & Download eBooks Full Online

Book Neural Correlates of Auditory Cognition

Download or read book Neural Correlates of Auditory Cognition written by Yale E. Cohen and published by Springer Science & Business Media. This book was released on 2012-10-19 with total page 336 pages. Available in PDF, EPUB and Kindle. Book excerpt: Hearing and communication present a variety of challenges to the nervous system. To be heard and understood, a communication signal must be transformed from a time-varying acoustic waveform to a perceptual representation to an even more abstract representation that integrates memory stores with semantic/referential information. Finally, this complex, abstract representation must be interpreted to form categorical decisions that guide behavior. Did I hear the stimulus? From where and whom did it come? What does it tell me? How can I use this information to plan an action? All of these issues and questions underlie auditory cognition. Since the early 1990s, there has been a re-birth of studies that test the neural correlates of auditory cognition with a unique emphasis on the use of awake, behaving animals as model. Continuing today, how and where in the brain neural correlates of auditory cognition are formed is an intensive and active area of research. Importantly, our understanding of the role that the cortex plays in hearing has the potential to impact the next generation of cochlear- and brainstem-auditory implants and consequently help those with hearing impairments. Thus, it is timely to produce a volume that brings together this exciting literature on the neural correlates of auditory cognition. This volume compliments and extends many recent SHAR volumes such as Sound Source Localization (2005) Auditory Perception of Sound Sources (2007), and Human Auditory Cortex (2010). For example, in many of these volumes, similar issues are discussed such as auditory-object identification and perception with different emphases: in Auditory Perception of Sound Sources, authors discuss the underlying psychophysics/behavior, whereas in the Human Auditory Cortex, fMRI data are presented. The unique contribution of the proposed volume is that the authors will integrate both of these factors to highlight the neural correlates of cognition/behavior. Moreover, unlike other these other volumes, the neurophysiological data will emphasize the exquisite spatial and temporal resolution of single-neuron [as opposed to more coarse fMRI or MEG data] responses in order to reveal the elegant representations and computations used by the nervous system.

Book Neural Correlates of Auditory Cognition

Download or read book Neural Correlates of Auditory Cognition written by Yale E. Cohen and published by Springer Science & Business Media. This book was released on 2012-10-20 with total page 336 pages. Available in PDF, EPUB and Kindle. Book excerpt: Hearing and communication present a variety of challenges to the nervous system. To be heard and understood, a communication signal must be transformed from a time-varying acoustic waveform to a perceptual representation to an even more abstract representation that integrates memory stores with semantic/referential information. Finally, this complex, abstract representation must be interpreted to form categorical decisions that guide behavior. Did I hear the stimulus? From where and whom did it come? What does it tell me? How can I use this information to plan an action? All of these issues and questions underlie auditory cognition. Since the early 1990s, there has been a re-birth of studies that test the neural correlates of auditory cognition with a unique emphasis on the use of awake, behaving animals as model. Continuing today, how and where in the brain neural correlates of auditory cognition are formed is an intensive and active area of research. Importantly, our understanding of the role that the cortex plays in hearing has the potential to impact the next generation of cochlear- and brainstem-auditory implants and consequently help those with hearing impairments. Thus, it is timely to produce a volume that brings together this exciting literature on the neural correlates of auditory cognition. This volume compliments and extends many recent SHAR volumes such as Sound Source Localization (2005) Auditory Perception of Sound Sources (2007), and Human Auditory Cortex (2010). For example, in many of these volumes, similar issues are discussed such as auditory-object identification and perception with different emphases: in Auditory Perception of Sound Sources, authors discuss the underlying psychophysics/behavior, whereas in the Human Auditory Cortex, fMRI data are presented. The unique contribution of the proposed volume is that the authors will integrate both of these factors to highlight the neural correlates of cognition/behavior. Moreover, unlike other these other volumes, the neurophysiological data will emphasize the exquisite spatial and temporal resolution of single-neuron [as opposed to more coarse fMRI or MEG data] responses in order to reveal the elegant representations and computations used by the nervous system.

Book Neural Correlates of Auditory visual Speech Perception in Noise

Download or read book Neural Correlates of Auditory visual Speech Perception in Noise written by Jaimie Gilbert and published by ProQuest. This book was released on 2009 with total page 173 pages. Available in PDF, EPUB and Kindle. Book excerpt: Speech perception in noise may be facilitated by presenting the concurrent optic stimulus of observable speech gestures. Objective measures such as event-related potentials (ERPs) are crucial to understanding the processes underlying a facilitation of auditory-visual speech perception. Previous research has demonstrated that in quiet acoustic conditions auditory-visual speech perception occurs faster (decreased latency) and with less neural activity (decreased amplitude) than auditory-only speech perception. These empirical observations provide support for the construct of auditory-visual neural facilitation. Auditory-visual facilitation was quantified with response time and accuracy measures and the N1/P2 ERP waveform response as a function of changes in audibility (manipulation of the acoustic environment by testing a range of signal-to-noise ratios) and content of optic cue (manipulation of the types of cues available, e.g., speech, nonspeech-static, or non-speech-dynamic cues). Experiment 1 (Response Time Measures) evaluated participant responses in a speeded-response task investigating effects of both audibility and type of optic cue. Results revealed better accuracy and response times with visible speech gestures compared to those for any non-speech cue. Experiment 2 (Audibility) investigated the influence of audibility on auditory-visual facilitation in response time measures and the N1/P2 response. ERP measures showed effects of reduced audibility (slower latency, decreased amplitude) for both types of facial motion, i.e., speech and non-speech dynamic facial optic cues, compared to measures in quiet conditions. Experiment 3 (Optic Cues) evaluated the influence of the type of optic cue on auditory-visual facilitation with response time measures and the N1/P2 response. N1 latency was faster with both types of facial motion tested in this experiment, but N1 amplitude was decreased only with concurrent presentation of auditory and visual speech. The N1 ERP results of these experiments reveal that the effect of audibility alone does not explain auditory-visual facilitation in noise. The decreased N1 amplitude associated with the visible speech gesture and the concurrent auditory speech suggests that processing of the visible speech gesture either stimulates N1 generators or interacts with processing in N1 generators. A likely generator of the N1 response is the auditory cortex, which matures differently without auditory stimulation during a critical period. The impact of auditory-visual integration deprivation on neural development and ability to make use of optic cues must also be investigated. Further scientific understanding of any maturational differences or differences in processing due to auditory-visual integration deprivation is needed to promote utilization of auditory-visual facilitation of speech perception for individuals with auditory impairment. Research and (re)habilitation therapies for speech perception in noise must continue to emphasize the benefit of associating and integrating auditory and visual speech cues.

Book Neural Correlates of Auditory Word Processing in Infants and Adults

Download or read book Neural Correlates of Auditory Word Processing in Infants and Adults written by Katherine Elizabeth Travis and published by . This book was released on 2011 with total page 122 pages. Available in PDF, EPUB and Kindle. Book excerpt: For the majority of people, words are first learned and are communicated in high proportions in the auditory modality. However, the neural dynamics underlying speech perception are poorly understood. Even more limited, is knowledge of the neurophysiological processes and neuroanatomical structures that afford developing language abilities in infants. This dissertation investigates these issues in a series of related studies that are aimed at characterizing the spatial and temporal neural dynamics of auditory word processing in both developing 12-19 month old infants and adults. The first study, performed in adults, reveals new evidence for a neural response that is selective for auditory words, relative to acoustically-matched control sounds. This response appears to index a stage in speech processing wherein an incoming word sound is translated from an acoustic signal into a linguistically relevant code. This information can then be passed along the speech processing stream so that eventually the appropriate meaning of a word can be selected amongst representations stored within associative left fronto-temporal networks. The second study, performed in both adults and 12-18 month old infants, demonstrates that the neural mechanism responsible for encoding lexico-semantic word information has similar spatial and temporal characteristics in infants and adults. Prior work has not been able to establish whether infants and adults share similar neural substrates for language, and these findings suggest that the neurophysiological processes important for word understanding reside within similar neural networks throughout the lifespan. Finally, to gain a better understanding of the regional neuroanatomical changes that take place in the developing cortex of 12-19 month old-infants, the third study examines age-related changes tissues signal properties assessed with magnetic resonance imaging. This a period in development that is pivotal for emerging linguistic, cognitive and sensorimotor behaviors, however, the maturational changes that occur brain structures are poorly understood at these ages. This study reveals large changes in structural measures within precisely the specific areas that were demonstrated to generate lexico-semantic activity in study two. Together, these studies help to advance current understanding of neurophysiological processing stages and neural structures involved in auditory word processing in both the developing and mature brain. These findings invite a host of new studies that will continue to further knowledge of how speech processing is instantiated within the brain. Finally, with the use of multimodal imaging techniques such as those described in the present studies, there is increasing potential for new research aimed at understanding the neurobiological underpinning of language and other cognitive behaviors.

Book Timbre  Acoustics  Perception  and Cognition

Download or read book Timbre Acoustics Perception and Cognition written by Kai Siedenburg and published by Springer. This book was released on 2019-05-07 with total page 389 pages. Available in PDF, EPUB and Kindle. Book excerpt: Roughly defined as any property other than pitch, duration, and loudness that allows two sounds to be distinguished, timbre is a foundational aspect of hearing. The remarkable ability of humans to recognize sound sources and events (e.g., glass breaking, a friend’s voice, a tone from a piano) stems primarily from a capacity to perceive and process differences in the timbre of sounds. Timbre raises many important issues in psychology and the cognitive sciences, musical acoustics, speech processing, medical engineering, and artificial intelligence. Current research on timbre perception unfolds along three main fronts: On the one hand, researchers explore the principal perceptual processes that orchestrate timbre processing, such as the structure of its perceptual representation, sound categorization and recognition, memory for timbre, and its ability to elicit rich semantic associations, as well as the underlying neural mechanisms. On the other hand, timbre is studied as part of specific scenarios, including the perception of the human voice, as a structuring force in music, as perceived with cochlear implants, and through its role in affecting sound quality and sound design. Finally, computational acoustic models are sought through prediction of psychophysical data, physiologically inspired representations, and audio analysis-synthesis techniques. Along these three scientific fronts, significant breakthroughs have been achieved during the last decade. This volume will be the first book dedicated to a comprehensive and authoritative presentation of timbre perception and cognition research and the acoustic modeling of timbre. The volume will serve as a natural complement to the SHAR volumes on the basic auditory parameters of Pitch edited by Plack, Oxenham, Popper, and Fay, and Loudness by Florentine, Popper, and Fay. Moreover, through the integration of complementary scientific methods ranging from signal processing to brain imaging, the book has the potential to leverage new interdisciplinary synergies in hearing science. For these reasons, the volume will be exceptionally valuable to various subfields of hearing science, including cognitive auditory neuroscience, psychoacoustics, music perception and cognition, but may even exert significant influence on fields such as musical acoustics, music information retrieval, and acoustic signal processing. It is expected that the volume will have broad appeal to psychologists, neuroscientists, and acousticians involved in research on auditory perception and cognition. Specifically, this book will have a strong impact on hearing researchers with interest in timbre and will serve as the key publication and up-to-date reference on timbre for graduate students, postdoctoral researchers, as well as established scholars.

Book Neural Correlates of Auditory Processing and Language Impairment in Children with Autism Spectrum Disorders

Download or read book Neural Correlates of Auditory Processing and Language Impairment in Children with Autism Spectrum Disorders written by Shu Hui Yau and published by . This book was released on 2014 with total page 224 pages. Available in PDF, EPUB and Kindle. Book excerpt: The term autism spectrum disorders (ASD) refers to a group of neurodevelopmental disorders characterised by social and communication impairments, as well as restricted and repetitive patterns of behaviour. The thesis contains four studies using magnetoencephalography (MEG) to measure brain responses to auditory stimuli. The aim is to better understand the neural correlates of auditory processing defects in ASD, and determine how such deficits may be associated with spoken language impairment that affect many individuals on the autism spectrum.

Book Auditory Cognition and Human Performance

Download or read book Auditory Cognition and Human Performance written by Carryl L. Baldwin and published by CRC Press. This book was released on 2016-04-21 with total page 335 pages. Available in PDF, EPUB and Kindle. Book excerpt: Hearing and understanding sound- auditory processing- greatly enriches everyday life and enhances our ability to perform many tasks essential to survival. The complex soundscape in which we live influences where we direct our attention, how we communicate with each other, and how we interact with technological systems. Auditory Cognition and Human

Book Neural Correlates of Thinking

Download or read book Neural Correlates of Thinking written by Eduard Kraft and published by Springer Science & Business Media. This book was released on 2008-11-14 with total page 287 pages. Available in PDF, EPUB and Kindle. Book excerpt: The advances in neuroimaging technologies have led to substantial progress in understanding the neural mechanisms of cognitive functions. Thinking and reasoning have only recently been addressed by using neuroimaging techniques. The present book comprehensively explores current approaches and contributions to understanding the neural mechanisms of thinking in a concise and readable manner. It provides an insight into the state of the art and the potentials, but also the limitations of current neuroimaging methods for studying cognitive functions. The book will be a valuable companion for everyone interested in one of the most fascinating topics of cognitive neuroscience.

Book Language and Cognition

Download or read book Language and Cognition written by Kuniyoshi L. Sakai and published by Frontiers Media SA. This book was released on 2015-07-07 with total page 127 pages. Available in PDF, EPUB and Kindle. Book excerpt: Interaction between language and cognition remains an unsolved scientific problem. What are the differences in neural mechanisms of language and cognition? Why do children acquire language by the age of six, while taking a lifetime to acquire cognition? What is the role of language and cognition in thinking? Is abstract cognition possible without language? Is language just a communication device, or is it fundamental in developing thoughts? Why are there no animals with human thinking but without human language? Combinations even among 100 words and 100 objects (multiple words can represent multiple objects) exceed the number of all the particles in the Universe, and it seems that no amount of experience would suffice to learn these associations. How does human brain overcome this difficulty? Since the 19th century we know about involvement of Broca’s and Wernicke’s areas in language. What new knowledge of language and cognition areas has been found with fMRI and other brain imaging methods? Every year we know more about their anatomical and functional/effective connectivity. What can be inferred about mechanisms of their interaction, and about their functions in language and cognition? Why does the human brain show hemispheric (i.e., left or right) dominance for some specific linguistic and cognitive processes? Is understanding of language and cognition processed in the same brain area, or are there differences in language-semantic and cognitive-semantic brain areas? Is the syntactic process related to the structure of our conceptual world? Chomsky has suggested that language is separable from cognition. On the opposite, cognitive and construction linguistics emphasized a single mechanism of both. Neither has led to a computational theory so far. Evolutionary linguistics has emphasized evolution leading to a mechanism of language acquisition, yet proposed approaches also lead to incomputable complexity. There are some more related issues in linguistics and language education as well. Which brain regions govern phonology, lexicon, semantics, and syntax systems, as well as their acquisitions? What are the differences in acquisition of the first and second languages? Which mechanisms of cognition are involved in reading and writing? Are different writing systems affect relations between language and cognition? Are there differences in language-cognition interactions among different language groups (such as Indo-European, Chinese, Japanese, Semitic) and types (different degrees of analytic-isolating, synthetic-inflected, fused, agglutinative features)? What can be learned from sign languages? Rizzolatti and Arbib have proposed that language evolved on top of earlier mirror-neuron mechanism. Can this proposal answer the unknown questions about language and cognition? Can it explain mechanisms of language-cognition interaction? How does it relate to known brain areas and their interactions identified in brain imaging? Emotional and conceptual contents of voice sounds in animals are fused. Evolution of human language has demanded splitting of emotional and conceptual contents and mechanisms, although language prosody still carries emotional content. Is it a dying-off remnant, or is it fundamental for interaction between language and cognition? If language and cognitive mechanisms differ, unifying these two contents requires motivation, hence emotions. What are these emotions? Can they be measured? Tonal languages use pitch contours for semantic contents, are there differences in language-cognition interaction among tonal and atonal languages? Are emotional differences among cultures exclusively cultural, or also depend on languages? Interaction of language and cognition is thus full of mysteries, and we encourage papers addressing any aspect of this topic.

Book Neural Correlates of Auditory Perceptual Organization Measured with Direct Cortical Recordings in Humans

Download or read book Neural Correlates of Auditory Perceptual Organization Measured with Direct Cortical Recordings in Humans written by Andrew Richard Dykstra and published by . This book was released on 2011 with total page 181 pages. Available in PDF, EPUB and Kindle. Book excerpt: One of the primary functions of the human auditory system is to separate the complex mixture of sound arriving at the ears into neural representations of individual sound sources. This function is thought to be crucial for survival and communication in noisy settings, and allows listeners to selectively and dynamically attend to a sound source of interest while suppressing irrelevant information. How the brain works to perceptually organize the acoustic environment remains unclear despite the multitude of recent studies utilizing microelectrode recordings in experimental animals or non-invasive human neuroimaging. In particular, the role that brain areas outside the auditory cortex might play is, comparatively, vastly understudied. The experiments described in this thesis combined classic behavioral paradigms with electrical recordings made directly from the cortical surface of neurosurgical patients undergoing clinically-indicated invasive monitoring for localization of epileptogenic foci. By sampling from widespread brain areas with high temporal resolution while participants simultaneously engaged in streaming and jittered multi-tone masking paradigms, the present experiments sought to overcome limitations inherent in previous work, namely sampling extent, resolution in time and space, and direct knowledge of the perceptual experience of the listener. In experiment 1, participants listened to sequences of tones alternating in frequency (i.e., ABA-) and indicated whether they perceived the tones as grouped ("1 stream") or segregated ("2 streams"). As has been reported in neurologically-normal listeners since the 1950s, patients heard the sequences as grouped when the frequency separation between the A and B tones was small and segregated when it was large. Evoked potentials from widespread brain areas showed amplitude correlations with frequency separation but surprisingly did not differ based solely on perceptual organization in the absence of changes in the stimuli. In experiment 2, participants listened to sequences of jittered multi-tone masking stimuli on which a regularly-repeating target stream of tones was sometimes superimposed and indicated when they heard the target stream. Target detectability, as indexed behaviorally, increased throughout the course of each sequence. Evoked potentials and high-gamma activity differed strongly based on the listener's subjective perception of the target tones. These results extend and constrain theories of how the brain subserves auditory perceptual organization and suggests several new avenues of research for understanding the neural mechanisms underlying this critical function.

Book Neural Correlates of Quality Perception for Complex Speech Signals

Download or read book Neural Correlates of Quality Perception for Complex Speech Signals written by Jan-Niklas Antons and published by Springer. This book was released on 2015-02-11 with total page 108 pages. Available in PDF, EPUB and Kindle. Book excerpt: This book interconnects two essential disciplines to study the perception of speech: Neuroscience and Quality of Experience, which to date have rarely been used together for the purposes of research on speech quality perception. In five key experiments, the book demonstrates the application of standard clinical methods in neurophysiology on the one hand and of methods used in fields of research concerned with speech quality perception on the other. Using this combination, the book shows that speech stimuli with different lengths and different quality impairments are accompanied by physiological reactions related to quality variations, e.g., a positive peak in an event-related potential. Furthermore, it demonstrates that – in most cases – quality impairment intensity has an impact on the intensity of physiological reactions.

Book The Neural Correlates and Mechanisms Mediating the Integration of Auditory and Visual Information in the Human Brain

Download or read book The Neural Correlates and Mechanisms Mediating the Integration of Auditory and Visual Information in the Human Brain written by Sebastian Werner and published by . This book was released on 2010 with total page 145 pages. Available in PDF, EPUB and Kindle. Book excerpt:

Book An Introduction to the Event Related Potential Technique  second edition

Download or read book An Introduction to the Event Related Potential Technique second edition written by Steven J. Luck and published by MIT Press. This book was released on 2014-05-30 with total page 417 pages. Available in PDF, EPUB and Kindle. Book excerpt: An essential guide to designing, conducting, and analyzing event-related potential (ERP) experiments, completely updated for this edition. The event-related potential (ERP) technique, in which neural responses to specific events are extracted from the EEG, provides a powerful noninvasive tool for exploring the human brain. This volume describes practical methods for ERP research along with the underlying theoretical rationale. It offers researchers and students an essential guide to designing, conducting, and analyzing ERP experiments. This second edition has been completely updated, with additional material, new chapters, and more accessible explanations. Freely available supplementary material, including several online-only chapters, offer expanded or advanced treatment of selected topics. The first half of the book presents essential background information, describing the origins of ERPs, the nature of ERP components, and the design of ERP experiments. The second half of the book offers a detailed treatment of the main steps involved in conducting ERP experiments, covering such topics as recording the EEG, filtering the EEG and ERP waveforms, and quantifying amplitudes and latencies. Throughout, the emphasis is on rigorous experimental design and relatively simple analyses. New material in the second edition includes entire chapters devoted to components, artifacts, measuring amplitudes and latencies, and statistical analysis; updated coverage of recording technologies; concrete examples of experimental design; and many more figures. Online chapters cover such topics as overlap, localization, writing and reviewing ERP papers, and setting up and running an ERP lab.

Book Hearing Voices

    Book Details:
  • Author : Katrina R. Koslov
  • Publisher :
  • Release : 2005
  • ISBN :
  • Pages : 52 pages

Download or read book Hearing Voices written by Katrina R. Koslov and published by . This book was released on 2005 with total page 52 pages. Available in PDF, EPUB and Kindle. Book excerpt:

Book Brain and Music

    Book Details:
  • Author : Stefan Koelsch
  • Publisher : John Wiley & Sons
  • Release : 2012-03-22
  • ISBN : 1119943116
  • Pages : 323 pages

Download or read book Brain and Music written by Stefan Koelsch and published by John Wiley & Sons. This book was released on 2012-03-22 with total page 323 pages. Available in PDF, EPUB and Kindle. Book excerpt: A comprehensive survey of the latest neuroscientific research into the effects of music on the brain Covers a variety of topics fundamental for music perception, including musical syntax, musical semantics, music and action, music and emotion Includes general introductory chapters to engage a broad readership, as well as a wealth of detailed research material for experts Offers the most empirical (and most systematic) work on the topics of neural correlates of musical syntax and musical semantics Integrates research from different domains (such as music, language, action and emotion both theoretically and empirically, to create a comprehensive theory of music psychology

Book Neural Correlates of Top down Musical Temporal Processing

Download or read book Neural Correlates of Top down Musical Temporal Processing written by Emily Graber and published by . This book was released on 2018 with total page pages. Available in PDF, EPUB and Kindle. Book excerpt: For performers and listeners, mentally processing time during musical sequences is essential for executing and understanding the underlying structure and expressive intention in music. In performing classical chamber music for example, the musicians must dynamically monitor their own playing as well as that of their group members for synchrony, they must collectively prepare for and perform rubato or explicit tempo changes, and they must actively track time even when the group has rests instead of notes. Audience members, too, can do such temporal processes as anticipate expressive or scored tempo changes (if they happen to have prior familiarity with the piece), make predictions about the development of a changing musical tempo, and expect certain continuations of the music in time. Notably, the temporal processes mentioned here for both performers and listeners are active, based on voluntary effort put into creating, conveying, or engaging with ongoing music. Moreover, these active temporal processes may be done on top of whatever automatic processes occur due to simply hearing sounds arranged in time. In order to characterize the neural activities that reflect automatic temporal processing, many previous studies have taken bottom-up approaches, driving fixed temporal expectations with fixed stimulus properties and measuring the brain responses elicited by expectation violations after deviating patterns. Only recently have some studies started to investigate the neural activities that result from active temporal processing without manipulating the stimuli. In this dissertation, top-down musical temporal processes were studied in particular by using a unique paradigm with controlled stimuli designed to drive the deliberate processes that musicians and audiences regularly engage in. Specifically, experienced musicians were required to listen to and anticipate tempo accelerations, decelerations, or steady beats after visual cues without knowing when the tempo changes would start. Their electroencephalograms (EEGs) were recorded during this anticipation task, while they continued to monitor the beat sequences for continuity and smoothness during the actual tempo changes, and after each sequence finished. This dissertation presents three studies based on the obtained EEG recordings. Dynamic power modulations of neural oscillations in the beta-band (13-30 Hz) and endogenous evoked response components were analyzed before, during, and after tempo changes. How the targeted neural responses reflected the top-down musical temporal processes of (1) anticipation, (2) temporal interval prediction, and (3) expectation strength are reported in detail. To the field of auditory cognition, this work contributes evidence in support of the hypothesis that the brain actively interacts and engages with stimuli based on top-down goals. Anticipation, direction-specific temporal predictions, and expectations in silence were able to be decoded from ongoing beta modulations and evoked responses, contributing to a neuroscientific understanding of the behavior of beta modulations as well as a musical understanding of what musical temporal processing entails. This work was also the first to document beta modulations during silence after auditory sequences. Finally, compared to standard methods of driving temporal expectation, the paradigm here introduced a more ecologically valid yet EEG-friendly approach to investigate musical temporal processing under experimentally controlled conditions.