<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Sam V. Norman-Haignere</style></author><author><style face="normal" font="default" size="100%">Jenelle Feather</style></author><author><style face="normal" font="default" size="100%">Dana Boebinger</style></author><author><style face="normal" font="default" size="100%">Peter Brunner</style></author><author><style face="normal" font="default" size="100%">Anthony Ritaccio</style></author><author><style face="normal" font="default" size="100%">Josh H. McDermott</style></author><author><style face="normal" font="default" size="100%">Gerwin Schalk</style></author><author><style face="normal" font="default" size="100%">Nancy Kanwisher</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">A neural population selective for song in human auditory cortex</style></title><secondary-title><style face="normal" font="default" size="100%">Current Biology</style></secondary-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">Auditory Cortex</style></keyword><keyword><style  face="normal" font="default" size="100%">component</style></keyword><keyword><style  face="normal" font="default" size="100%">ECoG</style></keyword><keyword><style  face="normal" font="default" size="100%">Electrocorticography</style></keyword><keyword><style  face="normal" font="default" size="100%">fMRI</style></keyword><keyword><style  face="normal" font="default" size="100%">music</style></keyword><keyword><style  face="normal" font="default" size="100%">natural sounds</style></keyword><keyword><style  face="normal" font="default" size="100%">song</style></keyword><keyword><style  face="normal" font="default" size="100%">Speech</style></keyword><keyword><style  face="normal" font="default" size="100%">voice</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2022</style></year></dates><urls><web-urls><url><style face="normal" font="default" size="100%">https://www.sciencedirect.com/science/article/pii/S0960982222001312</style></url></web-urls></urls><volume><style face="normal" font="default" size="100%">32</style></volume><pages><style face="normal" font="default" size="100%">1470-1484.e12</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Summary How is music represented in the brain? While neuroimaging has revealed some spatial segregation between responses to music versus other sounds, little is known about the neural code for music itself. To address this question, we developed a method to infer canonical response components of human auditory cortex using intracranial responses to natural sounds, and further used the superior coverage of fMRI to map their spatial distribution. The inferred components replicated many prior findings, including distinct neural selectivity for speech and music, but also revealed a novel component that responded nearly exclusively to music with singing. Song selectivity was not explainable by standard acoustic features, was located near speech- and music-selective responses, and was also evident in individual electrodes. These results suggest that representations of music are fractionated into subpopulations selective for different types of music, one of which is specialized for the analysis of song.</style></abstract></record></records></xml>