<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>10</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Miller, John W</style></author><author><style face="normal" font="default" size="100%">Hermes, Dora</style></author><author><style face="normal" font="default" size="100%">Gerwin Schalk</style></author><author><style face="normal" font="default" size="100%">Ramsey, Nick F</style></author><author><style face="normal" font="default" size="100%">Jagadeesh, Bharathi</style></author><author><style face="normal" font="default" size="100%">den Nijs, Marcel</style></author><author><style face="normal" font="default" size="100%">Ojemann, J G</style></author><author><style face="normal" font="default" size="100%">Rao, Rajesh P N</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Detection of spontaneous class-specific visual stimuli with high temporal accuracy in human electrocorticography.</style></title><secondary-title><style face="normal" font="default" size="100%">Conf Proc IEEE Eng Med Biol Soc</style></secondary-title><alt-title><style face="normal" font="default" size="100%">Conf Proc IEEE Eng Med Biol Soc</style></alt-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">Algorithms</style></keyword><keyword><style  face="normal" font="default" size="100%">Electrocardiography</style></keyword><keyword><style  face="normal" font="default" size="100%">Evoked Potentials, Visual</style></keyword><keyword><style  face="normal" font="default" size="100%">Humans</style></keyword><keyword><style  face="normal" font="default" size="100%">Male</style></keyword><keyword><style  face="normal" font="default" size="100%">Pattern Recognition, Automated</style></keyword><keyword><style  face="normal" font="default" size="100%">Pattern Recognition, Visual</style></keyword><keyword><style  face="normal" font="default" size="100%">Photic Stimulation</style></keyword><keyword><style  face="normal" font="default" size="100%">Reproducibility of Results</style></keyword><keyword><style  face="normal" font="default" size="100%">Sensitivity and Specificity</style></keyword><keyword><style  face="normal" font="default" size="100%">User-Computer Interface</style></keyword><keyword><style  face="normal" font="default" size="100%">Visual Cortex</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2009</style></year><pub-dates><date><style  face="normal" font="default" size="100%">2009</style></date></pub-dates></dates><volume><style face="normal" font="default" size="100%">2009</style></volume><pages><style face="normal" font="default" size="100%">6465-8</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">Most brain-computer interface classification experiments from electrical potential recordings have been focused on the identification of classes of stimuli or behavior where the timing of experimental parameters is known or pre-designated. Real world experience, however, is spontaneous, and to this end we describe an experiment predicting the occurrence, timing, and types of visual stimuli perceived by a human subject from electrocorticographic recordings. All 300 of 300 presented stimuli were correctly detected, with a temporal precision of order 20 ms. The type of stimulus (face/house) was correctly identified in 95% of these cases. There were approximately 20 false alarm events, corresponding to a late 2nd neuronal response to a previously identified event.</style></abstract></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Brendan Z. Allison</style></author><author><style face="normal" font="default" size="100%">Dennis J. McFarland</style></author><author><style face="normal" font="default" size="100%">Gerwin Schalk</style></author><author><style face="normal" font="default" size="100%">Zheng, Shi Dong</style></author><author><style face="normal" font="default" size="100%">Moore-Jackson, Melody</style></author><author><style face="normal" font="default" size="100%">Jonathan Wolpaw</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Towards an independent brain-computer interface using steady state visual evoked potentials.</style></title><secondary-title><style face="normal" font="default" size="100%">Clin Neurophysiol</style></secondary-title><alt-title><style face="normal" font="default" size="100%">Clin Neurophysiol</style></alt-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">Adolescent</style></keyword><keyword><style  face="normal" font="default" size="100%">Adult</style></keyword><keyword><style  face="normal" font="default" size="100%">Attention</style></keyword><keyword><style  face="normal" font="default" size="100%">Brain</style></keyword><keyword><style  face="normal" font="default" size="100%">Brain Mapping</style></keyword><keyword><style  face="normal" font="default" size="100%">Dose-Response Relationship, Radiation</style></keyword><keyword><style  face="normal" font="default" size="100%">Electroencephalography</style></keyword><keyword><style  face="normal" font="default" size="100%">Evoked Potentials, Visual</style></keyword><keyword><style  face="normal" font="default" size="100%">Female</style></keyword><keyword><style  face="normal" font="default" size="100%">Humans</style></keyword><keyword><style  face="normal" font="default" size="100%">Male</style></keyword><keyword><style  face="normal" font="default" size="100%">Pattern Recognition, Visual</style></keyword><keyword><style  face="normal" font="default" size="100%">Photic Stimulation</style></keyword><keyword><style  face="normal" font="default" size="100%">Spectrum Analysis</style></keyword><keyword><style  face="normal" font="default" size="100%">User-Computer Interface</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2008</style></year><pub-dates><date><style  face="normal" font="default" size="100%">02/2008</style></date></pub-dates></dates><urls><web-urls><url><style face="normal" font="default" size="100%">http://www.ncbi.nlm.nih.gov/pubmed/18077208</style></url></web-urls></urls><volume><style face="normal" font="default" size="100%">119</style></volume><pages><style face="normal" font="default" size="100%">399-408</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">&lt;h4 style=&quot;font-size: 13px; margin: 0px 0.25em 0px 0px; text-transform: uppercase; float: left; font-family: arial, helvetica, clean, sans-serif; line-height: 17px;&quot;&gt;OBJECTIVE:&amp;nbsp;&lt;/h4&gt;
&lt;p style=&quot;margin: 0px 0px 0.5em; font-family: arial, helvetica, clean, sans-serif; font-size: 13px; line-height: 17px;&quot;&gt;Brain-computer interface (BCI) systems using steady state visual evoked potentials (SSVEPs) have allowed healthy subjects to communicate. However, these systems may not work in severely disabled users because they may depend on gaze shifting. This study evaluates the hypothesis that overlapping stimuli can evoke changes in SSVEP activity sufficient to control a BCI. This would provide evidence that SSVEP BCIs could be used without shifting gaze.&lt;/p&gt;
&lt;h4 style=&quot;font-size: 13px; margin: 0px 0.25em 0px 0px; text-transform: uppercase; float: left; font-family: arial, helvetica, clean, sans-serif; line-height: 17px;&quot;&gt;METHODS:&amp;nbsp;&lt;/h4&gt;
&lt;p style=&quot;margin: 0px 0px 0.5em; font-family: arial, helvetica, clean, sans-serif; font-size: 13px; line-height: 17px;&quot;&gt;Subjects viewed a display containing two images that each oscillated at a different frequency. Different conditions used overlapping or non-overlapping images to explore dependence on gaze function. Subjects were asked to direct attention to one or the other of these images during each of 12 one-minute runs.&lt;/p&gt;
&lt;h4 style=&quot;font-size: 13px; margin: 0px 0.25em 0px 0px; text-transform: uppercase; float: left; font-family: arial, helvetica, clean, sans-serif; line-height: 17px;&quot;&gt;RESULTS:&amp;nbsp;&lt;/h4&gt;
&lt;p style=&quot;margin: 0px 0px 0.5em; font-family: arial, helvetica, clean, sans-serif; font-size: 13px; line-height: 17px;&quot;&gt;Half of the subjects produced differences in SSVEP activity elicited by overlapping stimuli that could support BCI control. In all remaining users, differences did exist at corresponding frequencies but were not strong enough to allow effective control.&lt;/p&gt;
&lt;h4 style=&quot;font-size: 13px; margin: 0px 0.25em 0px 0px; text-transform: uppercase; float: left; font-family: arial, helvetica, clean, sans-serif; line-height: 17px;&quot;&gt;CONCLUSIONS:&amp;nbsp;&lt;/h4&gt;
&lt;p style=&quot;margin: 0px 0px 0.5em; font-family: arial, helvetica, clean, sans-serif; font-size: 13px; line-height: 17px;&quot;&gt;The&amp;nbsp;&lt;span class=&quot;highlight&quot;&gt;data&lt;/span&gt;&amp;nbsp;demonstrate that SSVEP differences sufficient for BCI control may be elicited by selective attention to one of two overlapping stimuli. Thus, some SSVEP-based BCI approaches may not depend on gaze control. The nature and extent of any BCI's dependence on muscle activity is a function of many factors, including the display, task, environment, and user.&lt;/p&gt;
&lt;h4 style=&quot;font-size: 13px; margin: 0px 0.25em 0px 0px; text-transform: uppercase; float: left; font-family: arial, helvetica, clean, sans-serif; line-height: 17px;&quot;&gt;SIGNIFICANCE:&amp;nbsp;&lt;/h4&gt;
&lt;p style=&quot;margin: 0px 0px 0.5em; font-family: arial, helvetica, clean, sans-serif; font-size: 13px; line-height: 17px;&quot;&gt;SSVEP BCIs might function in severely disabled users unable to reliably control gaze. Further research with these users is necessary to explore the optimal parameters of such a system and validate online performance in a home environment.&lt;/p&gt;</style></abstract><issue><style face="normal" font="default" size="100%">2</style></issue></record></records></xml>