<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
 <record>
  <leader>     naa a22        4500</leader>
  <controlfield tag="001">510739776</controlfield>
  <controlfield tag="003">CHVBK</controlfield>
  <controlfield tag="005">20180411083024.0</controlfield>
  <controlfield tag="007">cr unu---uuuuu</controlfield>
  <controlfield tag="008">180411e20131101xx      s     000 0 eng  </controlfield>
  <datafield tag="024" ind1="7" ind2="0">
   <subfield code="a">10.1007/s12193-013-0122-3</subfield>
   <subfield code="2">doi</subfield>
  </datafield>
  <datafield tag="035" ind1=" " ind2=" ">
   <subfield code="a">(NATIONALLICENCE)springer-10.1007/s12193-013-0122-3</subfield>
  </datafield>
  <datafield tag="245" ind1="0" ind2="0">
   <subfield code="a">Induction, recording and recognition of natural emotions from facial expressions and speech prosody</subfield>
   <subfield code="h">[Elektronische Daten]</subfield>
   <subfield code="c">[Kostas Karpouzis, George Caridakis, Roddy Cowie, Ellen Douglas-Cowie]</subfield>
  </datafield>
  <datafield tag="520" ind1="3" ind2=" ">
   <subfield code="a">Recording and annotating a multimodal database of natural expressivity is a task that requires careful planning and implementation, before even starting to apply feature extraction and recognition algorithms. Requirements and characteristics of such databases are inherently different than those of acted behaviour, both in terms of unconstrained expressivity of the human participants, and in terms of the expressed emotions. In this paper, we describe a method to induce, record and annotate natural emotions, which was used to provide multimodal data for dynamic emotion recognition from facial expressions and speech prosody; results from a dynamic recognition algorithm, based on recurrent neural networks, indicate that multimodal processing surpasses both speech and visual analysis by a wide margin. The SAL database was used in the framework of the Humaine Network of Excellence as a common ground for research in everyday, natural emotions.</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
   <subfield code="a">OpenInterface Association, 2013</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Affective computing</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Multimodal emotion recognition</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="690" ind1=" " ind2="7">
   <subfield code="a">Emotion induction methods</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Karpouzis</subfield>
   <subfield code="D">Kostas</subfield>
   <subfield code="u">Image, Video and Multimedia Systems Laboratory, National Technical University of Athens, Athens, Greece</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Caridakis</subfield>
   <subfield code="D">George</subfield>
   <subfield code="u">Image, Video and Multimedia Systems Laboratory, National Technical University of Athens, Athens, Greece</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Cowie</subfield>
   <subfield code="D">Roddy</subfield>
   <subfield code="u">School of Psychology, Queen's University, Belfast, Belfast, UK</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="700" ind1="1" ind2=" ">
   <subfield code="a">Douglas-Cowie</subfield>
   <subfield code="D">Ellen</subfield>
   <subfield code="u">School of Psychology, Queen's University, Belfast, Belfast, UK</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="773" ind1="0" ind2=" ">
   <subfield code="t">Journal on Multimodal User Interfaces</subfield>
   <subfield code="d">Springer Berlin Heidelberg</subfield>
   <subfield code="g">7/3(2013-11-01), 195-206</subfield>
   <subfield code="x">1783-7677</subfield>
   <subfield code="q">7:3&lt;195</subfield>
   <subfield code="1">2013</subfield>
   <subfield code="2">7</subfield>
   <subfield code="o">12193</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2="0">
   <subfield code="u">https://doi.org/10.1007/s12193-013-0122-3</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="908" ind1=" " ind2=" ">
   <subfield code="D">1</subfield>
   <subfield code="a">research-article</subfield>
   <subfield code="2">jats</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">856</subfield>
   <subfield code="E">40</subfield>
   <subfield code="u">https://doi.org/10.1007/s12193-013-0122-3</subfield>
   <subfield code="q">text/html</subfield>
   <subfield code="z">Onlinezugriff via DOI</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Karpouzis</subfield>
   <subfield code="D">Kostas</subfield>
   <subfield code="u">Image, Video and Multimedia Systems Laboratory, National Technical University of Athens, Athens, Greece</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Caridakis</subfield>
   <subfield code="D">George</subfield>
   <subfield code="u">Image, Video and Multimedia Systems Laboratory, National Technical University of Athens, Athens, Greece</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Cowie</subfield>
   <subfield code="D">Roddy</subfield>
   <subfield code="u">School of Psychology, Queen's University, Belfast, Belfast, UK</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">700</subfield>
   <subfield code="E">1-</subfield>
   <subfield code="a">Douglas-Cowie</subfield>
   <subfield code="D">Ellen</subfield>
   <subfield code="u">School of Psychology, Queen's University, Belfast, Belfast, UK</subfield>
   <subfield code="4">aut</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="P">773</subfield>
   <subfield code="E">0-</subfield>
   <subfield code="t">Journal on Multimodal User Interfaces</subfield>
   <subfield code="d">Springer Berlin Heidelberg</subfield>
   <subfield code="g">7/3(2013-11-01), 195-206</subfield>
   <subfield code="x">1783-7677</subfield>
   <subfield code="q">7:3&lt;195</subfield>
   <subfield code="1">2013</subfield>
   <subfield code="2">7</subfield>
   <subfield code="o">12193</subfield>
  </datafield>
  <datafield tag="900" ind1=" " ind2="7">
   <subfield code="a">Metadata rights reserved</subfield>
   <subfield code="b">Springer special CC-BY-NC licence</subfield>
   <subfield code="2">nationallicence</subfield>
  </datafield>
  <datafield tag="898" ind1=" " ind2=" ">
   <subfield code="a">BK010053</subfield>
   <subfield code="b">XK010053</subfield>
   <subfield code="c">XK010000</subfield>
  </datafield>
  <datafield tag="949" ind1=" " ind2=" ">
   <subfield code="B">NATIONALLICENCE</subfield>
   <subfield code="F">NATIONALLICENCE</subfield>
   <subfield code="b">NL-springer</subfield>
  </datafield>
 </record>
</collection>
