Serveur d'exploration sur l'opéra

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA

Identifieur interne : 000B01 ( Ncbi/Merge ); précédent : 000B00; suivant : 000B02

Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA

Auteurs : Teija Waaramaa [Finlande] ; Timo Leisiö [Finlande]

Source :

RBID : PMC:3689256

Abstract

The present study focused on voice quality and the perception of the basic emotions from speech samples in cross-cultural conditions. It was examined whether voice quality, cultural, or language background, age, or gender were related to the identification of the emotions. Professional actors (n2) and actresses (n2) produced non-sense sentences (n32) and protracted vowels (n8) expressing the six basic emotions, interest, and a neutral emotional state. The impact of musical interests on the ability to distinguish between emotions or valence (on an axis positivity – neutrality – negativity) from voice samples was studied. Listening tests were conducted on location in five countries: Estonia, Finland, Russia, Sweden, and the USA with 50 randomly chosen participants (25 males and 25 females) in each country. The participants (total N = 250) completed a questionnaire eliciting their background information and musical interests. The responses in the listening test and the questionnaires were statistically analyzed. Voice quality parameters and the share of the emotions and valence identified correlated significantly with each other for both genders. The percentage of emotions and valence identified was clearly above the chance level in each of the five countries studied, however, the countries differed significantly from each other for the identified emotions and the gender of the speaker. The samples produced by females were identified significantly better than those produced by males. Listener's age was a significant variable. Only minor gender differences were found for the identification. Perceptual confusion in the listening test between emotions seemed to be dependent on their similar voice production types. Musical interests tended to have a positive effect on the identification of the emotions. The results also suggest that identifying emotions from speech samples may be easier for those listeners who share a similar language or cultural background with the speaker.


Url:
DOI: 10.3389/fpsyg.2013.00344
PubMed: 23801972
PubMed Central: 3689256

Links toward previous steps (curation, corpus...)


Links to Exploration step

PMC:3689256

Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA</title>
<author>
<name sortKey="Waaramaa, Teija" sort="Waaramaa, Teija" uniqKey="Waaramaa T" first="Teija" last="Waaramaa">Teija Waaramaa</name>
<affiliation wicri:level="1">
<nlm:aff id="aff1">
<institution>School of Communication Media and Theatre, University of Tampere</institution>
<country>Tampere, Finland</country>
</nlm:aff>
<country xml:lang="fr">Finlande</country>
<wicri:regionArea></wicri:regionArea>
<wicri:regionArea># see nlm:aff region in country</wicri:regionArea>
</affiliation>
</author>
<author>
<name sortKey="Leisio, Timo" sort="Leisio, Timo" uniqKey="Leisio T" first="Timo" last="Leisiö">Timo Leisiö</name>
<affiliation wicri:level="1">
<nlm:aff id="aff2">
<institution>School of Social Sciences and Humanities, University of Tampere</institution>
<country>Tampere, Finland</country>
</nlm:aff>
<country xml:lang="fr">Finlande</country>
<wicri:regionArea></wicri:regionArea>
<wicri:regionArea># see nlm:aff region in country</wicri:regionArea>
</affiliation>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">PMC</idno>
<idno type="pmid">23801972</idno>
<idno type="pmc">3689256</idno>
<idno type="url">http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3689256</idno>
<idno type="RBID">PMC:3689256</idno>
<idno type="doi">10.3389/fpsyg.2013.00344</idno>
<date when="2013">2013</date>
<idno type="wicri:Area/Pmc/Corpus">000E16</idno>
<idno type="wicri:Area/Pmc/Curation">000E16</idno>
<idno type="wicri:Area/Pmc/Checkpoint">000169</idno>
<idno type="wicri:Area/Ncbi/Merge">000B01</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en" level="a" type="main">Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA</title>
<author>
<name sortKey="Waaramaa, Teija" sort="Waaramaa, Teija" uniqKey="Waaramaa T" first="Teija" last="Waaramaa">Teija Waaramaa</name>
<affiliation wicri:level="1">
<nlm:aff id="aff1">
<institution>School of Communication Media and Theatre, University of Tampere</institution>
<country>Tampere, Finland</country>
</nlm:aff>
<country xml:lang="fr">Finlande</country>
<wicri:regionArea></wicri:regionArea>
<wicri:regionArea># see nlm:aff region in country</wicri:regionArea>
</affiliation>
</author>
<author>
<name sortKey="Leisio, Timo" sort="Leisio, Timo" uniqKey="Leisio T" first="Timo" last="Leisiö">Timo Leisiö</name>
<affiliation wicri:level="1">
<nlm:aff id="aff2">
<institution>School of Social Sciences and Humanities, University of Tampere</institution>
<country>Tampere, Finland</country>
</nlm:aff>
<country xml:lang="fr">Finlande</country>
<wicri:regionArea></wicri:regionArea>
<wicri:regionArea># see nlm:aff region in country</wicri:regionArea>
</affiliation>
</author>
</analytic>
<series>
<title level="j">Frontiers in Psychology</title>
<idno type="e-ISSN">1664-1078</idno>
<imprint>
<date when="2013">2013</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass></textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">
<p>The present study focused on voice quality and the perception of the basic emotions from speech samples in cross-cultural conditions. It was examined whether voice quality, cultural, or language background, age, or gender were related to the identification of the emotions. Professional actors (n2) and actresses (n2) produced non-sense sentences (n32) and protracted vowels (n8) expressing the six basic emotions, interest, and a neutral emotional state. The impact of musical interests on the ability to distinguish between emotions or valence (on an axis positivity – neutrality – negativity) from voice samples was studied. Listening tests were conducted on location in five countries: Estonia, Finland, Russia, Sweden, and the USA with 50 randomly chosen participants (25 males and 25 females) in each country. The participants (total
<italic>N</italic>
= 250) completed a questionnaire eliciting their background information and musical interests. The responses in the listening test and the questionnaires were statistically analyzed. Voice quality parameters and the share of the emotions and valence identified correlated significantly with each other for both genders. The percentage of emotions and valence identified was clearly above the chance level in each of the five countries studied, however, the countries differed significantly from each other for the identified emotions and the gender of the speaker. The samples produced by females were identified significantly better than those produced by males. Listener's age was a significant variable. Only minor gender differences were found for the identification. Perceptual confusion in the listening test between emotions seemed to be dependent on their similar voice production types. Musical interests tended to have a positive effect on the identification of the emotions. The results also suggest that identifying emotions from speech samples may be easier for those listeners who share a similar language or cultural background with the speaker.</p>
</div>
</front>
<back>
<div1 type="bibliography">
<listBibl>
<biblStruct>
<analytic>
<author>
<name sortKey="Abelin, " uniqKey="Abelin ">Å Abelin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Abelin, " uniqKey="Abelin ">Å. Abelin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Abelin, " uniqKey="Abelin ">Å. Abelin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Abelin, " uniqKey="Abelin ">Å. Abelin</name>
</author>
<author>
<name sortKey="Allwood, J" uniqKey="Allwood J">J. Allwood</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Balkwill, L L" uniqKey="Balkwill L">L.-L. Balkwill</name>
</author>
<author>
<name sortKey="Thompson, W F" uniqKey="Thompson W">W. F. Thompson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Banse, R" uniqKey="Banse R">R. Banse</name>
</author>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Belin, P" uniqKey="Belin P">P. Belin</name>
</author>
<author>
<name sortKey="Bestelmeyer, P E G" uniqKey="Bestelmeyer P">P. E. G. Bestelmeyer</name>
</author>
<author>
<name sortKey="Latinus, M" uniqKey="Latinus M">M. Latinus</name>
</author>
<author>
<name sortKey="Watson, R" uniqKey="Watson R">R. Watson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Besson, M" uniqKey="Besson M">M. Besson</name>
</author>
<author>
<name sortKey="Magne, C" uniqKey="Magne C">C. Magne</name>
</author>
<author>
<name sortKey="Schon, D" uniqKey="Schon D">D. Schön</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Cross, I" uniqKey="Cross I">I. Cross</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Ekman, P" uniqKey="Ekman P">P. Ekman</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Falk, D" uniqKey="Falk D">D. Falk</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fant, G" uniqKey="Fant G">G. Fant</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fecteau, S" uniqKey="Fecteau S">S. Fecteau</name>
</author>
<author>
<name sortKey="Armony, J L" uniqKey="Armony J">J. L. Armony</name>
</author>
<author>
<name sortKey="Yves, J" uniqKey="Yves J">J. Yves</name>
</author>
<author>
<name sortKey="Belin, P" uniqKey="Belin P">P. Belin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="F Nagy, I" uniqKey="F Nagy I">I. Fónagy</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fritz, T" uniqKey="Fritz T">T. Fritz</name>
</author>
<author>
<name sortKey="Jentschke, S" uniqKey="Jentschke S">S. Jentschke</name>
</author>
<author>
<name sortKey="Gosselin, N" uniqKey="Gosselin N">N. Gosselin</name>
</author>
<author>
<name sortKey="Sammler, D" uniqKey="Sammler D">D. Sammler</name>
</author>
<author>
<name sortKey="Peretz, I" uniqKey="Peretz I">I. Peretz</name>
</author>
<author>
<name sortKey="Turner, R" uniqKey="Turner R">R. Turner</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fr Kj R Jensen, B" uniqKey="Fr Kj R Jensen B">B. Frøkjær-Jensen</name>
</author>
<author>
<name sortKey="Prytz, S" uniqKey="Prytz S">S. Prytz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Gauffin, J" uniqKey="Gauffin J">J. Gauffin</name>
</author>
<author>
<name sortKey="Sundberg, J" uniqKey="Sundberg J">J. Sundberg</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Gentilucci, M" uniqKey="Gentilucci M">M. Gentilucci</name>
</author>
<author>
<name sortKey="Corballis, M C" uniqKey="Corballis M">M. C. Corballis</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Hannon, E E" uniqKey="Hannon E">E. E. Hannon</name>
</author>
<author>
<name sortKey="Trehub, S E" uniqKey="Trehub S">S. E. Trehub</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Higgins, M B" uniqKey="Higgins M">M. B. Higgins</name>
</author>
<author>
<name sortKey="Schulte, L" uniqKey="Schulte L">L. Schulte</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Imaizumi, S" uniqKey="Imaizumi S">S. Imaizumi</name>
</author>
<author>
<name sortKey="Homma, M" uniqKey="Homma M">M. Homma</name>
</author>
<author>
<name sortKey="Ozawa, Y" uniqKey="Ozawa Y">Y. Ozawa</name>
</author>
<author>
<name sortKey="Maruishi, M" uniqKey="Maruishi M">M. Maruishi</name>
</author>
<author>
<name sortKey="Muranaka, H" uniqKey="Muranaka H">H. Muranaka</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Iversen, J R" uniqKey="Iversen J">J. R. Iversen</name>
</author>
<author>
<name sortKey="Patel, A D" uniqKey="Patel A">A. D. Patel</name>
</author>
<author>
<name sortKey="Ohgushi, K" uniqKey="Ohgushi K">K. Ohgushi</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Izard, C E" uniqKey="Izard C">C. E. Izard</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Juslin, P N" uniqKey="Juslin P">P. N. Juslin</name>
</author>
<author>
<name sortKey="Laukka, P" uniqKey="Laukka P">P. Laukka</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Juslin, P N" uniqKey="Juslin P">P. N. Juslin</name>
</author>
<author>
<name sortKey="Laukka, P" uniqKey="Laukka P">P. Laukka</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Juslin, P N" uniqKey="Juslin P">P. N. Juslin</name>
</author>
<author>
<name sortKey="V Stf Ll, D" uniqKey="V Stf Ll D">D. Västfäll</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Koeda, M" uniqKey="Koeda M">M. Koeda</name>
</author>
<author>
<name sortKey="Belin, P" uniqKey="Belin P">P. Belin</name>
</author>
<author>
<name sortKey="Hama, T" uniqKey="Hama T">T. Hama</name>
</author>
<author>
<name sortKey="Masuda, T" uniqKey="Masuda T">T. Masuda</name>
</author>
<author>
<name sortKey="Matsuura, M" uniqKey="Matsuura M">M. Matsuura</name>
</author>
<author>
<name sortKey="Okubo, Y" uniqKey="Okubo Y">Y. Okubo</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kotlyar, G M" uniqKey="Kotlyar G">G. M. Kotlyar</name>
</author>
<author>
<name sortKey="Morozov, V P" uniqKey="Morozov V">V. P. Morozov</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Ladd, D R" uniqKey="Ladd D">D. R. Ladd</name>
</author>
<author>
<name sortKey="Silverman, K E A" uniqKey="Silverman K">K. E. A. Silverman</name>
</author>
<author>
<name sortKey="Tolk Mitt, F" uniqKey="Tolk Mitt F">F. Tolk-mitt</name>
</author>
<author>
<name sortKey="Bergmann, G" uniqKey="Bergmann G">G. Bergmann</name>
</author>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Laukka, P" uniqKey="Laukka P">P. Laukka</name>
</author>
<author>
<name sortKey="Juslin, P N" uniqKey="Juslin P">P. N. Juslin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Laukkanen, A M" uniqKey="Laukkanen A">A.-M. Laukkanen</name>
</author>
<author>
<name sortKey="Alku, P" uniqKey="Alku P">P. Alku</name>
</author>
<author>
<name sortKey="Airas, M" uniqKey="Airas M">M. Airas</name>
</author>
<author>
<name sortKey="Waaramaa, T" uniqKey="Waaramaa T">T. Waaramaa</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Laukkanen, A M" uniqKey="Laukkanen A">A.-M. Laukkanen</name>
</author>
<author>
<name sortKey="Vilkman, E" uniqKey="Vilkman E">E. Vilkman</name>
</author>
<author>
<name sortKey="Alku, P" uniqKey="Alku P">P. Alku</name>
</author>
<author>
<name sortKey="Oksanen, H" uniqKey="Oksanen H">H. Oksanen</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Laver, J" uniqKey="Laver J">J. Laver</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Leisio, T" uniqKey="Leisio T">T. Leisiö</name>
</author>
<author>
<name sortKey="Ebeling, M" uniqKey="Ebeling M">M. Ebeling</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Levitin, D J" uniqKey="Levitin D">D. J. Levitin</name>
</author>
<author>
<name sortKey="Tirovolas, A K" uniqKey="Tirovolas A">A. K. Tirovolas</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Levitin, D J" uniqKey="Levitin D">D. J. Levitin</name>
</author>
<author>
<name sortKey="Tirovolas, A K" uniqKey="Tirovolas A">A. K. Tirovolas</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Liberman, A M" uniqKey="Liberman A">A. M. Liberman</name>
</author>
<author>
<name sortKey="Mattingly, I G" uniqKey="Mattingly I">I. G. Mattingly</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Liberman, A M" uniqKey="Liberman A">A. M. Liberman</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Lima, C F" uniqKey="Lima C">C. F. Lima</name>
</author>
<author>
<name sortKey="Castro, S L" uniqKey="Castro S">S. L. Castro</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Matsumoto, D" uniqKey="Matsumoto D">D. Matsumoto</name>
</author>
<author>
<name sortKey="Franklin, B" uniqKey="Franklin B">B. Franklin</name>
</author>
<author>
<name sortKey="Choi, J W" uniqKey="Choi J">J.-W. Choi</name>
</author>
<author>
<name sortKey="Rogers, D" uniqKey="Rogers D">D. Rogers</name>
</author>
<author>
<name sortKey="Tatani, H" uniqKey="Tatani H">H. Tatani</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Mithen, S" uniqKey="Mithen S">S. Mithen</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Morrison, S J" uniqKey="Morrison S">S. J. Morrison</name>
</author>
<author>
<name sortKey="Demorest, S M" uniqKey="Demorest S">S. M. Demorest</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Murray, I R" uniqKey="Murray I">I. R. Murray</name>
</author>
<author>
<name sortKey="Arnott, J L" uniqKey="Arnott J">J. L. Arnott</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Nordenberg, M" uniqKey="Nordenberg M">M. Nordenberg</name>
</author>
<author>
<name sortKey="Sundberg, J" uniqKey="Sundberg J">J. Sundberg</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Panksepp, J" uniqKey="Panksepp J">J. Panksepp</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Richman, B" uniqKey="Richman B">B. Richman</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Rizzolatti, G" uniqKey="Rizzolatti G">G. Rizzolatti</name>
</author>
<author>
<name sortKey="Fadiga, L" uniqKey="Fadiga L">L. Fadiga</name>
</author>
<author>
<name sortKey="Gallese, V" uniqKey="Gallese V">V. Gallese</name>
</author>
<author>
<name sortKey="Fogassi, L" uniqKey="Fogassi L">L. Fogassi</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Sauter, D A" uniqKey="Sauter D">D. A. Sauter</name>
</author>
<author>
<name sortKey="Eisner, F" uniqKey="Eisner F">F. Eisner</name>
</author>
<author>
<name sortKey="Ekman, P" uniqKey="Ekman P">P. Ekman</name>
</author>
<author>
<name sortKey="Scott, S K" uniqKey="Scott S">S. K. Scott</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
<author>
<name sortKey="Banse, R" uniqKey="Banse R">R. Banse</name>
</author>
<author>
<name sortKey="Wall Bott, H G" uniqKey="Wall Bott H">H. G. Wall-bott</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
<author>
<name sortKey="Banse, R" uniqKey="Banse R">R. Banse</name>
</author>
<author>
<name sortKey="Wallbott, H G" uniqKey="Wallbott H">H. G. Wallbott</name>
</author>
<author>
<name sortKey="Goldbeck, T" uniqKey="Goldbeck T">T. Goldbeck</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Scherer, K R" uniqKey="Scherer K">K. R. Scherer</name>
</author>
<author>
<name sortKey="Ellgring, H" uniqKey="Ellgring H">H. Ellgring</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Schirmer, A" uniqKey="Schirmer A">A. Schirmer</name>
</author>
<author>
<name sortKey="Kotz, S A" uniqKey="Kotz S">S. A. Kotz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Schirmer, A" uniqKey="Schirmer A">A. Schirmer</name>
</author>
<author>
<name sortKey="Kotz, S A" uniqKey="Kotz S">S. A. Kotz</name>
</author>
<author>
<name sortKey="Friederici, A" uniqKey="Friederici A">A. Friederici</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Schirmer, A" uniqKey="Schirmer A">A. Schirmer</name>
</author>
<author>
<name sortKey="Simpson, E" uniqKey="Simpson E">E. Simpson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Strait, D L" uniqKey="Strait D">D. L. Strait</name>
</author>
<author>
<name sortKey="Kraus, N" uniqKey="Kraus N">N. Kraus</name>
</author>
<author>
<name sortKey="Skoe, E" uniqKey="Skoe E">E. Skoe</name>
</author>
<author>
<name sortKey="Ashley, R" uniqKey="Ashley R">R. Ashley</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Sundberg, J" uniqKey="Sundberg J">J. Sundberg</name>
</author>
<author>
<name sortKey="Nordenberg, M" uniqKey="Nordenberg M">M. Nordenberg</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Thompson, W E" uniqKey="Thompson W">W. E. Thompson</name>
</author>
<author>
<name sortKey="Schellenberg, E G" uniqKey="Schellenberg E">E. G. Schellenberg</name>
</author>
<author>
<name sortKey="Husain, G" uniqKey="Husain G">G. Husain</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Trehub, S E" uniqKey="Trehub S">S. E. Trehub</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Trimmer, C G" uniqKey="Trimmer C">C. G. Trimmer</name>
</author>
<author>
<name sortKey="Cuddy, L L" uniqKey="Cuddy L">L. L. Cuddy</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Waaramaa, T" uniqKey="Waaramaa T">T. Waaramaa</name>
</author>
<author>
<name sortKey="Alku, P" uniqKey="Alku P">P. Alku</name>
</author>
<author>
<name sortKey="Laukka Nen, A M" uniqKey="Laukka Nen A">A.-M. Laukka-nen</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Waaramaa, T" uniqKey="Waaramaa T">T. Waaramaa</name>
</author>
<author>
<name sortKey="Kankare, E" uniqKey="Kankare E">E. Kankare</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Waaramaa, T" uniqKey="Waaramaa T">T. Waaramaa</name>
</author>
<author>
<name sortKey="Laukkanen, A M" uniqKey="Laukkanen A">A.-M. Laukkanen</name>
</author>
<author>
<name sortKey="Airas, M" uniqKey="Airas M">M. Airas</name>
</author>
<author>
<name sortKey="Alku, P" uniqKey="Alku P">P. Alku</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Waaramaa, T" uniqKey="Waaramaa T">T. Waaramaa</name>
</author>
<author>
<name sortKey="Laukkanen, A M" uniqKey="Laukkanen A">A.-M. Laukkanen</name>
</author>
<author>
<name sortKey="Alku, P" uniqKey="Alku P">P. Alku</name>
</author>
<author>
<name sortKey="V Yrynen, E" uniqKey="V Yrynen E">E. Väyrynen</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Waaramaa M Ki Kulmala, T" uniqKey="Waaramaa M Ki Kulmala T">T. Waaramaa-Mäki-Kulmala</name>
</author>
</analytic>
</biblStruct>
</listBibl>
</div1>
</back>
</TEI>
<pmc article-type="research-article">
<pmc-dir>properties open_access</pmc-dir>
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">Front Psychol</journal-id>
<journal-id journal-id-type="iso-abbrev">Front Psychol</journal-id>
<journal-id journal-id-type="publisher-id">Front. Psychol.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Psychology</journal-title>
</journal-title-group>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="pmid">23801972</article-id>
<article-id pub-id-type="pmc">3689256</article-id>
<article-id pub-id-type="doi">10.3389/fpsyg.2013.00344</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research Article</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Waaramaa</surname>
<given-names>Teija</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Leisiö</surname>
<given-names>Timo</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Communication Media and Theatre, University of Tampere</institution>
<country>Tampere, Finland</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>School of Social Sciences and Humanities, University of Tampere</institution>
<country>Tampere, Finland</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Anjali Bhatara, Université Paris Descartes, France</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Michihiko Koeda, University of Glasgow, UK; Åsa Abelin, University of Gothenburg, Sweden</p>
</fn>
<corresp id="fn001">*Correspondence: Teija Waaramaa, School of Communication Media and Theatre, University of Tampere, Kalevantie 4, Tampere 33014, Finland e-mail:
<email xlink:type="simple">teija.waaramaa@uta.fi</email>
</corresp>
<fn fn-type="other" id="fn002">
<p>This article was submitted to Frontiers in Emotion Science, a specialty of Frontiers in Psychology.</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>21</day>
<month>6</month>
<year>2013</year>
</pub-date>
<pub-date pub-type="collection">
<year>2013</year>
</pub-date>
<volume>4</volume>
<elocation-id>344</elocation-id>
<history>
<date date-type="received">
<day>26</day>
<month>2</month>
<year>2013</year>
</date>
<date date-type="accepted">
<day>27</day>
<month>5</month>
<year>2013</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright © 2013 Waaramaa and Leisiö.</copyright-statement>
<copyright-year>2013</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/3.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in other forums, provided the original authors and source are credited and subject to any copyright notices concerning any third-party graphics etc.</license-p>
</license>
</permissions>
<abstract>
<p>The present study focused on voice quality and the perception of the basic emotions from speech samples in cross-cultural conditions. It was examined whether voice quality, cultural, or language background, age, or gender were related to the identification of the emotions. Professional actors (n2) and actresses (n2) produced non-sense sentences (n32) and protracted vowels (n8) expressing the six basic emotions, interest, and a neutral emotional state. The impact of musical interests on the ability to distinguish between emotions or valence (on an axis positivity – neutrality – negativity) from voice samples was studied. Listening tests were conducted on location in five countries: Estonia, Finland, Russia, Sweden, and the USA with 50 randomly chosen participants (25 males and 25 females) in each country. The participants (total
<italic>N</italic>
= 250) completed a questionnaire eliciting their background information and musical interests. The responses in the listening test and the questionnaires were statistically analyzed. Voice quality parameters and the share of the emotions and valence identified correlated significantly with each other for both genders. The percentage of emotions and valence identified was clearly above the chance level in each of the five countries studied, however, the countries differed significantly from each other for the identified emotions and the gender of the speaker. The samples produced by females were identified significantly better than those produced by males. Listener's age was a significant variable. Only minor gender differences were found for the identification. Perceptual confusion in the listening test between emotions seemed to be dependent on their similar voice production types. Musical interests tended to have a positive effect on the identification of the emotions. The results also suggest that identifying emotions from speech samples may be easier for those listeners who share a similar language or cultural background with the speaker.</p>
</abstract>
<kwd-group>
<kwd>voice quality</kwd>
<kwd>expression</kwd>
<kwd>perception of emotions</kwd>
<kwd>valence</kwd>
<kwd>musical interests</kwd>
<kwd>cross-cultural</kwd>
</kwd-group>
<counts>
<fig-count count="1"></fig-count>
<table-count count="11"></table-count>
<equation-count count="0"></equation-count>
<ref-count count="65"></ref-count>
<page-count count="13"></page-count>
<word-count count="10147"></word-count>
</counts>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Basic emotions are thought to be universal in their manifestation since they are considered to be phylogenetic, evolutionary-survival related affects (Izard,
<xref ref-type="bibr" rid="B23">2007</xref>
). The vocal expression and perception of these emotions tend to be based firstly on genetically inherited, and secondly on culturally learnt elements (Matsumoto et al.,
<xref ref-type="bibr" rid="B40">2002</xref>
). Also, the expression and perception of emotions expressed by music tends to be affected by both inherited characteristics and by cultural learning (Morrison and Demorest,
<xref ref-type="bibr" rid="B41">2009</xref>
), and even by individual preferences, e.g., a piece of music may emotionally move one person but not another (Cross,
<xref ref-type="bibr" rid="B9">2001</xref>
). In this paper it is hypothesized that the origin of speech and temporal experiences such as emotional and musical expressions are linked together in the evolution (Juslin and Laukka,
<xref ref-type="bibr" rid="B24">2003a</xref>
). According to Richman (
<xref ref-type="bibr" rid="B45">2001</xref>
) “in the beginning speech and music making were one and the same: they were collective, real-time repetitions of formulaic sequences.” Moreover, Thompson et al. (
<xref ref-type="bibr" rid="B57">2004</xref>
) have suggested that “it seems unlikely that human evolution led to duplicate mechanisms for associating pitch and temporal cues with emotions.”</p>
<p>In voice research, voice quality is traditionally defined as the coloring of the speaker's voice (Laver,
<xref ref-type="bibr" rid="B33">1980</xref>
), and in a narrower sense, as a combination of voice source (the air flow and vocal fold vibration), and filter functions (the vocal tract and formant frequencies) (Fant,
<xref ref-type="bibr" rid="B12">1970</xref>
). The amount of subglottal air pressure and adduction of the vocal folds in the glottis determine the phonation type, whether it is hyperfunctional or hypofunctional. In a hyperfunctional phonation type the spectral slope is flatter and there is more energy and stronger overtones in the high frequency area than in a hypofunctional phonation type, where the slope is steeper and the overtones are weaker (Gauffin and Sundberg,
<xref ref-type="bibr" rid="B17">1989</xref>
). Hyperfunctional phonation type is perceived as pressed voice quality and hypofunctional as breathy voice quality. Perceptual interpretations of the voice quality may either clarify or blur the meaning of the message, or change the whole information sent by a speaker.</p>
<p>Similarly to music, vocal expressions always have a fundamental frequency (F0) (excluding whisper), intensity (sound pressure level, SPL), and duration. These are the traditional parameters studied from the voice quality in emotional expressions. As sound is transmitted via vibrating objects there is no music without movement (Cross,
<xref ref-type="bibr" rid="B9">2001</xref>
; Levitin and Tirovolas,
<xref ref-type="bibr" rid="B35">2009</xref>
), and this connection between sound and movement tends to be evolutionarily based (Liberman,
<xref ref-type="bibr" rid="B38">1981</xref>
; Liberman and Mattingly,
<xref ref-type="bibr" rid="B37">1985</xref>
; Rizzolatti et al.,
<xref ref-type="bibr" rid="B46">1996</xref>
). As in voice production, the air pressure from the lungs makes the vocal folds vibrate, and without this action there is no vocal sound. According to the motor control theory and also the more recent theory of the mirror neurons speech is said to be understood rather in terms of its production than from the characteristics of the acoustic cues (Liberman,
<xref ref-type="bibr" rid="B38">1981</xref>
; Liberman and Mattingly,
<xref ref-type="bibr" rid="B37">1985</xref>
; Gentilucci and Corballis,
<xref ref-type="bibr" rid="B18">2006</xref>
). In turn, the acoustic cues are connected to the physiological principles, and are the carriers of the emotional content of speech (see e.g., Juslin and Laukka,
<xref ref-type="bibr" rid="B25">2003b</xref>
).</p>
<p>Human vocal communication inevitably conveys emotional messages – whether intended or not. Cultural differences do occur in humans in spite of the genetically based similarities in the expression and perception of the basic emotions (Matsumoto et al.,
<xref ref-type="bibr" rid="B40">2002</xref>
; Abelin,
<xref ref-type="bibr" rid="B1">2004</xref>
). The cultural differentiation in music seems to occur by the end of the first year of life (Hannon and Trehub,
<xref ref-type="bibr" rid="B19">2005</xref>
; Belin et al.,
<xref ref-type="bibr" rid="B7">2011</xref>
), and the cultural conventions of the music are learnt by the age of five (Trehub,
<xref ref-type="bibr" rid="B58">2003</xref>
; Hannon and Trehub,
<xref ref-type="bibr" rid="B19">2005</xref>
).</p>
<p>Typical of music, always based on harmonic relations between tones, are the rules (syntaxes) which govern the ways a tune is allowed to be composed. These rules are local and they deal with various alternating combinations of 1, 2, 3, 4, 5, or 6 tones (Leisiö and Ebeling,
<xref ref-type="bibr" rid="B34">2010</xref>
). Typicality creates expectations and predictions of the characteristics of the musical sounds in a particular culture (Levitin and Tirovolas,
<xref ref-type="bibr" rid="B35">2009</xref>
). However, the three basic elements of musical expression, frequency, intensity, and duration are not culture-specific as such.</p>
<p>There also appear to be similarities in the musical emotional expressions between cultures, e.g., emotional content of happy, sad, and fearful Western music has been reported to be recognized clearly above chance level by African listeners (Fritz et al.,
<xref ref-type="bibr" rid="B15">2009</xref>
). Balkwill and Thompson (
<xref ref-type="bibr" rid="B5">1999</xref>
) studied the perception of emotions in Western and Indian music and suggested that listeners are sensitive to unfamiliar tonal systems.</p>
<p>However, recognition of the emotions is more demanding in the absence of the familiar perceptual cues. This was also verified by Scherer et al. (
<xref ref-type="bibr" rid="B49">2001</xref>
), who conducted an extensive research project on the perception of vocal emotional utterances in seven countries, in Europe, Indonesia, and the USA. The vocal language-free portrayals used were produced by German professional actors, who expressed four emotions and a neutral emotional state. The emotions were perceived with 66% accuracy across countries. However, as the dissimilarities between the languages increased the accuracy of the perception decreased. As a result, the researchers stated that culture and language specific patterns may have an influence on the decoding processes of emotional vocal portrayals.</p>
<p>Sauter et al. (
<xref ref-type="bibr" rid="B47">2009</xref>
) studied perception of English and Himba non-verbal vocalizations representing basic emotions. Their results showed that listeners from both groups could identify the emotions, however, better accuracy was achieved when the producer and the listener were from the same culture.</p>
<p>Similar results were reported by Koeda et al. (
<xref ref-type="bibr" rid="B27">2013</xref>
) in a recent investigation of non-verbal “ah” affect bursts. The vocalizations were produced by French-Canadian actors. Canadian and Japanese participants served as listeners. It was found that the Canadian listeners recognized the emotions expressed, both positive and negative, more accurately than did the Japanese listeners.</p>
<p>Thompson et al. (
<xref ref-type="bibr" rid="B57">2004</xref>
), and Lima and Castro (
<xref ref-type="bibr" rid="B39">2011</xref>
) investigated whether music training assists speech prosody decoding. The researchers concluded that music training may facilitate the recognition of the emotional content of speech. Trimmer and Cuddy (
<xref ref-type="bibr" rid="B59">2008</xref>
) came to somewhat opposite conclusion. They reported that music training does not seem to be linked to the ability to recognize emotional speech prosody. Instead, emotional intelligence may predict sensitivity to emotion recognition from speech prosody, and this tends to require different processes than those required in musical or acoustical sensitivity. Strait et al. (
<xref ref-type="bibr" rid="B55">2009</xref>
) have stated that subcortical mechanisms are involved in the auditory processing of emotions, and musical training enhances these processes: training when younger than 7 years facilitates pitch and timber perception, and duration of training impacts processing of temporal features.</p>
<p>The present study was concerned with whether the voice quality of emotional speech samples affects the identification of emotions and emotional valence (on the axis positivity – neutrality – negativity). The second aim was to investigate cross-cultural perception, whether it is dependent on language or cultural background, age, or gender. Thirdly, whether the ability to recognize emotional states is related to musical interests was studied (Thompson et al.,
<xref ref-type="bibr" rid="B57">2004</xref>
; Trimmer and Cuddy,
<xref ref-type="bibr" rid="B59">2008</xref>
; Levitin and Tirovolas,
<xref ref-type="bibr" rid="B35">2009</xref>
; Strait et al.,
<xref ref-type="bibr" rid="B55">2009</xref>
). Therefore, the participants of the listening tests were asked on a questionnaire about their subjective musical interests. Listening tests for 250 randomly chosen, volunteer participants were conducted on location in five countries: Estonia, Finland, Russia, Sweden, and the USA.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>Materials and methods</title>
<sec>
<title>Acoustic and statistical analyses</title>
<p>Emotionally loaded sentences (n32) and protracted vowels [a:], [i:], [u:] (n8) were produced by Finnish professional actors (n2) and actresses (n2). They read aloud a non-sense text
<italic>(Elki neiku ko:tsa, fonta tegoa vi:fif:i askepan:a æspa. Fis:afi: te:ki sta:ku porkas talu.)</italic>
expressing six basic emotions, namely anger, disgust, fear, joy, sadness, surprise, and a neutral emotional state. These emotions were chosen since 4–6 of them (depending on the source) are thought to be universal (Murray and Arnott,
<xref ref-type="bibr" rid="B42">1993</xref>
; Juslin and Laukka,
<xref ref-type="bibr" rid="B24">2003a</xref>
; Mithen,
<xref ref-type="bibr" rid="B40a">2006</xref>
). Interest is sometimes also listed as one of the basic emotions since it is seen as the principle force in organizing consciousness and focusing attention (Izard,
<xref ref-type="bibr" rid="B23">2007</xref>
, see also Scherer and Ellgring,
<xref ref-type="bibr" rid="B51">2007</xref>
). Based on this definition, interest was included in the present investigation. The recordings were made by Sony Sound Forge 9.0 recording and editing system, and Rode NTK microphone at a professional recording studio MediaBeat in Tampere, Finland. The speakers' distance from the microphone was 40 cm. In the tests the listeners used Sennheiser HD 598 headphones.</p>
<p>Acoustic parameters were measured with Praat Software, version 5.2.18. A frequency range of 0–5 kHz and cross-correlation were used. F0, maximum pitch, SPL, filter characteristics (formant frequencies F1, F2, F3, F4), duration, mean harmonics-to-noise ratio (HNR, dB), number of pulses, and number and degree of voice breaks were measured. HNR measures perturbation in the voice signal. The number of voice breaks is the ratio between the number of pulse distances (min 1.25) and the pitch floor. Degree of voice breaks is the ratio between the non-voiced breaks and duration of the signal. (
<ext-link ext-link-type="uri" xlink:href="http://www.fon.hum.uva.nl/praat/manual/Voice_1__Voice_breaks.html">http://www.fon.hum.uva.nl/praat/manual/Voice_1__Voice_breaks.html</ext-link>
.) The vowels were replayed consecutively to the participants in the listening tests. As the stress is always on the first syllable in Finnish language and thus carries the main communicational information, the acoustic parameters were studied only for the first [a:] vowel. Alpha ratio was calculated by subtracting the SPL in the range 50 Hz–1 kHz from SPL in the range 1–5 kHz (Frøkjær-Jensen and Prytz,
<xref ref-type="bibr" rid="B16">1973</xref>
). Alpha ratio is used to get an illustration from the spectral energy distribution.</p>
<p>Emotional valence was coded by the researcher: positive valence (interest, joy and surprise) = 1, a neutral emotional state = 0, negative valence (anger, disgust, fear and sadness) = −1.</p>
<p>Statistical analyses were conducted using Excel and IBM SPSS Statistics 19 to investigate whether the voice parameters measured correlated with the identification of the emotions or valence and whether the perception of emotions differed by country, age, gender, or self-reported musical interests.</p>
</sec>
<sec>
<title>Questionnaire and listening tests</title>
<p>Listening tests were conducted on location in five countries with different cultural and/or language backgrounds: Estonia, Finland, Russia, Sweden, and the USA. American English, Russian, and Swedish are related as members of the Indo-European linguistic family while Estonian and Finnish belong to the same Finno-Ugric language genus. As Nordic countries Finland and Sweden share a similar cultural background.</p>
<p>Fifty randomly chosen listeners in each country (25 males and 25 females × 5 countries = 250 listeners) participated in the perception test. The only criteria for participation was that the listeners were native speakers of the specific main language in each country, i.e., Estonian in Estonia, Finnish in Finland, Russian in Russia, Swedish in Sweden, and American English in the USA, and that the participants had lived most of their lives in the country. In Sweden, some of the listeners had one parent from another country, and one listener was adopted to Sweden as a baby, however, every listener spoke Swedish as their first language. The listeners were adults (18+ years old), mean age 33 years (Finland 47.5 years, Russia 34.5 years, Estonia 32 years, Sweden 27 years, and the USA 23 years).</p>
<p>The contact universities in the countries studied published the research project and called for volunteers to participate in the listening tests. Neither personal data registers nor invasive methods were used. All participants' anonymity was ensured. Consequently, no permission of the ethics committee was needed. The participants recruited in the USA were offered a course credit for participating.</p>
<p>The listening tests were conducted one by one with the listeners in an office (Finland and partly the USA), or in normal classroom conditions (Estonia, Russia, and partly the USA) or in a soundproof studio (Sweden). The researcher was alone with the listener in the test, except when a translator was needed in Russia. Listening tests are traditionally conducted in soundproof studio conditions. In the present study this was not required so as to be able to conduct the research independently using the facilities the universities in different countries were able to offer a visiting researcher. Furthermore, it was of interest to replicate the conditions of a normal social situation where people talk to each other having some random sounds around them, and nevertheless, focusing on listening to the speech and the voice of their interlocutors.</p>
<p>The participants completed a questionnaire eliciting background information, and responded to the following statements concerning their musical activities: (1) I like to listen to music. (2) It is easy for me to respond to music. (3) I am interested in singing. (4) I play a musical instrument. (5) I am interested in dancing. (6) It is easy for me to dance in the correct rhythm. (7) It is easy for me to learn a new melody. (8) Music may affect my mood. (9) Music may cause me physical reactions. The idea was to study the participants' subjective opinion about their relation to music, not to measure their activity or education in music.</p>
<p>The questionnaire and the emotion labels were translated by university teachers, either native speakers of the language (Estonian and English) or Finnish teachers in Swedish and Russian.</p>
<p>In the perception test the listener first heard four two-sentence non-sense samples, one from each speaker, and then one example of each emotion expressed by the four speakers. The researcher named the samples by the emotion before replaying them one by one in order to familiarize the listener with the speakers' voices and the vocal variation the speakers used in the emotional expressions. Next, the researcher replayed the 32 emotional nonsense sentences one by one (eight emotions × four speakers), and the listener reported orally which emotion he/she perceived. The researcher wrote down the answers given. Finally, the listener heard eight simple protracted vowel samples, two emotions from each speaker, and chose his/her answer again from the list of the eight emotions expressed. Free choice was not used. The test took about 35 min for each listener.</p>
<p>All the samples were replayed in the same random order from the researcher's computer to the participants. The listeners did not have to use any equipment while listening and answering. In unclear cases the participants were instructed to choose the nearest emotion to what they assumed to be the target. They were asked to choose neutral only when they thought there was no particular emotion expressed. The participants were instructed to answer as briefly as possible. On the other hand they were allowed to listen to a sample as many times as they felt they needed to (usually 1–2 times). They were also allowed to listen to the previous samples again so as to avoid possible order effects.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec>
<title>Voice quality</title>
<p>In vowel [a:] alpha ratio correlated significantly negatively with duration in both genders. In the sentences alpha ratio and SPL correlated significantly positively. Alpha ratio and SPL have been shown to vary together (Nordenberg and Sundberg,
<xref ref-type="bibr" rid="B43">2003</xref>
; Sundberg and Nordenberg,
<xref ref-type="bibr" rid="B56">2006</xref>
). Duration correlated negatively with F0. These results suggest that in hypofunctionally produced samples duration is longer than in hyperfunctional produced samples.</p>
<p>Significant correlations with share of identified
<italic>emotions</italic>
and voice parameters were found in both genders for mean HNR, number of voice breaks and SPL, and in females also for maximum pitch and number of pulses. Significant correlations with share of
<italic>valence</italic>
and voice parameters identified were found in both genders for number of pulses and number of voice breaks, and in males also for duration (Table
<xref ref-type="table" rid="T1">1</xref>
).</p>
<table-wrap id="T1" position="float">
<label>Table 1</label>
<caption>
<p>
<bold>Significant results for Pearson correlation between voice quality parameters and the share of identified emotions and valence (
<italic>p</italic>
< 0.05)</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>
<italic>r</italic>
-values</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Max pitch</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>N pulses</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>N voice breaks</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>HNR (dB)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Duration</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>SPL (dB)</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="7" rowspan="1">
<bold>SIGNIFICANT CORRELATIONS OF VOICE PARAMETERS WITH SHARE OF IDENTIFIED EMOTIONS AND VALENCE</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Identified emotions, male listeners</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">0.427</td>
<td align="left" rowspan="1" colspan="1">0.44</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">-0.356</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Identified emotions, female listeners</td>
<td align="left" rowspan="1" colspan="1">0.391</td>
<td align="left" rowspan="1" colspan="1">0.342</td>
<td align="left" rowspan="1" colspan="1">0.448</td>
<td align="left" rowspan="1" colspan="1">0.462</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">-0.314</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Identified valence, male listeners</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">0.395</td>
<td align="left" rowspan="1" colspan="1">0.385</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">0.353</td>
<td align="left" rowspan="1" colspan="1">ns</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Identified valence, female listeners</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">0.334</td>
<td align="left" rowspan="1" colspan="1">0.344</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">ns</td>
<td align="left" rowspan="1" colspan="1">ns</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Number of voice breaks was highest for sadness and lowest for anger, and degree of voice breaks was highest for fear and lowest for joy. The voice production type in sadness and fear tends to be more hypofunctional than in anger and joy thus, having less energy e in the higher frequency area of the spectrum.</p>
<p>The mean duration of the sentence samples was 9652 ms, and vowels 930 ms. Anger in males, and joy in females had the lowest durations for the sentences. Negative emotion of sadness followed by fear had the longest durations in both genders.</p>
</sec>
<sec>
<title>Questionnaire</title>
<p>Degree of tiredness or mood tended to be non-significant features in relation to the identification accuracy of the emotional samples. Seventeen participants reported impaired hearing (Estonia 1, Finland 8, Russia 2, Sweden 5, and the USA 1).</p>
<p>The results of the Student's
<italic>T</italic>
test showed that those who reported impaired hearing did not identify the emotions less successfully (69% identified) than those with normal hearing (70% identified).</p>
<p>The listeners were divided into two groups, under 40 years and 40+ years in order to study perceptual age differences. The younger group identified emotions with 70% accuracy and valence with 91% accuracy, and the older group emotions with 68% and valence with 90% accuracy. When Pearson correlation was studied by country, a slight negative correlation between age and the identification of the emotions was found for Finland, Russia, and the USA (Table
<xref ref-type="table" rid="T2">2</xref>
).</p>
<table-wrap id="T2" position="float">
<label>Table 2</label>
<caption>
<p>
<bold>Percentages and significance for identified emotions and valence by age and country</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Country</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Gender</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Age (years)</bold>
</th>
<th align="center" colspan="2" rowspan="1">
<bold>Identification %</bold>
</th>
</tr>
<tr>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th align="left" rowspan="1" colspan="1">
<bold>Emotion (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Valence (%)</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="5" rowspan="1">
<bold>IDENTIFIED EMOTIONS AND VALENCE BY AGE AND COUNTRIES</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Estonia</td>
<td align="left" rowspan="1" colspan="1">Male</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">71</td>
<td align="left" rowspan="1" colspan="1">94</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">77</td>
<td align="left" rowspan="1" colspan="1">94</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">70</td>
<td align="left" rowspan="1" colspan="1">92</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">73</td>
<td align="left" rowspan="1" colspan="1">92</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Finland
<xref ref-type="table-fn" rid="TN1">
<sup>*</sup>
</xref>
</td>
<td align="left" rowspan="1" colspan="1">Male</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">96</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">74</td>
<td align="left" rowspan="1" colspan="1">94</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">81</td>
<td align="left" rowspan="1" colspan="1">97</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">78</td>
<td align="left" rowspan="1" colspan="1">96</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Russia
<xref ref-type="table-fn" rid="TN1">
<sup>*</sup>
</xref>
</td>
<td align="left" rowspan="1" colspan="1">Male</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">65</td>
<td align="left" rowspan="1" colspan="1">89</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">54</td>
<td align="left" rowspan="1" colspan="1">84</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">64</td>
<td align="left" rowspan="1" colspan="1">85</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">61</td>
<td align="left" rowspan="1" colspan="1">86</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Sweden</td>
<td align="left" rowspan="1" colspan="1">Male</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">69</td>
<td align="left" rowspan="1" colspan="1">89</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">89</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">73</td>
<td align="left" rowspan="1" colspan="1">91</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">68</td>
<td align="left" rowspan="1" colspan="1">83</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">USA
<xref ref-type="table-fn" rid="TN1">
<sup>*</sup>
</xref>
</td>
<td align="left" rowspan="1" colspan="1">Male</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">64</td>
<td align="left" rowspan="1" colspan="1">88</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">48</td>
<td align="left" rowspan="1" colspan="1">77</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female</td>
<td align="left" rowspan="1" colspan="1"><40</td>
<td align="left" rowspan="1" colspan="1">61</td>
<td align="left" rowspan="1" colspan="1">86</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">40+</td>
<td align="left" rowspan="1" colspan="1">.%</td>
<td align="left" rowspan="1" colspan="1">.%</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The results are presented for the first four samples, sentences and vowels (There were no female listeners 40+ years in the USA).</p>
<fn id="TN1">
<label>*</label>
<p>Significant negative correlation with age: Finland r = −0.333, Russia r = −0.350, USA r = −0.302.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>The first statement in the questionnaire was “I like to listen to music.” By this statement the idea was to measure the degree of consumption of music. The results showed that the degree of consumption by listening to the music did not seem to be associated with the emotions or valence identified in the vocal samples (Table
<xref ref-type="table" rid="T3">3</xref>
).</p>
<table-wrap id="T3" position="float">
<label>Table 3</label>
<caption>
<p>
<bold>The results of the 250 questionnaires from the five countries studied and their relation to the identified emotions</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>“Yes” answers (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Estonia (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Finland (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Russia (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Sweden (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>USA (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Total (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Sig.</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="8" rowspan="1">
<bold>RESULTS OF THE QUESTIONNAIRES BY COUNTRY AND THEIR RELATION TO THE EMOTIONAL SAMPLES IDENTIFIED</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">I like to listen to music</td>
<td align="left" rowspan="1" colspan="1">90</td>
<td align="left" rowspan="1" colspan="1">92</td>
<td align="left" rowspan="1" colspan="1">90</td>
<td align="left" rowspan="1" colspan="1">96</td>
<td align="left" rowspan="1" colspan="1">100</td>
<td align="left" rowspan="1" colspan="1">94</td>
<td align="left" rowspan="1" colspan="1">ns</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">It is easy for me to respond to the music</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">78</td>
<td align="left" rowspan="1" colspan="1">36</td>
<td align="left" rowspan="1" colspan="1">68</td>
<td align="left" rowspan="1" colspan="1">78</td>
<td align="left" rowspan="1" colspan="1">65</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">I am interested in singing</td>
<td align="left" rowspan="1" colspan="1">24</td>
<td align="left" rowspan="1" colspan="1">28</td>
<td align="left" rowspan="1" colspan="1">10</td>
<td align="left" rowspan="1" colspan="1">48</td>
<td align="left" rowspan="1" colspan="1">18</td>
<td align="left" rowspan="1" colspan="1">26</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">I play a musical instrument</td>
<td align="left" rowspan="1" colspan="1">24</td>
<td align="left" rowspan="1" colspan="1">24</td>
<td align="left" rowspan="1" colspan="1">22</td>
<td align="left" rowspan="1" colspan="1">42</td>
<td align="left" rowspan="1" colspan="1">14</td>
<td align="left" rowspan="1" colspan="1">25</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN4">*</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">I am interested in dancing</td>
<td align="left" rowspan="1" colspan="1">16</td>
<td align="left" rowspan="1" colspan="1">18</td>
<td align="left" rowspan="1" colspan="1">6</td>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">44</td>
<td align="left" rowspan="1" colspan="1">18</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">It is easy for me to dance in the correct rhythm</td>
<td align="left" rowspan="1" colspan="1">70</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">40</td>
<td align="left" rowspan="1" colspan="1">67</td>
<td align="left" rowspan="1" colspan="1">32</td>
<td align="left" rowspan="1" colspan="1">55</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">It is easy for me to learn a new melody</td>
<td align="left" rowspan="1" colspan="1">48</td>
<td align="left" rowspan="1" colspan="1">58</td>
<td align="left" rowspan="1" colspan="1">38</td>
<td align="left" rowspan="1" colspan="1">60</td>
<td align="left" rowspan="1" colspan="1">30</td>
<td align="left" rowspan="1" colspan="1">47</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN3">**</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Music may affect my mood</td>
<td align="left" rowspan="1" colspan="1">70</td>
<td align="left" rowspan="1" colspan="1">100</td>
<td align="left" rowspan="1" colspan="1">70</td>
<td align="left" rowspan="1" colspan="1">92</td>
<td align="left" rowspan="1" colspan="1">76</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Music may cause me physical reactions</td>
<td align="left" rowspan="1" colspan="1">30</td>
<td align="left" rowspan="1" colspan="1">62</td>
<td align="left" rowspan="1" colspan="1">62</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">32</td>
<td align="left" rowspan="1" colspan="1">54</td>
<td align="left" rowspan="1" colspan="1">
<xref ref-type="table-fn" rid="TN2">***</xref>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The percentages are the “Yes” answers to the statements. “Yes” answers and the identification of the emotions and valence were significantly associated (excluding the first statement). Significance of the relationship appears on the far right.</p>
<p>Statements:</p>
<fn id="TN2">
<label>***</label>
<p>p < 0.001,</p>
</fn>
<fn id="TN3">
<label>**</label>
<p>p < 0.01,</p>
</fn>
<fn id="TN4">
<label>*</label>
<p>p < 0.05, ns) non-significant in independent samples.</p>
</fn>
<p>Student's T test for equality of means.</p>
</table-wrap-foot>
</table-wrap>
<p>The other statements concerning musical interests were statistically significantly associated with the emotions and valence identified. Those participants who reported engaging in musical interests and responding to music were compared to those who did not have a clear response to these activities. It was found that the listeners reportedly engaging in music differed significantly in the share of the identified emotions and valence from the listeners who did not report musical interests or sensitive response to music (Table
<xref ref-type="table" rid="T3">3</xref>
).</p>
<p>Females reported significantly more often than males being interested in singing while males reported playing a musical instrument significantly more often than females. When studied by country, those who were interested in singing and who played a musical instrument were most often Swedish listeners. “I am interested in dancing” was most often answered “Yes” by the US listeners.</p>
<p>Emotional states of fear, interest, and joy were most frequently associated with musical interests. Neutrality was not associated with any of the musical interests. “It is easy for me to learn a new melody” and “I am interested in singing.” were the statements which seemed to be engaged with most of the identified emotions. The statement “It is easy for me to dance in the correct rhythm” was not emotion specific and was not associated with any particular emotion (Table
<xref ref-type="table" rid="T4">4</xref>
).</p>
<table-wrap id="T4" position="float">
<label>Table 4</label>
<caption>
<p>
<bold>Emotions significantly associated with musical interests</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td align="left" colspan="3" rowspan="1">
<bold>MUSICAL ACTIVITIES ASSOCIATED WITH IDENTIFICATION OF EMOTIONS</bold>
</td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  I like to listen to music</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Sadness
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Only in the background</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  It is easy for me to respond to the music</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Fear
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  I am interested in singing</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Anger
<xref ref-type="table-fn" rid="TN5">
<sup>***</sup>
</xref>
, disgust
<xref ref-type="table-fn" rid="TN5">
<sup>***</sup>
</xref>
, fear
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not in public</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  I play a musical instrument</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Fear
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, joy
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not any more</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  I am interested in dancing</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Fear
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not in public</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  It is easy for me to dance in the correct rhythm</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Ns</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  It is easy for me to learn a new melody</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Anger
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
, disgust
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, fear
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, joy
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, surprise
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  Music may affect my mood</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Fear
<xref ref-type="table-fn" rid="TN5">
<sup>***</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, joy
<xref ref-type="table-fn" rid="TN6">
<sup>**</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" colspan="2" rowspan="1">  Music may cause me physical reactions</td>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Fear
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, interest
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
, joy
<xref ref-type="table-fn" rid="TN7">
<sup>*</sup>
</xref>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
<td rowspan="1" colspan="1"></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Statements:</p>
<fn id="TN5">
<label>***</label>
<p>p < 0.001,</p>
</fn>
<fn id="TN6">
<label>**</label>
<p>p < 0.01,</p>
</fn>
<fn id="TN7">
<label>*</label>
<p>p < 0.05, ns.) non-significant in independent samples.</p>
</fn>
<p>Student's T test for equality of means.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec>
<title>Listening tests</title>
<p>Crohnbach's alpha for the listening test by country was: Finland 0.945, Estonia 0.929, Sweden 0.905, the USA 0.874, and Russia 0.871. The results showed that the percentage of emotions and valence identified was clearly above the chance level in each of the five countries with different language and/or cultural backgrounds. A confusion matrix in percentages and numbers for the emotions identified is shown in Table
<xref ref-type="table" rid="T5">5</xref>
. Sadness and fear were the most frequently chosen emotional states for an answer, followed by neutrality. Anger was the most rarely chosen answer (Figure
<xref ref-type="fig" rid="F1">1</xref>
).</p>
<table-wrap id="T5" position="float">
<label>Table 5</label>
<caption>
<p>
<bold>The line “Count” in the confusion matrix of the emotions expressed and emotions perceived shows the numbers of answers given</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Integer =
<italic>N</italic>
</bold>
</th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th align="center" colspan="3" rowspan="1">
<bold>Emotion expressed</bold>
</th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
</tr>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Emotion perceived</bold>
</th>
<th rowspan="1" colspan="1"></th>
<th align="left" rowspan="1" colspan="1">
<bold>Joy</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Disgust</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Interest</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Neutral</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Fear</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Sadness</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Anger</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Surprise</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Total</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="11" rowspan="1">
<bold>EMOTION PERCEIVED × EMOTION EXPRESSED CROSSTABULATION</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Joy</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">
<bold>952</bold>
</td>
<td align="left" rowspan="1" colspan="1">11</td>
<td align="left" rowspan="1" colspan="1">61</td>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">6</td>
<td align="left" rowspan="1" colspan="1">0</td>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">178</td>
<td align="left" rowspan="1" colspan="1">1217</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">78.2%</td>
<td align="left" rowspan="1" colspan="1">0.9%</td>
<td align="left" rowspan="1" colspan="1">5%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">0.5%</td>
<td align="left" rowspan="1" colspan="1">0%</td>
<td align="left" rowspan="1" colspan="1">0.6%</td>
<td align="left" rowspan="1" colspan="1">14.6%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">64.2%</td>
<td align="left" rowspan="1" colspan="1">0.7%</td>
<td align="left" rowspan="1" colspan="1">4.9%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">0.4%</td>
<td align="left" rowspan="1" colspan="1">0%</td>
<td align="left" rowspan="1" colspan="1">0.6%</td>
<td align="left" rowspan="1" colspan="1">14.2%</td>
<td align="left" rowspan="1" colspan="1">11.1%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Disgust</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">45</td>
<td align="left" rowspan="1" colspan="1">
<bold>846</bold>
</td>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">37</td>
<td align="left" rowspan="1" colspan="1">62</td>
<td align="left" rowspan="1" colspan="1">35</td>
<td align="left" rowspan="1" colspan="1">293</td>
<td align="left" rowspan="1" colspan="1">51</td>
<td align="left" rowspan="1" colspan="1">1372</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">3.3%</td>
<td align="left" rowspan="1" colspan="1">61.7%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">2.7%</td>
<td align="left" rowspan="1" colspan="1">4.5%</td>
<td align="left" rowspan="1" colspan="1">2.6%</td>
<td align="left" rowspan="1" colspan="1">21.4%</td>
<td align="left" rowspan="1" colspan="1">3.7%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">3%</td>
<td align="left" rowspan="1" colspan="1">57%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">3%</td>
<td align="left" rowspan="1" colspan="1">4.2%</td>
<td align="left" rowspan="1" colspan="1">2.4%</td>
<td align="left" rowspan="1" colspan="1">23.4%</td>
<td align="left" rowspan="1" colspan="1">4.1%</td>
<td align="left" rowspan="1" colspan="1">12.6%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Interest</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">128</td>
<td align="left" rowspan="1" colspan="1">25</td>
<td align="left" rowspan="1" colspan="1">
<bold>742</bold>
</td>
<td align="left" rowspan="1" colspan="1">26</td>
<td align="left" rowspan="1" colspan="1">41</td>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">77</td>
<td align="left" rowspan="1" colspan="1">246</td>
<td align="left" rowspan="1" colspan="1">1288</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">9.9%</td>
<td align="left" rowspan="1" colspan="1">1.9%</td>
<td align="left" rowspan="1" colspan="1">57.6%</td>
<td align="left" rowspan="1" colspan="1">2%</td>
<td align="left" rowspan="1" colspan="1">3.2%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">6%</td>
<td align="left" rowspan="1" colspan="1">19.1%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">8.6%</td>
<td align="left" rowspan="1" colspan="1">1.7%</td>
<td align="left" rowspan="1" colspan="1">59.4%</td>
<td align="left" rowspan="1" colspan="1">2.1%</td>
<td align="left" rowspan="1" colspan="1">2.8%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">6.2%</td>
<td align="left" rowspan="1" colspan="1">19.7%</td>
<td align="left" rowspan="1" colspan="1">11.8%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Neutral</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">38</td>
<td align="left" rowspan="1" colspan="1">53</td>
<td align="left" rowspan="1" colspan="1">81</td>
<td align="left" rowspan="1" colspan="1">
<bold>1106</bold>
</td>
<td align="left" rowspan="1" colspan="1">37</td>
<td align="left" rowspan="1" colspan="1">17</td>
<td align="left" rowspan="1" colspan="1">106</td>
<td align="left" rowspan="1" colspan="1">10</td>
<td align="left" rowspan="1" colspan="1">1448</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">2.6%</td>
<td align="left" rowspan="1" colspan="1">3.7%</td>
<td align="left" rowspan="1" colspan="1">5.6%</td>
<td align="left" rowspan="1" colspan="1">76.4%</td>
<td align="left" rowspan="1" colspan="1">2.6%</td>
<td align="left" rowspan="1" colspan="1">1.2%</td>
<td align="left" rowspan="1" colspan="1">7.3%</td>
<td align="left" rowspan="1" colspan="1">0.7%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">2.6%</td>
<td align="left" rowspan="1" colspan="1">3.6%</td>
<td align="left" rowspan="1" colspan="1">6.5%</td>
<td align="left" rowspan="1" colspan="1">88.5%</td>
<td align="left" rowspan="1" colspan="1">2.5%</td>
<td align="left" rowspan="1" colspan="1">1.1%</td>
<td align="left" rowspan="1" colspan="1">8.5%</td>
<td align="left" rowspan="1" colspan="1">0.8%</td>
<td align="left" rowspan="1" colspan="1">13.2%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Fear</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">84</td>
<td align="left" rowspan="1" colspan="1">93</td>
<td align="left" rowspan="1" colspan="1">17</td>
<td align="left" rowspan="1" colspan="1">12</td>
<td align="left" rowspan="1" colspan="1">
<bold>1174</bold>
</td>
<td align="left" rowspan="1" colspan="1">154</td>
<td align="left" rowspan="1" colspan="1">24</td>
<td align="left" rowspan="1" colspan="1">16</td>
<td align="left" rowspan="1" colspan="1">1574</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">5.3%</td>
<td align="left" rowspan="1" colspan="1">5.9%</td>
<td align="left" rowspan="1" colspan="1">1.1%</td>
<td align="left" rowspan="1" colspan="1">0.8%</td>
<td align="left" rowspan="1" colspan="1">74.6%</td>
<td align="left" rowspan="1" colspan="1">9.8%</td>
<td align="left" rowspan="1" colspan="1">1.5%</td>
<td align="left" rowspan="1" colspan="1">1%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">5.7%</td>
<td align="left" rowspan="1" colspan="1">6.3%</td>
<td align="left" rowspan="1" colspan="1">1.4%</td>
<td align="left" rowspan="1" colspan="1">1%</td>
<td align="left" rowspan="1" colspan="1">79.2%</td>
<td align="left" rowspan="1" colspan="1">10.4%</td>
<td align="left" rowspan="1" colspan="1">1.9%</td>
<td align="left" rowspan="1" colspan="1">1.3%</td>
<td align="left" rowspan="1" colspan="1">14.4%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Sadness</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">28</td>
<td align="left" rowspan="1" colspan="1">192</td>
<td align="left" rowspan="1" colspan="1">16</td>
<td align="left" rowspan="1" colspan="1">49</td>
<td align="left" rowspan="1" colspan="1">75</td>
<td align="left" rowspan="1" colspan="1">
<bold>1266</bold>
</td>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">1634</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">1.7%</td>
<td align="left" rowspan="1" colspan="1">11.8%</td>
<td align="left" rowspan="1" colspan="1">1%</td>
<td align="left" rowspan="1" colspan="1">3%</td>
<td align="left" rowspan="1" colspan="1">4.6%</td>
<td align="left" rowspan="1" colspan="1">77.5%</td>
<td align="left" rowspan="1" colspan="1">0.4%</td>
<td align="left" rowspan="1" colspan="1">0.1%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">1.9%</td>
<td align="left" rowspan="1" colspan="1">12.9%</td>
<td align="left" rowspan="1" colspan="1">1.3%</td>
<td align="left" rowspan="1" colspan="1">3.9%</td>
<td align="left" rowspan="1" colspan="1">5.1%</td>
<td align="left" rowspan="1" colspan="1">85.4%</td>
<td align="left" rowspan="1" colspan="1">0.6%</td>
<td align="left" rowspan="1" colspan="1">0.1%</td>
<td align="left" rowspan="1" colspan="1">14.9%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Anger</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">21</td>
<td align="left" rowspan="1" colspan="1">249</td>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">18</td>
<td align="left" rowspan="1" colspan="1">36</td>
<td align="left" rowspan="1" colspan="1">6</td>
<td align="left" rowspan="1" colspan="1">
<bold>708</bold>
</td>
<td align="left" rowspan="1" colspan="1">37</td>
<td align="left" rowspan="1" colspan="1">1079</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">1.9%</td>
<td align="left" rowspan="1" colspan="1">23.1%</td>
<td align="left" rowspan="1" colspan="1">0.4%</td>
<td align="left" rowspan="1" colspan="1">1.7%</td>
<td align="left" rowspan="1" colspan="1">3.3%</td>
<td align="left" rowspan="1" colspan="1">0.6%</td>
<td align="left" rowspan="1" colspan="1">65.6%</td>
<td align="left" rowspan="1" colspan="1">3.4%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">1.4%</td>
<td align="left" rowspan="1" colspan="1">16.8%</td>
<td align="left" rowspan="1" colspan="1">0.3%</td>
<td align="left" rowspan="1" colspan="1">1.4%</td>
<td align="left" rowspan="1" colspan="1">2.4%</td>
<td align="left" rowspan="1" colspan="1">0.4%</td>
<td align="left" rowspan="1" colspan="1">56.6%</td>
<td align="left" rowspan="1" colspan="1">3%</td>
<td align="left" rowspan="1" colspan="1">9.9%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Surprise</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">187</td>
<td align="left" rowspan="1" colspan="1">14</td>
<td align="left" rowspan="1" colspan="1">326</td>
<td align="left" rowspan="1" colspan="1">0</td>
<td align="left" rowspan="1" colspan="1">52</td>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">28</td>
<td align="left" rowspan="1" colspan="1">
<bold>711</bold>
</td>
<td align="left" rowspan="1" colspan="1">1320</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">14.2%</td>
<td align="left" rowspan="1" colspan="1">1.1%</td>
<td align="left" rowspan="1" colspan="1">24.7%</td>
<td align="left" rowspan="1" colspan="1">0%</td>
<td align="left" rowspan="1" colspan="1">3.9%</td>
<td align="left" rowspan="1" colspan="1">0.2%</td>
<td align="left" rowspan="1" colspan="1">2.1%</td>
<td align="left" rowspan="1" colspan="1">53.9%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">12.6%</td>
<td align="left" rowspan="1" colspan="1">0.9%</td>
<td align="left" rowspan="1" colspan="1">26.1%</td>
<td align="left" rowspan="1" colspan="1">0%</td>
<td align="left" rowspan="1" colspan="1">3.5%</td>
<td align="left" rowspan="1" colspan="1">0.1%</td>
<td align="left" rowspan="1" colspan="1">2.2%</td>
<td align="left" rowspan="1" colspan="1">56.9%</td>
<td align="left" rowspan="1" colspan="1">12.1%</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Total</td>
<td align="left" rowspan="1" colspan="1">Count</td>
<td align="left" rowspan="1" colspan="1">1483</td>
<td align="left" rowspan="1" colspan="1">1483</td>
<td align="left" rowspan="1" colspan="1">1250</td>
<td align="left" rowspan="1" colspan="1">1250</td>
<td align="left" rowspan="1" colspan="1">1483</td>
<td align="left" rowspan="1" colspan="1">1483</td>
<td align="left" rowspan="1" colspan="1">1250</td>
<td align="left" rowspan="1" colspan="1">1250</td>
<td align="left" rowspan="1" colspan="1">
<bold>10932</bold>
</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion perceived</td>
<td align="left" rowspan="1" colspan="1">13.6%</td>
<td align="left" rowspan="1" colspan="1">13.6%</td>
<td align="left" rowspan="1" colspan="1">11.4%</td>
<td align="left" rowspan="1" colspan="1">11.4%</td>
<td align="left" rowspan="1" colspan="1">13.6%</td>
<td align="left" rowspan="1" colspan="1">13.6%</td>
<td align="left" rowspan="1" colspan="1">11.4%</td>
<td align="left" rowspan="1" colspan="1">11.4%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">% within emotion expressed</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
<td align="left" rowspan="1" colspan="1">100%</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The integers presented in bold face are the emotions identified in numbers. The other integers on the “Count” line show the numbers of confusions with the other emotions. The line “% within emotion perceived” shows the percentage of the answers given for each emotion. The line “% within emotion expressed” shows the percentage for the identification of the emotion in question.</p>
</table-wrap-foot>
</table-wrap>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption>
<p>
<bold>Results of the listening test by emotion and country including the first four samples, the sentences and the vowels</bold>
.</p>
</caption>
<graphic xlink:href="fpsyg-04-00344-g0001"></graphic>
</fig>
<p>For the first four samples the percentage of identified emotions was 59% and valence 87%, for the sentences 70% and valence 90%, and for the vowels 69 and 90% respectively. The result for the first four samples was from the 233 participants since the first 17 Finnish listeners missed these samples at the beginning of the present research project. As the accuracy percentage of identification was higher for the sentences than for the first four samples it may be assumed that familiarizing the listeners with the variations of the speakers' voices may have improved their recognition of the target emotions. The familiarizing did not seem to affect the recognition as much of emotional valence which was fairly high already before the familiarizing. Negative emotions were identified slightly more accurately than positive ones.</p>
<p>The younger listeners identified sadness significantly better than the older listeners (
<italic>p</italic>
= 0.036), who identified joy (
<italic>p</italic>
= 0.021), surprise (
<italic>p</italic>
= 0.002), and neutrality (
<italic>p</italic>
= 0.024) significantly better than the younger ones.</p>
<p>The binomial test conducted on the samples showed that 10 samples were identified with under 50% accuracy: two from the first four samples, disgust (24%), and fear (45% accuracy), from the sentences two samples of anger (13%, 31%), disgust (38%), joy (26%) interest (44%) and surprise (43%), and from the vowels joy (36%) and surprise (42% accuracy). Seven of these samples were produced by male speakers.</p>
<p>A number of confusions of the emotions perceived occurred in the listening test. Hypofunctionally produced emotions of sadness and fear were frequently confused with one another, likewise the hyperfunctionally produced negative emotions of anger and disgust. On the other hand, disgust was also confused with sadness by the listeners in Russia, Sweden, and the USA but not in Estonia or Finland. Positive emotions of joy, surprise, and interest were confused with one another, and thus the percentage for their identification was relatively low.</p>
<p>There was a tendency in the perception test that the more similar the listeners' language or cultural background was to those of the speakers', the more accurate the emotion recognition was, and conversely, the more different the language or cultural background was the less accurate the emotion recognition was. The quartiles studied by country showed that 1/4 of the listeners, e.g., in Estonia identified 55%, 1/2 identified 68%, and 1/4 at least 77% of the emotion samples. Variation was widest for Finland. The percentages fall into the quartiles roughly similarly for Estonia and Sweden, and for Russia and the USA. Finnish listeners were most accurate in the identification (Table
<xref ref-type="table" rid="T6">6</xref>
).</p>
<table-wrap id="T6" position="float">
<label>Table 6</label>
<caption>
<p>
<bold>The quartiles for the shares of identified emotions studied by country</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td align="left" colspan="4" rowspan="1">
<bold>LISTENER QUARTILES BY COUNTRY</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">% = Share of identified emotions</td>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Listener quartiles</td>
<td align="left" rowspan="1" colspan="1">25</td>
<td align="left" rowspan="1" colspan="1">50</td>
<td align="left" rowspan="1" colspan="1">75</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Estonia (%)</td>
<td align="left" rowspan="1" colspan="1">55</td>
<td align="left" rowspan="1" colspan="1">68</td>
<td align="left" rowspan="1" colspan="1">77</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Finland (%)</td>
<td align="left" rowspan="1" colspan="1">41</td>
<td align="left" rowspan="1" colspan="1">76</td>
<td align="left" rowspan="1" colspan="1">85</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Russia (%)</td>
<td align="left" rowspan="1" colspan="1">52</td>
<td align="left" rowspan="1" colspan="1">59</td>
<td align="left" rowspan="1" colspan="1">68</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Sweden (%)</td>
<td align="left" rowspan="1" colspan="1">55</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">77</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">USA (%)</td>
<td align="left" rowspan="1" colspan="1">50</td>
<td align="left" rowspan="1" colspan="1">59</td>
<td align="left" rowspan="1" colspan="1">67</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The logistic regression model (Response = Emotion identified/not identified) showed that the five countries perceived the emotions expressed significantly differently. The identification was connected to the age of the listener. The interaction effect of the speaker gender, country and emotion expressed was significant. The greatest difference between the emotion identification and the gender of the speaker was found for Estonia and Russia. There, most of the non-identified samples were produced by males. Listener's gender was non-significant (Table
<xref ref-type="table" rid="T7">7</xref>
).</p>
<table-wrap id="T7" position="float">
<label>Table 7</label>
<caption>
<p>
<bold>Test of model effect in the logistic regression model of the combined effects on the identification of the emotions</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Main effects + all significant (p < 0.05) 2-way and 3-way effects</bold>
</th>
<th align="left" colspan="3" rowspan="1">
<bold>Type III Wald df Sig. Chi-Square</bold>
</th>
</tr>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Source</bold>
</th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="4" rowspan="1">
<bold>TESTS OF MODEL EFFECTS</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">(Intercept)</td>
<td align="left" rowspan="1" colspan="1">366.2</td>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Speaker gender</td>
<td align="left" rowspan="1" colspan="1">323.9</td>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Listener gender</td>
<td align="left" rowspan="1" colspan="1">1.00</td>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">0.32</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Country</td>
<td align="left" rowspan="1" colspan="1">113.5</td>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Emotion expressed</td>
<td align="left" rowspan="1" colspan="1">629.6</td>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Age</td>
<td align="left" rowspan="1" colspan="1">19.3</td>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Speaker gender × country</td>
<td align="left" rowspan="1" colspan="1">12.3</td>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">0.015</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Speaker gender × emotion expressed</td>
<td align="left" rowspan="1" colspan="1">204.9</td>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Country × emotion expressed</td>
<td align="left" rowspan="1" colspan="1">103</td>
<td align="left" rowspan="1" colspan="1">28</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Speaker gender × country × emotion expressed</td>
<td align="left" rowspan="1" colspan="1">97.6</td>
<td align="left" rowspan="1" colspan="1">28</td>
<td align="left" rowspan="1" colspan="1">0.000</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>When studied by country, gender differences were found for only two countries: Estonian males recognized the valence of the first four samples significantly better than did the Estonian females. Swedish males recognized emotions from the sentences significantly better than did Swedish females. However, the differences between genders did not vary significantly among all five countries (Table
<xref ref-type="table" rid="T8">8</xref>
).</p>
<table-wrap id="T8" position="float">
<label>Table 8</label>
<caption>
<p>
<bold>Accuracy of the identification of the emotions and valence in percentages when studied by country and gender</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th align="center" colspan="2" rowspan="1">
<bold>Male speakers</bold>
</th>
<th align="center" colspan="2" rowspan="1">
<bold>Females speakers</bold>
</th>
</tr>
<tr>
<th rowspan="1" colspan="1"></th>
<th rowspan="1" colspan="1"></th>
<th align="left" rowspan="1" colspan="1">
<bold>Emotions (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Valence (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Emotions (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Valence (%)</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" colspan="6" rowspan="1">
<bold>EMOTION AND EMOTIONALVALENCE IDENTIFIED BY COUNTRY</bold>
</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Finland</td>
<td align="left" rowspan="1" colspan="1">Male listeners</td>
<td align="left" rowspan="1" colspan="1">70</td>
<td align="left" rowspan="1" colspan="1">93</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">95</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female listeners</td>
<td align="left" rowspan="1" colspan="1">73</td>
<td align="left" rowspan="1" colspan="1">96</td>
<td align="left" rowspan="1" colspan="1">85</td>
<td align="left" rowspan="1" colspan="1">96</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Estonia</td>
<td align="left" rowspan="1" colspan="1">Male listeners</td>
<td align="left" rowspan="1" colspan="1">63</td>
<td align="left" rowspan="1" colspan="1">91</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">97</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female listeners</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">90</td>
<td align="left" rowspan="1" colspan="1">76</td>
<td align="left" rowspan="1" colspan="1">95</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Sweden</td>
<td align="left" rowspan="1" colspan="1">Male listeners</td>
<td align="left" rowspan="1" colspan="1">64</td>
<td align="left" rowspan="1" colspan="1">86</td>
<td align="left" rowspan="1" colspan="1">73</td>
<td align="left" rowspan="1" colspan="1">93</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female listeners</td>
<td align="left" rowspan="1" colspan="1">68</td>
<td align="left" rowspan="1" colspan="1">87</td>
<td align="left" rowspan="1" colspan="1">77</td>
<td align="left" rowspan="1" colspan="1">93</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">USA</td>
<td align="left" rowspan="1" colspan="1">Male listeners</td>
<td align="left" rowspan="1" colspan="1">59</td>
<td align="left" rowspan="1" colspan="1">84</td>
<td align="left" rowspan="1" colspan="1">66</td>
<td align="left" rowspan="1" colspan="1">91</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female listeners</td>
<td align="left" rowspan="1" colspan="1">55</td>
<td align="left" rowspan="1" colspan="1">83</td>
<td align="left" rowspan="1" colspan="1">68</td>
<td align="left" rowspan="1" colspan="1">90</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Russia</td>
<td align="left" rowspan="1" colspan="1">Male listeners</td>
<td align="left" rowspan="1" colspan="1">54</td>
<td align="left" rowspan="1" colspan="1">82</td>
<td align="left" rowspan="1" colspan="1">69</td>
<td align="left" rowspan="1" colspan="1">93</td>
</tr>
<tr>
<td rowspan="1" colspan="1"></td>
<td align="left" rowspan="1" colspan="1">Female listeners</td>
<td align="left" rowspan="1" colspan="1">54</td>
<td align="left" rowspan="1" colspan="1">79</td>
<td align="left" rowspan="1" colspan="1">71</td>
<td align="left" rowspan="1" colspan="1">91</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The emotions produced by males were perceived with 62% accuracy and valence with 87% accuracy, those produced by females corresponding with 74 and 94% accuracy. The difference was statistically significant (Table
<xref ref-type="table" rid="T9">9</xref>
).</p>
<table-wrap id="T9" position="float">
<label>Table 9</label>
<caption>
<p>
<bold>Results of the emotionally loaded samples identified in percentages by gender of speakers and listeners</bold>
.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" rowspan="1" colspan="1">
<bold>Listeners</bold>
</th>
<th align="center" colspan="4" rowspan="1">
<bold>Target matching vocal samples</bold>
</th>
</tr>
<tr>
<th rowspan="1" colspan="1"></th>
<th align="center" colspan="2" rowspan="1">
<bold>Male speakers</bold>
</th>
<th align="center" colspan="2" rowspan="1">
<bold>Female speakers</bold>
</th>
</tr>
<tr>
<th rowspan="1" colspan="1"></th>
<th align="left" rowspan="1" colspan="1">
<bold>Emotions (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Valence (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Emotions (%)</bold>
</th>
<th align="left" rowspan="1" colspan="1">
<bold>Valence (%)</bold>
</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" rowspan="1" colspan="1">Males</td>
<td align="left" rowspan="1" colspan="1">62</td>
<td align="left" rowspan="1" colspan="1">87</td>
<td align="left" rowspan="1" colspan="1">74</td>
<td align="left" rowspan="1" colspan="1">94</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">Females</td>
<td align="left" rowspan="1" colspan="1">63</td>
<td align="left" rowspan="1" colspan="1">87</td>
<td align="left" rowspan="1" colspan="1">75</td>
<td align="left" rowspan="1" colspan="1">94</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<sec>
<title>Voice quality</title>
<p>Identification of valence in both genders appeared to be connected to the number of pulses and number of voice breaks. In a hyper-functional voice quality (e.g., in joy and anger) number of pulses is higher per time-domain than in a hypofunctional voice production type (e.g., Waaramaa et al.,
<xref ref-type="bibr" rid="B60">2006</xref>
). Highest number of voice breaks was found for sadness, and highest degree of voice breaks for fear which were both hypofunctionally produced utterances. Voice breaks and perturbation of voice signal tended to be discriminating features connected to the pressed/breathy voice quality in the emotional utterances.</p>
<p>The results suggest that valence is more important in the perception process of the vocal expressions and is therefore of greater communicative importance than the actual emotions. It was shown in a recent study by Waaramaa and Kankare (
<xref ref-type="bibr" rid="B61">2012</xref>
) that statistically significant differences between valences were already found on micro level emotional expressions which were calculated from the electroglottogram (EGG) signal. EGG was used to measure the contact quotient (CQ
<sub>EGG</sub>
) of the vocal folds. When the vocal folds were 25% closed (25% threshold level) significant differences were already found between valences for the CQ
<sub>EGG</sub>
. Significant gender differences have been found at the 55% threshold level (Higgins and Schulte,
<xref ref-type="bibr" rid="B20">2002</xref>
). Consequently, differences between emotions may occur only on higher threshold levels, i.e., later in the expression. Glottal behavior has likewise been reported to affect valence perception by Laukkanen et al. (
<xref ref-type="bibr" rid="B32">1997</xref>
) and (Waaramaa et al.,
<xref ref-type="bibr" rid="B63">2008</xref>
,
<xref ref-type="bibr" rid="B62">2010</xref>
).) Thus, from the communicative perspective, expression of valence seems to precede the expression of gender or the actual emotion in speech samples.</p>
<p>Formant frequencies measured in vowel [a:] did not show significant differences between emotions in the present material. Nor was it expected for F1 and F2, since they are determined by the vowel expressed. Instead, in earlier investigations F3 and F4 have shown higher frequencies in positive emotions than in negative ones (Waaramaa-Mäki-Kulmala,
<xref ref-type="bibr" rid="B64">2009</xref>
). This was also the case in the present material, but not significant. Waaramaa et al. (
<xref ref-type="bibr" rid="B60">2006</xref>
) studied synthesized vowel [a:] samples with raised, lowered and removed third formant frequency (F3) and valence perception from the samples. The results showed that the raised F3 frequency was perceived more often as positive than the other samples. It was concluded that samples with sufficient energy in the high frequency area of F3 may affect perception of positive valence from a signal.</p>
<p>However, it has been suggested by Laukkanen et al. (
<xref ref-type="bibr" rid="B31">2008</xref>
) that at least valence – if not actual emotions – can be perceived from emotional expressions even with several vocal cues eliminated (see also Waaramaa et al.,
<xref ref-type="bibr" rid="B60">2006</xref>
). This concurs with the idea of motor control and mirror neuron theory that speech can be understood rather in terms of its production than from the characteristics of the acoustic cues (see Introduction in this paper). Thus, general acoustic patterns for emotions can be only roughly presented.</p>
</sec>
<sec>
<title>Questionnaire</title>
<p>Language differences emerged when the original Finnish questionnaire was translated into Russian and Swedish. It occurred that the statement “It is easy for me to respond to music.” was translated into Russian in such a way that the grammatical subject (me) was changed into the object (on me): “Music has a strong effect on me.” It can be speculated whether this has had an effect on the answering to this statement since the percentage of the “Yes” answers was about 50% less in Russia than in the other countries. Another problem with the translation occurred when the Finnish word for “anger,”
<italic>viha</italic>
was translated into Swedish as
<italic>hat</italic>
, “hate” instead of its correct equivalent of
<italic>ilska</italic>
, “ill temper,” “anger.” This problem was explained to the last 1/4 of the participants in the listening test in Sweden.</p>
<p>The statements “It is easy for me to learn a new melody” and “I am interested in singing” were connected to most of the identified emotions. This may partly refer to the underlying intonation of the speech (melody recognition) and partly to the similarities of vocally produced utterances recognized by those who were interested in singing which is also a form of vocal expressions.</p>
<p>Most of those listeners who were interested in singing and who played a musical instrument were Swedish listeners. This result may be affected by the fact that the listening tests were conducted with help from the Music Acoustics Group at KTH, Royal Institute of Technology in Stockholm. Thus, many of the participants were involved with music through their professions, studies or hobbies. In this respect, the participants in the other countries studied may have been more heterogeneous than those in Sweden.</p>
</sec>
<sec>
<title>Listening test</title>
<p>The results of the present study showed that the percentage of the emotion identification and valence was clearly above the chance level in each of the five countries with different language and/or cultural backgrounds. Gender had no role in the perception of emotions or valence between the five countries studied. This result concurs with the findings by Koeda et al. (
<xref ref-type="bibr" rid="B27">2013</xref>
). Yet individual differences may be significant.</p>
<p>The speakers of the voice samples spoke Finnish as their native language, hence they read the non-sense text aloud using the Finnish prosody. This may be the reason why the Finnish listeners scored highest on the identified samples. A similar result was reported by Scherer et al. (
<xref ref-type="bibr" rid="B49">2001</xref>
) and by Abelin and Allwood (
<xref ref-type="bibr" rid="B4">2000</xref>
). Matsumoto et al. (
<xref ref-type="bibr" rid="B40">2002</xref>
) and Abelin (
<xref ref-type="bibr" rid="B1">2004</xref>
) have suggested that interpretation of prosody is easier for native speakers of the language in question. Abelin (
<xref ref-type="bibr" rid="B1">2004</xref>
) also has stated that the prosody of emotional expression is always related to the particular language spoken, and never occurs in isolation (see also Iversen et al.,
<xref ref-type="bibr" rid="B22">2008</xref>
). Thus, the Finnish listeners were at an advantage in the perception test as they obviously recognized the prosody more easily than the other listeners in the other countries, and could connect the prosody to the linguistic expressions even without meaningful words used. Finnish listeners perceived most rarely neutrality and most frequently joy and interest – but also disgust when compared to the other countries.</p>
<p>In their earlier study Schirmer and Kotz (
<xref ref-type="bibr" rid="B52">2002</xref>
) used event-related potentials (ERP) to study how their participants judged the valence of the prosody of a German verb and the emotional meaning of the word.</p>
<p>Interaction between emotional prosody and word meaning was found in females but not in males. Males appeared to process the meaning and the emotional prosody independently of each other. The researchers also argued that females are faster and more accurate in judging emotional information than males (Schirmer and Kotz,
<xref ref-type="bibr" rid="B52">2002</xref>
; Schirmer et al.,
<xref ref-type="bibr" rid="B53">2002</xref>
, see also Besson et al.,
<xref ref-type="bibr" rid="B8">2002</xref>
; Imaizumi et al.,
<xref ref-type="bibr" rid="B21">2004</xref>
; Fecteau et al.,
<xref ref-type="bibr" rid="B13">2005</xref>
; Schirmer and Simpson,
<xref ref-type="bibr" rid="B54">2008</xref>
). In the present investigation non-sense utterances were used. Thus there was no meaning in the words. However, gender differences were not studied here by ERP, consequently, it can only be stated that no gender differences in the accuracy of the emotion or valence perception were found. This concurs with the findings combining brain evolution, gender differences, and music (Falk,
<xref ref-type="bibr" rid="B11">2000</xref>
).</p>
<p>The perceptual confusion of the three positive emotions interest, joy, and surprise may indicate that from the evolutionary-survival perspective it may not have been crucial to distinguish between these emotions. The emotional state of joy was poorly recognized Scherer et al. (
<xref ref-type="bibr" rid="B49">2001</xref>
) have reported similar results for joy Sauter et al. (
<xref ref-type="bibr" rid="B47">2009</xref>
) have stated that communication of positive emotions may be restricted to the members of the same social or cultural group and function as consolidation of that group.</p>
<p>Identification of anger was not particularly accurate in the present study. This may be in part due to the chosen expression types by the speakers. They tended to express more cold anger than hot anger or rage. Hot anger is undoubtedly easier to identify than cold anger. One reason for not using hot anger was that the expressions had to meet the quality criteria set by the software programs in order to conduct the acoustic analyses. Further, perception of anger (Ekman,
<xref ref-type="bibr" rid="B10">2004</xref>
; Abelin,
<xref ref-type="bibr" rid="B2">2008a</xref>
,
<xref ref-type="bibr" rid="B3">b</xref>
) and disgust (Banse and Scherer,
<xref ref-type="bibr" rid="B6">1996</xref>
) may be more dominated by the visual than auditive information. However, the negative emotions of anger and disgust have been reported to be confused in visual perception tests as well (Matsumoto et al.,
<xref ref-type="bibr" rid="B40">2002</xref>
). Matsumoto et al. have suggested that the semantics of these emotions is similar and they share the elicitors of the emotion. Also, it may be easier to distinguish between positive and negative emotions (i.e., to identify valence) than between emotions which share the same valence, e.g., two negative emotions (Thompson et al.,
<xref ref-type="bibr" rid="B57">2004</xref>
). Moreover, Koeda et al. (
<xref ref-type="bibr" rid="B27">2013</xref>
) have reported significant cross-cultural differences in the perception of anger, disgust, and fear.</p>
<p>In the present study, the emotional state of fear tended to be well recognized from the auditive characteristics (see also, Abelin,
<xref ref-type="bibr" rid="B2">2008a</xref>
,
<xref ref-type="bibr" rid="B3">b</xref>
). However, fear was frequently confused with sadness, obviously due to the similarities in their acoustic cues and the large number of voice breaks they shared. These negative emotions tended to be more irregularly expressed than the positive emotions (see also Juslin and Laukka,
<xref ref-type="bibr" rid="B24">2003a</xref>
). Accordingly, Kotlyar and Morozov,
<xref ref-type="bibr" rid="B28">1976</xref>
, see also Scherer,
<xref ref-type="bibr" rid="B48">1995</xref>
) have reported longer pauses between syllables and shorter syllable duration for fear than for the other emotions in the European opera singing tradition they studied. The confusion of sadness and fear concurs with the results of an earlier study by Scherer et al. (
<xref ref-type="bibr" rid="B49">2001</xref>
). Nevertheless, sadness and fear were well recognized: the two emotions together yielded 82% accuracy and valence 94% accuracy.</p>
<p>Laukka and Juslin (
<xref ref-type="bibr" rid="B30">2007</xref>
) and Lima and Castro (
<xref ref-type="bibr" rid="B39">2011</xref>
) have stated that recognition, especially of negative emotions, tends begin to change during middle age. In the present study a negative correlation was found between age and the emotions identified for Finland, Russia, and the USA. Young listeners have been reported to be more accurate than older listeners at recognizing disgust, fear and anger from speech samples (ibid.). This was also seen in the present results. Negative emotion of sadness was significantly better recognized by young listeners, and positive emotions of joy and surprise, and additionally neutrality were significantly better recognized by old listeners. Moreover, the US participants were the youngest listeners and they chose disgust most frequently as an answer to the sentences.</p>
<p>From the evolutionary-survival and reproduction viewpoints it may be important for young people to be able to recognize negative emotions. Additionally, sadness may be an emotion which strengthens the bond between the members of the community. An accurate identification of positive emotions may imply older people's higher tolerance or understanding for the less serious features.</p>
<p>As some of the US listeners were offered course credit for participating in the present test, it may be speculated whether they were completely volunteers or not, and if on the one hand the willingness, or on the other hand the advantage gained, was the “real” motive for participating. Either way, it may have had an effect on the US results.</p>
<p>Even though the speakers were professionals, significant differences occurred in the perception of emotions expressed. It must be stressed that the samples produced by one actress were easiest to recognize throughout the countries, and this may explain the bias in the results of the perception. Coincidentally, somewhat problematic differences in the vocal samples used have also been reported previously (Scherer et al.,
<xref ref-type="bibr" rid="B50">1991</xref>
). Speaker gender has previously been reported to have a significant effect on the identification of emotions (Koeda et al.,
<xref ref-type="bibr" rid="B27">2013</xref>
). Several studies of the vocal characteristics of emotional expressions have also shown that individual differences are significant (e.g., Ladd et al.,
<xref ref-type="bibr" rid="B29">1985</xref>
).</p>
<p>Whether actor portrayals should or should not be used in emotion research has frequently been discussed. Utterances produced by actors are claimed to be stereotypical and controlled, not genuine expressions. However, in such claims genuine is never defined. This raises another question about how genuine (or pure) our emotions are in “real life” as they are mixed in our minds with other ongoing emotions quite randomly and individually (see Izard,
<xref ref-type="bibr" rid="B23">2007</xref>
). Do we know how a pure single emotion always needs to be manifested by all humans? However, the emotional samples of the present study were fairly well recognized by the listeners. Thus, there must have been some cues, either universal or cultural, which the listeners thought they recognized as expressing the specific emotional states. A number of authorities, cultural, and social systems control and regulate our social and emotional behavior, competence, and skills (Banse and Scherer,
<xref ref-type="bibr" rid="B6">1996</xref>
; Sauter et al.,
<xref ref-type="bibr" rid="B47">2009</xref>
). To have social competence or skills requires subjective control. Thus, it does not seem reasonable to claim that in “real life” emotions are uncontrolled and hence, “genuine.” It seems rather that in “real social life” emotional expressions are restricted and socialized to fit the commonly accepted norms, rules, and limits of the particular society. Consequently, it may sometimes be difficult to interpret the emotional message if the verbal and non-verbal signals are ambiguous. The expressions produced by an actor may thus be more simple and clear as he only uses those vocal cues which are necessary to convey the target emotion. This in turn, may lack realistic situational constraints (Scherer and Ellgring,
<xref ref-type="bibr" rid="B51">2007</xref>
).</p>
</sec>
<sec>
<title>Vocal emotions and music</title>
<p>Humans tend to remember better the general structure of the melody line, i.e., the contour than the exact sizes of individual intervals between tones (Levitin and Tirovolas,
<xref ref-type="bibr" rid="B36">2010</xref>
). The prosodic contour of an utterance may underlie the significance of a musical phrase or proto-musical behavior (Cross,
<xref ref-type="bibr" rid="B9">2001</xref>
). According to Panksepp (
<xref ref-type="bibr" rid="B44">2009/2010</xref>
) it is possible that without prosodic pre-adaptations from evolving humans music might never have emerged. Juslin and Laukka (
<xref ref-type="bibr" rid="B24">2003a</xref>
) have suggested that the emotional expressiveness of music is based on the similarities of the emotional acoustic cues in vocal expressions. Hence, emotional music and speech may engage the same neural processes (Juslin and Västfäll,
<xref ref-type="bibr" rid="B26">2008</xref>
).</p>
<p>In the present investigation, the positive emotions were expressed with fewer voice breaks and in a more rhythmical manner than the negative emotions. Speaking in a friendly manner has been shown to carry more melodic characteristics than speaking in an unfriendly way (Fónagy,
<xref ref-type="bibr" rid="B14">1981</xref>
). Motherese, the speech directed to babies is also melodic and rhythmic (Trehub,
<xref ref-type="bibr" rid="B58">2003</xref>
). Melodicity has suggested to be a third dimension apart from pitch and time. Melodicity is defined as “the perceptual response to the higher or lower degree of regularity/continuity/predictability of the fundamental frequency curve within each syllable” (Fónagy,
<xref ref-type="bibr" rid="B14">1981</xref>
). Melodicity can also be used as a means in identifying the emotion. One male listener in the present study explained how he perceived the emotional samples as melodies and based on the melody he decided which emotion he heard. His identification was exceptionally accurate.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>Conclusion</title>
<p>Identification of emotions from speech samples tended to be affected by voice quality and by a similar language and/or cultural background. Hence, vocal non-verbal communication affects interpretation of emotions even in the absence of language. It tends to be interpreted differently by speakers of different languages. Musical interests facilitate distinguishing between emotions.</p>
<p>Finally, it has to be stated that all the five countries studied are culturally relatively close to each other. In the future study a clearly different culture representing a totally different language background should be included in the comparison of the countries. This culture and language will be Arabic in Egypt.</p>
<sec>
<title>Conflict of interest statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack>
<p>First of all the authors express their special gratitude to the participants in the listening tests in Finland, Estonia, Sweden, USA, and Russia, and the contact persons who made the listening tests possible: Director, Dr. Pille Pruulmann-Vengerfeldt and the staff, Institute of Journalism and Communication, University of Tartu, Tartu, Estonia; Professor Sten Ternström and his students Ragnar Schön and Evert Lagerberg, the Music Acoustics group, KTH, Royal Institute of Technology, Stockholm, Sweden; Assistant Professor Graham D. Bodie and Dr. Christopher C. Gearhart, Department of Communication Studies, Louisiana State University, LA, USA; and Director, Dr. Pavel Skrelin and Tatiana Chukaeva, Department of Phonetics, Saint Petersburg State University, Saint Petersburg, Russia. The authors would also like to thank Hanna-Mari Puuska M.Sc. for statistical analyses, Virginia Mattila M. A. for language correction of the manuscript, and the translators for translating the questionnaire. This study was supported by the Academy of Finland (grant no. 1139321).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Abelin</surname>
<given-names>Å</given-names>
</name>
</person-group>
(
<year>2004</year>
).
<article-title>Cross-cultural multimodal interpretation of emotional expressions – an experimental study of Spanish and Swedish</article-title>
, in
<source>Proceedings of Speech Prosody. ISCA, March 23–26</source>
,
<publisher-loc>Naran</publisher-loc>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Abelin</surname>
<given-names>Å.</given-names>
</name>
</person-group>
(
<year>2008a</year>
).
<article-title>Anger or fear? Cross-cultural multimodal interpretations of emotional expressions</article-title>
, in
<source>Emotions in the Human Voice</source>
, Vol. 1, ed.
<person-group person-group-type="editor">
<name>
<surname>Izdebski</surname>
<given-names>K.</given-names>
</name>
</person-group>
(
<publisher-loc>San Diego</publisher-loc>
:
<publisher-name>Plural Publishing</publisher-name>
),
<fpage>65</fpage>
<lpage>73</lpage>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Abelin</surname>
<given-names>Å.</given-names>
</name>
</person-group>
(
<year>2008b</year>
).
<article-title>Seeing glee but hearing fear? Emotional McGurk effect in Swedish</article-title>
, in
<source>Proceedings of Speech Prosody. May 6–9</source>
,
<publisher-loc>Campinas</publisher-loc>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Abelin</surname>
<given-names>Å.</given-names>
</name>
<name>
<surname>Allwood</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<year>2000</year>
).
<article-title>Cross linguistic interpretation of emotional prosody</article-title>
, in
<source>ISCA ITRW Workshop on Speech and Emotion, September 5–7</source>
,
<publisher-loc>Newcastle</publisher-loc>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Balkwill</surname>
<given-names>L.-L.</given-names>
</name>
<name>
<surname>Thompson</surname>
<given-names>W. F.</given-names>
</name>
</person-group>
(
<year>1999</year>
).
<article-title>A cross-cultural investigation of the perception of emotion in music: psychophysical and cultural cues</article-title>
.
<source>Music Percept</source>
.
<volume>17</volume>
,
<fpage>43</fpage>
<lpage>64</lpage>
<pub-id pub-id-type="doi">10.2307/40285811</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Banse</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
</person-group>
(
<year>1996</year>
).
<article-title>Acoustic profiles in vocal emotion expression</article-title>
.
<source>J. Pers. Soc. Psychol</source>
.
<volume>70</volume>
,
<fpage>614</fpage>
<lpage>636</lpage>
<pub-id pub-id-type="doi">10.1037/0022-3514.70.3.614</pub-id>
<pub-id pub-id-type="pmid">8851745</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Belin</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Bestelmeyer</surname>
<given-names>P. E. G.</given-names>
</name>
<name>
<surname>Latinus</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Watson</surname>
<given-names>R.</given-names>
</name>
</person-group>
(
<year>2011</year>
).
<article-title>Understanding voice perception</article-title>
.
<source>Br. J. Psychol</source>
.
<volume>102</volume>
,
<fpage>711</fpage>
<lpage>725</lpage>
<pub-id pub-id-type="doi">10.1111/j.2044-8295.2011.02041.x</pub-id>
<pub-id pub-id-type="pmid">21988380</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Besson</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Magne</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Schön</surname>
<given-names>D.</given-names>
</name>
</person-group>
(
<year>2002</year>
).
<article-title>Emotional prosody: sex differences in sensitivity to speech melody</article-title>
.
<source>Trends Cogn. Sci. (Regul. Ed.)</source>
<volume>6</volume>
,
<fpage>405</fpage>
<lpage>407</lpage>
<pub-id pub-id-type="doi">10.1016/S1364-6613(02)01975-7</pub-id>
<pub-id pub-id-type="pmid">12413566</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cross</surname>
<given-names>I.</given-names>
</name>
</person-group>
(
<year>2001</year>
).
<article-title>Music, cognition, culture and evolution</article-title>
.
<source>Ann. N. Y. Acad. Sci</source>
.
<volume>930</volume>
,
<fpage>28</fpage>
<lpage>42</lpage>
<pub-id pub-id-type="doi">10.1111/j.1749-6632.2001.tb05723.x</pub-id>
<pub-id pub-id-type="pmid">11458835</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ekman</surname>
<given-names>P.</given-names>
</name>
</person-group>
(
<year>2004</year>
).
<source>Emotions Revealed Recognizing Faces and Feelings to Improve Communication and Emo-tional Life</source>
.
<publisher-loc>New York</publisher-loc>
:
<publisher-name>Owl Books.</publisher-name>
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Falk</surname>
<given-names>D.</given-names>
</name>
</person-group>
(
<year>2000</year>
).
<article-title>Hominid brain evolution and the origins of music</article-title>
,in
<source>The Origins of Music</source>
, eds
<person-group person-group-type="editor">
<name>
<surname>Wallin</surname>
<given-names>N. L.</given-names>
</name>
<name>
<surname>Merker</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>S.</given-names>
</name>
</person-group>
(
<publisher-loc>Cambridge</publisher-loc>
:
<publisher-name>The MIT Press</publisher-name>
),
<fpage>197</fpage>
<lpage>216</lpage>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Fant</surname>
<given-names>G.</given-names>
</name>
</person-group>
(
<year>1970</year>
).
<source>Acoustic Theory of Speech Production. With Calculations Based on X-ray Studies of Russian Articulations</source>
,
<edition>2nd Edn.</edition>
<publisher-loc>The Hague</publisher-loc>
:
<publisher-name>Mouton.</publisher-name>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fecteau</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Armony</surname>
<given-names>J. L.</given-names>
</name>
<name>
<surname>Yves</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Belin</surname>
<given-names>P.</given-names>
</name>
</person-group>
(
<year>2005</year>
).
<article-title>Judgment of emotional nonlinguistic vocalization: age-related differences</article-title>
.
<source>Appl. Neurophysiol</source>
.
<volume>12</volume>
,
<fpage>40</fpage>
<lpage>48</lpage>
<pub-id pub-id-type="doi">10.1207/s15324826an1201_7</pub-id>
<pub-id pub-id-type="pmid">15788222</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Fónagy</surname>
<given-names>I.</given-names>
</name>
</person-group>
(
<year>1981</year>
).
<article-title>Emotions, voice music</article-title>
, in
<source>Research Aspects on Singing, Royal Swedish Academy of Music</source>
, Vol. 33, ed.
<person-group person-group-type="editor">
<name>
<surname>Sundberg</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<publisher-loc>Budapest</publisher-loc>
:
<publisher-name>Akadèmiai kiado</publisher-name>
),
<fpage>51</fpage>
<lpage>79</lpage>
[Originally in: Fónagy, I., and Magdics, K. (1967). A beszéd dallama].</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fritz</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Jentschke</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gosselin</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Sammler</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Peretz</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Turner</surname>
<given-names>R.</given-names>
</name>
<etal></etal>
</person-group>
(
<year>2009</year>
).
<article-title>Universal recognition of three basic emotions in music</article-title>
.
<source>Curr. Biol</source>
.
<volume>19</volume>
,
<fpage>573</fpage>
<lpage>576</lpage>
<pub-id pub-id-type="doi">10.1016/j.cub.2009.02.058</pub-id>
<pub-id pub-id-type="pmid">19303300</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Frøkjær-Jensen</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Prytz</surname>
<given-names>S.</given-names>
</name>
</person-group>
(
<year>1973</year>
).
<article-title>Registration of voice quality</article-title>
.
<source>Brüel Kjœr Tech. Rev</source>
.
<volume>3</volume>
,
<fpage>3</fpage>
<lpage>17</lpage>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gauffin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sundberg</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<year>1989</year>
).
<article-title>Spectral correlates of glottal voice source waveform characteristics</article-title>
.
<source>J. Speech Hear. Res</source>
.
<volume>32</volume>
,
<fpage>556</fpage>
<lpage>565</lpage>
<pub-id pub-id-type="pmid">2779199</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gentilucci</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Corballis</surname>
<given-names>M. C.</given-names>
</name>
</person-group>
(
<year>2006</year>
).
<article-title>From manual gesture to speech: a gradual transition</article-title>
.
<source>Neurosci. Biobehav. Rev</source>
.
<volume>30</volume>
,
<fpage>949</fpage>
<lpage>960</lpage>
<pub-id pub-id-type="doi">10.1016/j.neubiorev.2006.02.004</pub-id>
<pub-id pub-id-type="pmid">16620983</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hannon</surname>
<given-names>E. E.</given-names>
</name>
<name>
<surname>Trehub</surname>
<given-names>S. E.</given-names>
</name>
</person-group>
(
<year>2005</year>
).
<article-title>Metrical categories in infancy and adulthood</article-title>
.
<source>Psychol. Sci</source>
.
<volume>16</volume>
,
<fpage>48</fpage>
<lpage>55</lpage>
<pub-id pub-id-type="doi">10.1111/j.0956-7976.2005.00779.x</pub-id>
<pub-id pub-id-type="pmid">15660851</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Higgins</surname>
<given-names>M. B.</given-names>
</name>
<name>
<surname>Schulte</surname>
<given-names>L.</given-names>
</name>
</person-group>
(
<year>2002</year>
).
<article-title>Gender differences in vocal fold contact computed from electroglottographic signals: the influence of measurement threshold</article-title>
.
<source>J. Acoust. Soc. Am</source>
.
<volume>111</volume>
,
<fpage>1865</fpage>
<lpage>1871</lpage>
<pub-id pub-id-type="doi">10.1121/1.1456517</pub-id>
<pub-id pub-id-type="pmid">12002870</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Imaizumi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Homma</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ozawa</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Maruishi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Muranaka</surname>
<given-names>H.</given-names>
</name>
</person-group>
(
<year>2004</year>
).
<article-title>Gender differences in the functional organization of the brain for emotional prosody processing</article-title>
.
<source>ISCA Speech Prosody</source>
March 23–26,
<pub-id pub-id-type="doi">10.1080/1357650X.2011.586702</pub-id>
<pub-id pub-id-type="pmid">22973808</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Iversen</surname>
<given-names>J. R.</given-names>
</name>
<name>
<surname>Patel</surname>
<given-names>A. D.</given-names>
</name>
<name>
<surname>Ohgushi</surname>
<given-names>K.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>Perception of rhythmic grouping depends on auditory experience</article-title>
.
<source>J. Acoust. Soc. Am</source>
.
<volume>124</volume>
,
<fpage>2263</fpage>
<lpage>2271</lpage>
<pub-id pub-id-type="doi">10.1121/1.2973189</pub-id>
<pub-id pub-id-type="pmid">19062864</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Izard</surname>
<given-names>C. E.</given-names>
</name>
</person-group>
(
<year>2007</year>
).
<article-title>Basic emotions, natural kinds, emotion schemas, and a new paradigm</article-title>
.
<source>Perspect. Psychol. Sci</source>
.
<volume>2</volume>
,
<fpage>260</fpage>
<lpage>280</lpage>
<pub-id pub-id-type="doi">10.1111/j.1745-6916.2007.00044.x</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Juslin</surname>
<given-names>P. N.</given-names>
</name>
<name>
<surname>Laukka</surname>
<given-names>P.</given-names>
</name>
</person-group>
(
<year>2003a</year>
).
<article-title>Communication of emotions in vocal expression and music performance: different channels, same code?</article-title>
<source>Psychol. Bull</source>
.
<volume>129</volume>
,
<fpage>770</fpage>
<lpage>814</lpage>
<pub-id pub-id-type="doi">10.1037/0033-2909.129.5.770</pub-id>
<pub-id pub-id-type="pmid">12956543</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Juslin</surname>
<given-names>P. N.</given-names>
</name>
<name>
<surname>Laukka</surname>
<given-names>P.</given-names>
</name>
</person-group>
(
<year>2003b</year>
).
<article-title>Emotional expression in speech and music, evidence of cross-modal similarities</article-title>
.
<source>Ann. N. Y. Acad. Sci</source>
.
<volume>1000</volume>
,
<fpage>279</fpage>
<lpage>282</lpage>
<pub-id pub-id-type="doi">10.1196/annals.1280.025</pub-id>
<pub-id pub-id-type="pmid">14766637</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Juslin</surname>
<given-names>P. N.</given-names>
</name>
<name>
<surname>Västfäll</surname>
<given-names>D.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>Emotional responses to music: the need to consider underlying mechanisms</article-title>
.
<source>Behav. Brain Sci</source>
.
<volume>31</volume>
,
<fpage>559</fpage>
<lpage>575</lpage>
<pub-id pub-id-type="doi">10.1017/S0140525X08005293</pub-id>
<pub-id pub-id-type="pmid">18826699</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Koeda</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Belin</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Hama</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Masuda</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Matsuura</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Okubo</surname>
<given-names>Y.</given-names>
</name>
</person-group>
(
<year>2013</year>
).
<article-title>Cross-cultural differences in the processing of non-verbal affective vocalizations by Japanese and Canadian listeners</article-title>
.
<source>Front. Psychol</source>
.
<volume>4</volume>
:
<issue>105</issue>
<pub-id pub-id-type="doi">10.3389/fpsyg.2013.00105</pub-id>
<pub-id pub-id-type="pmid">23516137</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kotlyar</surname>
<given-names>G. M.</given-names>
</name>
<name>
<surname>Morozov</surname>
<given-names>V. P.</given-names>
</name>
</person-group>
(
<year>1976</year>
).
<article-title>Acoustical correlates of the emotional content of vocalized speech</article-title>
.
<source>Sov. Phys. Acoust</source>
.
<volume>22</volume>
,
<fpage>208</fpage>
<lpage>211</lpage>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ladd</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Silverman</surname>
<given-names>K. E. A.</given-names>
</name>
<name>
<surname>Tolk-mitt</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Bergmann</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
</person-group>
(
<year>1985</year>
).
<article-title>Evidence for the independent function of intonation contour type, voice quality, and F0 range in signaling speaker affect</article-title>
.
<source>J. Acoust. Soc. Am</source>
.
<volume>78</volume>
,
<fpage>435</fpage>
<lpage>444</lpage>
<pub-id pub-id-type="doi">10.1121/1.392466</pub-id>
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Laukka</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Juslin</surname>
<given-names>P. N.</given-names>
</name>
</person-group>
(
<year>2007</year>
).
<article-title>Similar patterns of age-related differences in emotion recognition from speech and music</article-title>
.
<source>Motiv. Emot</source>
.
<volume>31</volume>
,
<fpage>182</fpage>
<lpage>191</lpage>
<pub-id pub-id-type="doi">10.1007/s11031-007-9063-z</pub-id>
</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Laukkanen</surname>
<given-names>A.-M.</given-names>
</name>
<name>
<surname>Alku</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Airas</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Waaramaa</surname>
<given-names>T.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>The role of voice in the expression and perception of emotions</article-title>
, in
<source>Emotions in the Human Voice</source>
,
<volume>Vol. I</volume>
, ed.
<person-group person-group-type="editor">
<name>
<surname>Izdebski</surname>
<given-names>K.</given-names>
</name>
</person-group>
(
<publisher-loc>San Diego</publisher-loc>
:
<publisher-name>Plural Publishing</publisher-name>
),
<fpage>171</fpage>
<lpage>184</lpage>
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Laukkanen</surname>
<given-names>A.-M.</given-names>
</name>
<name>
<surname>Vilkman</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Alku</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Oksanen</surname>
<given-names>H.</given-names>
</name>
</person-group>
(
<year>1997</year>
).
<article-title>On the perception of emotions in speech: the role of voice quality</article-title>
.
<source>Logoped. Phoniatr. Vocol</source>
.
<volume>22</volume>
,
<fpage>157</fpage>
<lpage>168</lpage>
<pub-id pub-id-type="doi">10.3109/14015439709075330</pub-id>
</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Laver</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<year>1980</year>
).
<source>The Phonetic Description of Voice Quality</source>
.
<publisher-loc>Cambridge</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Leisiö</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ebeling</surname>
<given-names>M.</given-names>
</name>
</person-group>
(
<year>2010</year>
).
<article-title>Neu-ronal basis of seeker tone theory. A mathematical solution</article-title>
.
<source>Musiikki</source>
<volume>2</volume>
,
<fpage>60</fpage>
<lpage>71</lpage>
</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Levitin</surname>
<given-names>D. J.</given-names>
</name>
<name>
<surname>Tirovolas</surname>
<given-names>A. K.</given-names>
</name>
</person-group>
(
<year>2009</year>
).
<article-title>Current advances in the cognitive neuroscience of music</article-title>
.
<source>J. Cogn. Neurosci</source>
.
<volume>1156</volume>
,
<fpage>211</fpage>
<lpage>231</lpage>
<pub-id pub-id-type="doi">10.1111/j.1749-6632.2009.04417.x</pub-id>
<pub-id pub-id-type="pmid">19338510</pub-id>
</mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Levitin</surname>
<given-names>D. J.</given-names>
</name>
<name>
<surname>Tirovolas</surname>
<given-names>A. K.</given-names>
</name>
</person-group>
(
<year>2010</year>
).
<source>Music cognition and perception. In: sage encyclopedia of perception</source>
.
<publisher-loc>Thousand Oaks, CA</publisher-loc>
:
<publisher-name>Sage Publications</publisher-name>
,
<fpage>599</fpage>
<lpage>606</lpage>
</mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liberman</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Mattingly</surname>
<given-names>I. G.</given-names>
</name>
</person-group>
(
<year>1985</year>
).
<article-title>The motor theory of speech perception revised</article-title>
.
<source>Cognition</source>
<volume>21</volume>
,
<fpage>1</fpage>
<lpage>36</lpage>
<pub-id pub-id-type="doi">10.1016/0010-0277(85)90021-6</pub-id>
<pub-id pub-id-type="pmid">4075760</pub-id>
</mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Liberman</surname>
<given-names>A. M.</given-names>
</name>
</person-group>
(
<year>1981</year>
).
<source>On finding that speech is special. Haskins Laboratories</source>
.
<publisher-loc>New Haven</publisher-loc>
:
<publisher-name>Status Report on Speech Research SR-67/68</publisher-name>
</mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lima</surname>
<given-names>C. F.</given-names>
</name>
<name>
<surname>Castro</surname>
<given-names>S. L.</given-names>
</name>
</person-group>
(
<year>2011</year>
).
<article-title>Speaking to the trained ear: musical expertise enhances the recognition of emotions in speech prosody</article-title>
.
<source>Emotion</source>
<volume>11</volume>
,
<fpage>1021</fpage>
<lpage>1031</lpage>
<pub-id pub-id-type="doi">10.1037/a0024521</pub-id>
<pub-id pub-id-type="pmid">21942696</pub-id>
</mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Matsumoto</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Franklin</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>J.-W.</given-names>
</name>
<name>
<surname>Rogers</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Tatani</surname>
<given-names>H.</given-names>
</name>
</person-group>
(
<year>2002</year>
).
<article-title>Cultural influences on the expression and perception of emotion</article-title>
, in
<source>Handbook of International and Inter-cultural Communication</source>
,
<edition>2nd Edn</edition>
, eds
<person-group person-group-type="editor">
<name>
<surname>Gudykunst</surname>
<given-names>W. B.</given-names>
</name>
<name>
<surname>Mody</surname>
<given-names>B.</given-names>
</name>
</person-group>
(
<publisher-loc>Thousand Oaks</publisher-loc>
:
<publisher-name>Sage Publications, Inc</publisher-name>
),
<fpage>107</fpage>
<lpage>125</lpage>
</mixed-citation>
</ref>
<ref id="B40a">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Mithen</surname>
<given-names>S.</given-names>
</name>
</person-group>
(
<year>2006</year>
).
<source>The Singing Neanderthals. The Origin of Music, Language, Mind and Body</source>
.
<publisher-loc>Great Britain</publisher-loc>
:
<publisher-name>Phoenix</publisher-name>
</mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Morrison</surname>
<given-names>S. J.</given-names>
</name>
<name>
<surname>Demorest</surname>
<given-names>S. M.</given-names>
</name>
</person-group>
(
<year>2009</year>
).
<article-title>Cultural constraints on music perception and cognition</article-title>
, in
<source>Progress in Brain Research. Cultural Neuroscience: Cultural Influences on Brain Function</source>
,
<volume>Vol. 178</volume>
, ed.
<person-group person-group-type="editor">
<name>
<surname>Chiao</surname>
<given-names>J. Y.</given-names>
</name>
</person-group>
(
<publisher-loc>Elsevier</publisher-loc>
),
<fpage>67</fpage>
<lpage>77</lpage>
<pub-id pub-id-type="doi">10.1016/S0079-6123(09)17805-6</pub-id>
</mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Murray</surname>
<given-names>I. R.</given-names>
</name>
<name>
<surname>Arnott</surname>
<given-names>J. L.</given-names>
</name>
</person-group>
(
<year>1993</year>
).
<article-title>Toward the simulation of emotion in synthetic speech: a review of the literature on human vocal emotion</article-title>
.
<source>J. Acoust. Soc. Am</source>
.
<volume>93</volume>
,
<fpage>1097</fpage>
<lpage>1108</lpage>
<pub-id pub-id-type="doi">10.1121/1.405558</pub-id>
<pub-id pub-id-type="pmid">8445120</pub-id>
</mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nordenberg</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Sundberg</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<year>2003</year>
).
<article-title>Effect on LTAS of vocal loudness variation</article-title>
.
<source>TMH-QPSR</source>
<volume>45</volume>
,
<fpage>93</fpage>
<lpage>100</lpage>
</mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Panksepp</surname>
<given-names>J.</given-names>
</name>
</person-group>
(
<year>2009/2010</year>
).
<article-title>The emotional antecedents to the evolution of music and language</article-title>
.
<source>Musicœ Sci.</source>
<volume>13</volume>
,
<fpage>229</fpage>
<lpage>259</lpage>
<pub-id pub-id-type="doi">10.1177/1029864909013002111</pub-id>
</mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Richman</surname>
<given-names>B.</given-names>
</name>
</person-group>
(
<year>2001</year>
).
<article-title>How music fixed “nonsense” into significant formulas: on rhythm, repetition and meaning</article-title>
, in
<source>The Origins of Music</source>
, eds
<person-group person-group-type="editor">
<name>
<surname>Wallin</surname>
<given-names>N. L.</given-names>
</name>
<name>
<surname>Merker</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>S.</given-names>
</name>
</person-group>
(
<publisher-loc>Cambridge</publisher-loc>
:
<publisher-name>Massachusetts Institute of Technology</publisher-name>
),
<fpage>301</fpage>
<lpage>314</lpage>
</mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rizzolatti</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Fadiga</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Gallese</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Fogassi</surname>
<given-names>L.</given-names>
</name>
</person-group>
(
<year>1996</year>
).
<article-title>Premotor cortex and the recognition of motor actions</article-title>
.
<source>Brain Res. Cogn. Brain Res</source>
.
<volume>3</volume>
,
<fpage>131</fpage>
<lpage>141</lpage>
<pub-id pub-id-type="doi">10.1016/0926-6410(95)00038-0</pub-id>
<pub-id pub-id-type="pmid">8713554</pub-id>
</mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sauter</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Eisner</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Ekman</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Scott</surname>
<given-names>S. K.</given-names>
</name>
</person-group>
(
<year>2009</year>
).
<article-title>Cross-culötural recognition of basic emotions through nonverbal emotional vocalizations</article-title>
.
<source>Proc. Natl. Acad. Sci. U.S.A</source>
.
<volume>107</volume>
,
<fpage>2408</fpage>
<lpage>2412</lpage>
<pub-id pub-id-type="doi">10.1073/pnas.0908239106</pub-id>
<pub-id pub-id-type="pmid">20133790</pub-id>
</mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
</person-group>
(
<year>1995</year>
).
<article-title>Expression of emotion in voice and music</article-title>
.
<source>J. Voice</source>
<volume>9</volume>
,
<fpage>235</fpage>
<lpage>248</lpage>
<pub-id pub-id-type="doi">10.1016/S0892-1997(05)80231-0</pub-id>
<pub-id pub-id-type="pmid">8541967</pub-id>
</mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
<name>
<surname>Banse</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Wall-bott</surname>
<given-names>H. G.</given-names>
</name>
</person-group>
(
<year>2001</year>
).
<article-title>Emotion inferences from vocal expression correlate across languages and cultures</article-title>
.
<source>J. Cross Cult. Psychol</source>
.
<volume>32</volume>
,
<fpage>76</fpage>
<lpage>92</lpage>
<pub-id pub-id-type="doi">10.1177/0022022101032001009</pub-id>
</mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
<name>
<surname>Banse</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Wallbott</surname>
<given-names>H. G.</given-names>
</name>
<name>
<surname>Goldbeck</surname>
<given-names>T.</given-names>
</name>
</person-group>
(
<year>1991</year>
).
<article-title>Vocal cues in emotion encoding and decoding</article-title>
.
<source>Motiv. Emot</source>
.
<volume>15</volume>
,
<fpage>123</fpage>
<lpage>148</lpage>
<pub-id pub-id-type="doi">10.1007/BF00995674</pub-id>
</mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scherer</surname>
<given-names>K. R.</given-names>
</name>
<name>
<surname>Ellgring</surname>
<given-names>H.</given-names>
</name>
</person-group>
(
<year>2007</year>
).
<article-title>Multimodal expression of emotion: affect programs or componential appraisal patterns?</article-title>
<source>Emotion</source>
<volume>7</volume>
,
<fpage>158</fpage>
<lpage>171</lpage>
<pub-id pub-id-type="doi">10.1037/1528-3542.7.1.158</pub-id>
<pub-id pub-id-type="pmid">17352571</pub-id>
</mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Schirmer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Kotz</surname>
<given-names>S. A.</given-names>
</name>
</person-group>
(
<year>2002</year>
).
<article-title>Sex differentiates the Stroop-effect in emotional speech: ERP evidence</article-title>
, in
<source>Proceedings of the 1st Speech Prosody Conference</source>
, (
<publisher-loc>France</publisher-loc>
:
<publisher-name>Aix-en-Provence</publisher-name>
),
<fpage>631</fpage>
<lpage>634</lpage>
<pub-id pub-id-type="doi">10.1162/089892903322598102</pub-id>
</mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schirmer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Kotz</surname>
<given-names>S. A.</given-names>
</name>
<name>
<surname>Friederici</surname>
<given-names>A.</given-names>
</name>
</person-group>
(
<year>2002</year>
).
<article-title>Sex differentiates the role of emotional prosody during word processing</article-title>
.
<source>Brain Res. Cogn. Brain Res</source>
.
<volume>14</volume>
,
<fpage>228</fpage>
<lpage>233</lpage>
<pub-id pub-id-type="doi">10.1016/S0926-6410(02)00108-8</pub-id>
<pub-id pub-id-type="pmid">12067695</pub-id>
</mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Schirmer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Simpson</surname>
<given-names>E.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>Brain correlates of vocal emotional processing in men and women</article-title>
, in
<source>Emotions in the Human Voice</source>
, ed.
<person-group person-group-type="editor">
<name>
<surname>Izdebski</surname>
<given-names>K.</given-names>
</name>
</person-group>
(
<publisher-loc>San Diego</publisher-loc>
:
<publisher-name>Plural Publishing</publisher-name>
),
<fpage>75</fpage>
<lpage>86</lpage>
<pub-id pub-id-type="doi">10.1037/1528-3542.6.3.406</pub-id>
</mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Strait</surname>
<given-names>D. L.</given-names>
</name>
<name>
<surname>Kraus</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Skoe</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Ashley</surname>
<given-names>R.</given-names>
</name>
</person-group>
(
<year>2009</year>
).
<article-title>Musical experience and neural efficiency – effects of training on subcortical processing of vocal expressions of emotion</article-title>
.
<source>Eur. J. Neurosci</source>
.
<volume>29</volume>
,
<fpage>661</fpage>
<lpage>668</lpage>
<pub-id pub-id-type="doi">10.1111/j.1460-9568.2009.06617.x</pub-id>
<pub-id pub-id-type="pmid">19222564</pub-id>
</mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sundberg</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Nordenberg</surname>
<given-names>M.</given-names>
</name>
</person-group>
(
<year>2006</year>
).
<article-title>Effects of vocal loudness variation on spectrum balance as reflected by the alpha measure of long-term-average spectra of speech</article-title>
.
<source>J. Acoust. Soc. Am</source>
.
<volume>120</volume>
,
<fpage>453</fpage>
<lpage>457</lpage>
<pub-id pub-id-type="doi">10.1121/1.2208451</pub-id>
<pub-id pub-id-type="pmid">16875241</pub-id>
</mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thompson</surname>
<given-names>W. E.</given-names>
</name>
<name>
<surname>Schellenberg</surname>
<given-names>E. G.</given-names>
</name>
<name>
<surname>Husain</surname>
<given-names>G.</given-names>
</name>
</person-group>
(
<year>2004</year>
).
<article-title>Decoding speech prosody: do music lessons help?</article-title>
<source>Emotion</source>
<volume>4</volume>
,
<fpage>46</fpage>
<lpage>64</lpage>
<pub-id pub-id-type="doi">10.1037/1528-3542.4.1.46</pub-id>
<pub-id pub-id-type="pmid">15053726</pub-id>
</mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Trehub</surname>
<given-names>S. E.</given-names>
</name>
</person-group>
(
<year>2003</year>
).
<article-title>The developmental origins of musicality</article-title>
.
<source>Nat. Neurosci</source>
.
<volume>6</volume>
,
<fpage>669</fpage>
<lpage>673</lpage>
<pub-id pub-id-type="doi">10.1038/nn1084</pub-id>
<pub-id pub-id-type="pmid">12830157</pub-id>
</mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Trimmer</surname>
<given-names>C. G.</given-names>
</name>
<name>
<surname>Cuddy</surname>
<given-names>L. L.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>Emotional intelligence, not music training, predicts recognition of emotional speech prosody</article-title>
.
<source>Emotion</source>
<volume>8</volume>
,
<fpage>838</fpage>
<lpage>849</lpage>
<pub-id pub-id-type="doi">10.1037/a0014080</pub-id>
<pub-id pub-id-type="pmid">19102595</pub-id>
</mixed-citation>
</ref>
<ref id="B60">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waaramaa</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Alku</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Laukka-nen</surname>
<given-names>A.-M.</given-names>
</name>
</person-group>
(
<year>2006</year>
).
<article-title>The role of F3 in the vocal expression of emotions</article-title>
.
<source>Logoped. Phoniatr. Vocol</source>
.
<volume>31</volume>
,
<fpage>153</fpage>
<lpage>156</lpage>
<pub-id pub-id-type="doi">10.1080/14015430500456739</pub-id>
<pub-id pub-id-type="pmid">17114127</pub-id>
</mixed-citation>
</ref>
<ref id="B61">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waaramaa</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kankare</surname>
<given-names>E.</given-names>
</name>
</person-group>
(
<year>2012</year>
).
<article-title>Acoustic and EGG analyses of emotional utterances</article-title>
.
<source>Logoped. Phoniatr. Vocol</source>
. (in press). Available at:
<ext-link ext-link-type="uri" xlink:href="http://www.ncbi.nlm.nih.gov/pubmed/22587654">http://www.ncbi.nlm.nih.gov/pubmed/22587654</ext-link>
,
<pub-id pub-id-type="doi">10.3109/14015439.2012.679966</pub-id>
<pub-id pub-id-type="pmid">22587654</pub-id>
</mixed-citation>
</ref>
<ref id="B62">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waaramaa</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Laukkanen</surname>
<given-names>A.-M.</given-names>
</name>
<name>
<surname>Airas</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Alku</surname>
<given-names>P.</given-names>
</name>
</person-group>
(
<year>2010</year>
).
<article-title>Perception of emotional valences and activity levels from vowel segments of continuous speech</article-title>
.
<source>J. Voice</source>
<volume>24</volume>
,
<fpage>30</fpage>
<lpage>38</lpage>
<pub-id pub-id-type="doi">10.1016/j.jvoice.2008.04.004</pub-id>
<pub-id pub-id-type="pmid">19111438</pub-id>
</mixed-citation>
</ref>
<ref id="B63">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Waaramaa</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Laukkanen</surname>
<given-names>A.-M.</given-names>
</name>
<name>
<surname>Alku</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Väyrynen</surname>
<given-names>E.</given-names>
</name>
</person-group>
(
<year>2008</year>
).
<article-title>Mono-pitched expression of emotions in different vowels</article-title>
.
<source>Folia Phoniatr. Logop</source>
.
<volume>60</volume>
,
<fpage>249</fpage>
<lpage>255</lpage>
<pub-id pub-id-type="doi">10.1159/000151762</pub-id>
<pub-id pub-id-type="pmid">18765945</pub-id>
</mixed-citation>
</ref>
<ref id="B64">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Waaramaa-Mäki-Kulmala</surname>
<given-names>T.</given-names>
</name>
</person-group>
(
<year>2009</year>
).
<source>Emotions in Voice. Acoustic and Perceptual Analysis of Voice Quality in the Vocal Expression of Emotions</source>
.
<publisher-loc>Academic dissertation. Tampere University Press. Tampere</publisher-loc>
</mixed-citation>
</ref>
</ref-list>
<app-group>
<app id="A1">
<title>Appendix</title>
<table-wrap id="d35e5152" position="anchor">
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">Name</td>
<td align="left" rowspan="1" colspan="1">——————————————–</td>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">Age</td>
<td align="left" rowspan="1" colspan="1">——————————</td>
<td rowspan="1" colspan="1"></td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">Gender</td>
<td align="left" rowspan="1" colspan="1">1 Male</td>
<td align="left" rowspan="1" colspan="1">2 Female</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">My hearing is</td>
<td align="left" rowspan="1" colspan="1">1 Normal</td>
<td align="left" rowspan="1" colspan="1">2 Impaired</td>
<td rowspan="1" colspan="1"></td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">5</td>
<td align="left" rowspan="1" colspan="1">At the moment I am</td>
<td align="left" rowspan="1" colspan="1">1 Alert</td>
<td align="left" rowspan="1" colspan="1">2 Tired</td>
<td align="left" rowspan="1" colspan="1">3 In between these two</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">6</td>
<td align="left" rowspan="1" colspan="1">At the moment my mood is</td>
<td align="left" rowspan="1" colspan="1">1 Good</td>
<td align="left" rowspan="1" colspan="1">2 Bad</td>
<td align="left" rowspan="1" colspan="1">3 Neutral</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">I like to listen to music</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Only in the background</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">8</td>
<td align="left" rowspan="1" colspan="1">It is easy for me to respond to the music</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">9</td>
<td align="left" rowspan="1" colspan="1">I am interested in singing</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not in public</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">10</td>
<td align="left" rowspan="1" colspan="1">I play a musical instrument</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not any more</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">11</td>
<td align="left" rowspan="1" colspan="1">I am interested in dancing</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Not in public</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">12</td>
<td align="left" rowspan="1" colspan="1">It is easy for me to dance in the correct rhythm</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">13</td>
<td align="left" rowspan="1" colspan="1">It is easy for me to learn a new melody</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">14</td>
<td align="left" rowspan="1" colspan="1">Music may affect my mood</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">15</td>
<td align="left" rowspan="1" colspan="1">Music may cause me physical reactions</td>
<td align="left" rowspan="1" colspan="1">1 Yes</td>
<td align="left" rowspan="1" colspan="1">2 No</td>
<td align="left" rowspan="1" colspan="1">3 Sometimes</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec>
<title>Please choose from these emotions which of them you hear:</title>
<table-wrap id="d35e5318" position="anchor">
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">Neutral</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">Sadness</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">Fear</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">Anger</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">5</td>
<td align="left" rowspan="1" colspan="1">Disgust</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">6</td>
<td align="left" rowspan="1" colspan="1">Joy</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">7</td>
<td align="left" rowspan="1" colspan="1">Surprise</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">8</td>
<td align="left" rowspan="1" colspan="1">Interest</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</app>
</app-group>
</back>
</pmc>
<affiliations>
<list>
<country>
<li>Finlande</li>
</country>
</list>
<tree>
<country name="Finlande">
<noRegion>
<name sortKey="Waaramaa, Teija" sort="Waaramaa, Teija" uniqKey="Waaramaa T" first="Teija" last="Waaramaa">Teija Waaramaa</name>
</noRegion>
<name sortKey="Leisio, Timo" sort="Leisio, Timo" uniqKey="Leisio T" first="Timo" last="Leisiö">Timo Leisiö</name>
</country>
</tree>
</affiliations>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Wicri/Musique/explor/OperaV1/Data/Ncbi/Merge
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000B01 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Ncbi/Merge/biblio.hfd -nk 000B01 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Wicri/Musique
   |area=    OperaV1
   |flux=    Ncbi
   |étape=   Merge
   |type=    RBID
   |clé=     PMC:3689256
   |texte=   Perception of emotionally loaded vocal expressions and its connection to responses to music. A cross-cultural investigation: Estonia, Finland, Sweden, Russia, and the USA
}}

Pour générer des pages wiki

HfdIndexSelect -h $EXPLOR_AREA/Data/Ncbi/Merge/RBID.i   -Sk "pubmed:23801972" \
       | HfdSelect -Kh $EXPLOR_AREA/Data/Ncbi/Merge/biblio.hfd   \
       | NlmPubMed2Wicri -a OperaV1 

Wicri

This area was generated with Dilib version V0.6.21.
Data generation: Thu Apr 14 14:59:05 2016. Site generation: Thu Jan 4 23:09:23 2024