Serveur d'exploration sur l'opéra

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Iconicity can ground the creation of vocal symbols

Identifieur interne : 000E00 ( Ncbi/Merge ); précédent : 000D99; suivant : 000E01

Iconicity can ground the creation of vocal symbols

Auteurs : Marcus Perlman [États-Unis] ; Rick Dale [États-Unis] ; Gary Lupyan [États-Unis]

Source :

RBID : PMC:4555852

Abstract

Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative ‘vocal’ charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic ‘meaning template’ we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.


Url:
DOI: 10.1098/rsos.150152
PubMed: 26361547
PubMed Central: 4555852

Links toward previous steps (curation, corpus...)


Links to Exploration step

PMC:4555852

Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Iconicity can ground the creation of vocal symbols</title>
<author>
<name sortKey="Perlman, Marcus" sort="Perlman, Marcus" uniqKey="Perlman M" first="Marcus" last="Perlman">Marcus Perlman</name>
<affiliation wicri:level="2">
<nlm:aff id="af1">
<addr-line>Department of Psychology</addr-line>
,
<institution>University of Wisconsin-Madison</institution>
,
<addr-line>Madison, WI, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Madison, WI</wicri:regionArea>
<placeName>
<region type="state">Wisconsin</region>
</placeName>
</affiliation>
</author>
<author>
<name sortKey="Dale, Rick" sort="Dale, Rick" uniqKey="Dale R" first="Rick" last="Dale">Rick Dale</name>
<affiliation wicri:level="2">
<nlm:aff id="af2">
<addr-line>Cognitive and Information Sciences</addr-line>
,
<institution>University of California</institution>
,
<addr-line>Merced, Merced, CA, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Merced, Merced, CA</wicri:regionArea>
<placeName>
<region type="state">Californie</region>
</placeName>
</affiliation>
</author>
<author>
<name sortKey="Lupyan, Gary" sort="Lupyan, Gary" uniqKey="Lupyan G" first="Gary" last="Lupyan">Gary Lupyan</name>
<affiliation wicri:level="2">
<nlm:aff id="af1">
<addr-line>Department of Psychology</addr-line>
,
<institution>University of Wisconsin-Madison</institution>
,
<addr-line>Madison, WI, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Madison, WI</wicri:regionArea>
<placeName>
<region type="state">Wisconsin</region>
</placeName>
</affiliation>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">PMC</idno>
<idno type="pmid">26361547</idno>
<idno type="pmc">4555852</idno>
<idno type="url">http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4555852</idno>
<idno type="RBID">PMC:4555852</idno>
<idno type="doi">10.1098/rsos.150152</idno>
<date when="2015">2015</date>
<idno type="wicri:Area/Pmc/Corpus">000047</idno>
<idno type="wicri:Area/Pmc/Curation">000047</idno>
<idno type="wicri:Area/Pmc/Checkpoint">000058</idno>
<idno type="wicri:Area/Ncbi/Merge">000E00</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en" level="a" type="main">Iconicity can ground the creation of vocal symbols</title>
<author>
<name sortKey="Perlman, Marcus" sort="Perlman, Marcus" uniqKey="Perlman M" first="Marcus" last="Perlman">Marcus Perlman</name>
<affiliation wicri:level="2">
<nlm:aff id="af1">
<addr-line>Department of Psychology</addr-line>
,
<institution>University of Wisconsin-Madison</institution>
,
<addr-line>Madison, WI, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Madison, WI</wicri:regionArea>
<placeName>
<region type="state">Wisconsin</region>
</placeName>
</affiliation>
</author>
<author>
<name sortKey="Dale, Rick" sort="Dale, Rick" uniqKey="Dale R" first="Rick" last="Dale">Rick Dale</name>
<affiliation wicri:level="2">
<nlm:aff id="af2">
<addr-line>Cognitive and Information Sciences</addr-line>
,
<institution>University of California</institution>
,
<addr-line>Merced, Merced, CA, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Merced, Merced, CA</wicri:regionArea>
<placeName>
<region type="state">Californie</region>
</placeName>
</affiliation>
</author>
<author>
<name sortKey="Lupyan, Gary" sort="Lupyan, Gary" uniqKey="Lupyan G" first="Gary" last="Lupyan">Gary Lupyan</name>
<affiliation wicri:level="2">
<nlm:aff id="af1">
<addr-line>Department of Psychology</addr-line>
,
<institution>University of Wisconsin-Madison</institution>
,
<addr-line>Madison, WI, USA</addr-line>
</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<wicri:regionArea>Madison, WI</wicri:regionArea>
<placeName>
<region type="state">Wisconsin</region>
</placeName>
</affiliation>
</author>
</analytic>
<series>
<title level="j">Royal Society Open Science</title>
<idno type="e-ISSN">2054-5703</idno>
<imprint>
<date when="2015">2015</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass></textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">
<p>Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative ‘vocal’ charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic ‘meaning template’ we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.</p>
</div>
</front>
<back>
<div1 type="bibliography">
<listBibl>
<biblStruct>
<analytic>
<author>
<name sortKey="De Vos, C" uniqKey="De Vos C">C De Vos</name>
</author>
<author>
<name sortKey="Pfau, R" uniqKey="Pfau R">R Pfau</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Sandler, W" uniqKey="Sandler W">W Sandler</name>
</author>
<author>
<name sortKey="Meir, I" uniqKey="Meir I">I Meir</name>
</author>
<author>
<name sortKey="Padden, C" uniqKey="Padden C">C Padden</name>
</author>
<author>
<name sortKey="Aronoff, M" uniqKey="Aronoff M">M Aronoff</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Senghas, A" uniqKey="Senghas A">A Senghas</name>
</author>
<author>
<name sortKey="Kita, S" uniqKey="Kita S">S Kita</name>
</author>
<author>
<name sortKey="Ozyurek, A" uniqKey="Ozyurek A">A Özyürek</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Frishberg, N" uniqKey="Frishberg N">N Frishberg</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Meir, I" uniqKey="Meir I">I Meir</name>
</author>
<author>
<name sortKey="Israel, A" uniqKey="Israel A">A Israel</name>
</author>
<author>
<name sortKey="Sandler, W" uniqKey="Sandler W">W Sandler</name>
</author>
<author>
<name sortKey="Padden, Ca" uniqKey="Padden C">CA Padden</name>
</author>
<author>
<name sortKey="Aronoff, M" uniqKey="Aronoff M">M Aronoff</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Goldin Meadow, S" uniqKey="Goldin Meadow S">S Goldin-Meadow</name>
</author>
<author>
<name sortKey="Feldman, H" uniqKey="Feldman H">H Feldman</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Umiker Sebeok, J" uniqKey="Umiker Sebeok J">J Umiker-Sebeok</name>
</author>
<author>
<name sortKey="Sebeok, Ta" uniqKey="Sebeok T">TA Sebeok</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kendon, A" uniqKey="Kendon A">A Kendon</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Goldin Meadow, S" uniqKey="Goldin Meadow S">S Goldin-Meadow</name>
</author>
<author>
<name sortKey="Mcneill, D" uniqKey="Mcneill D">D McNeill</name>
</author>
<author>
<name sortKey="Singleton, J" uniqKey="Singleton J">J Singleton</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Dediu, D" uniqKey="Dediu D">D Dediu</name>
</author>
<author>
<name sortKey="Levinson, Sc" uniqKey="Levinson S">SC Levinson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kendon, A" uniqKey="Kendon A">A Kendon</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Armstrong, Df" uniqKey="Armstrong D">DF Armstrong</name>
</author>
<author>
<name sortKey="Wilcox, Se" uniqKey="Wilcox S">SE Wilcox</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Meir, I" uniqKey="Meir I">I Meir</name>
</author>
<author>
<name sortKey="Padden, C" uniqKey="Padden C">C Padden</name>
</author>
<author>
<name sortKey="Aronoff, M" uniqKey="Aronoff M">M Aronoff</name>
</author>
<author>
<name sortKey="Sandler, W" uniqKey="Sandler W">W Sandler</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Perniss, P" uniqKey="Perniss P">P Perniss</name>
</author>
<author>
<name sortKey="Vigliocco, G" uniqKey="Vigliocco G">G Vigliocco</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Taub, Sf" uniqKey="Taub S">SF Taub</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Garrod, S" uniqKey="Garrod S">S Garrod</name>
</author>
<author>
<name sortKey="Fay, N" uniqKey="Fay N">N Fay</name>
</author>
<author>
<name sortKey="Lee, J" uniqKey="Lee J">J Lee</name>
</author>
<author>
<name sortKey="Oberlander, J" uniqKey="Oberlander J">J Oberlander</name>
</author>
<author>
<name sortKey="Macleod, T" uniqKey="Macleod T">T MacLeod</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Perlman, M" uniqKey="Perlman M">M Perlman</name>
</author>
<author>
<name sortKey="Clark, N" uniqKey="Clark N">N Clark</name>
</author>
<author>
<name sortKey="Johansson Falck, M" uniqKey="Johansson Falck M">M Johansson Falck</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Imai, M" uniqKey="Imai M">M Imai</name>
</author>
<author>
<name sortKey="Kita, S" uniqKey="Kita S">S Kita</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fay, N" uniqKey="Fay N">N Fay</name>
</author>
<author>
<name sortKey="Arbib, M" uniqKey="Arbib M">M Arbib</name>
</author>
<author>
<name sortKey="Garrod, S" uniqKey="Garrod S">S Garrod</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Goldin Meadow, S" uniqKey="Goldin Meadow S">S Goldin-Meadow</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Aronoff, M" uniqKey="Aronoff M">M Aronoff</name>
</author>
<author>
<name sortKey="Meir, I" uniqKey="Meir I">I Meir</name>
</author>
<author>
<name sortKey="Sandler, W" uniqKey="Sandler W">W Sandler</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Sandler, W" uniqKey="Sandler W">W Sandler</name>
</author>
<author>
<name sortKey="Aronoff, M" uniqKey="Aronoff M">M Aronoff</name>
</author>
<author>
<name sortKey="Meir, I" uniqKey="Meir I">I Meir</name>
</author>
<author>
<name sortKey="Padden, C" uniqKey="Padden C">C Padden</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Gelb, Ij" uniqKey="Gelb I">IJ Gelb</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Sampson, G" uniqKey="Sampson G">G Sampson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Vaccari, O" uniqKey="Vaccari O">O Vaccari</name>
</author>
<author>
<name sortKey="Vaccari, Ee" uniqKey="Vaccari E">EE Vaccari</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Arbib, Ma" uniqKey="Arbib M">MA Arbib</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Corballis, Mc" uniqKey="Corballis M">MC Corballis</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Tomasello, M" uniqKey="Tomasello M">M Tomasello</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Hockett, Cf" uniqKey="Hockett C">CF Hockett</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fay, N" uniqKey="Fay N">N Fay</name>
</author>
<author>
<name sortKey="Lister, Cj" uniqKey="Lister C">CJ Lister</name>
</author>
<author>
<name sortKey="Ellison, Tm" uniqKey="Ellison T">TM Ellison</name>
</author>
<author>
<name sortKey="Goldin Meadow, S" uniqKey="Goldin Meadow S">S Goldin-Meadow</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Quine, Wv" uniqKey="Quine W">WV Quine</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Spence, C" uniqKey="Spence C">C Spence</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Winter, B" uniqKey="Winter B">B Winter</name>
</author>
<author>
<name sortKey="Marghetis, T" uniqKey="Marghetis T">T Marghetis</name>
</author>
<author>
<name sortKey="Matlock, T" uniqKey="Matlock T">T Matlock</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Ramachandran, Vs" uniqKey="Ramachandran V">VS Ramachandran</name>
</author>
<author>
<name sortKey="Hubbard, Em" uniqKey="Hubbard E">EM Hubbard</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Walker, P" uniqKey="Walker P">P Walker</name>
</author>
<author>
<name sortKey="Bremner, Jg" uniqKey="Bremner J">JG Bremner</name>
</author>
<author>
<name sortKey="Mason, U" uniqKey="Mason U">U Mason</name>
</author>
<author>
<name sortKey="Spring, J" uniqKey="Spring J">J Spring</name>
</author>
<author>
<name sortKey="Mattock, K" uniqKey="Mattock K">K Mattock</name>
</author>
<author>
<name sortKey="Slater, A" uniqKey="Slater A">A Slater</name>
</author>
<author>
<name sortKey="Johnson, Sp" uniqKey="Johnson S">SP Johnson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Hinton, L" uniqKey="Hinton L">L Hinton</name>
</author>
<author>
<name sortKey="Nichols, J" uniqKey="Nichols J">J Nichols</name>
</author>
<author>
<name sortKey="Ohala, Jj" uniqKey="Ohala J">JJ Ohala</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Nuckolls, Jb" uniqKey="Nuckolls J">JB Nuckolls</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Perniss, P" uniqKey="Perniss P">P Perniss</name>
</author>
<author>
<name sortKey="Thompson, Rl" uniqKey="Thompson R">RL Thompson</name>
</author>
<author>
<name sortKey="Vigliocco, G" uniqKey="Vigliocco G">G Vigliocco</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Ohala, Jj" uniqKey="Ohala J">JJ Ohala</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Ahlner, F" uniqKey="Ahlner F">F Ahlner</name>
</author>
<author>
<name sortKey="Zlatev, J" uniqKey="Zlatev J">J Zlatev</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Greenberg, J" uniqKey="Greenberg J">J Greenberg</name>
</author>
<author>
<name sortKey="Jenkins, J" uniqKey="Jenkins J">J Jenkins</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Klink, Rr" uniqKey="Klink R">RR Klink</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Dingemanse, M" uniqKey="Dingemanse M">M Dingemanse</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Voeltz, Efk" uniqKey="Voeltz E">EFK Voeltz</name>
</author>
<author>
<name sortKey="Kilian Hatz, C" uniqKey="Kilian Hatz C">C Kilian-Hatz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Verhoef, T" uniqKey="Verhoef T">T Verhoef</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Boersma, P" uniqKey="Boersma P">P Boersma</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kirby, S" uniqKey="Kirby S">S Kirby</name>
</author>
<author>
<name sortKey="Cornish, H" uniqKey="Cornish H">H Cornish</name>
</author>
<author>
<name sortKey="Smith, K" uniqKey="Smith K">K Smith</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Monaghan, P" uniqKey="Monaghan P">P Monaghan</name>
</author>
<author>
<name sortKey="Shillcock, Rc" uniqKey="Shillcock R">RC Shillcock</name>
</author>
<author>
<name sortKey="Christiansen, Mh" uniqKey="Christiansen M">MH Christiansen</name>
</author>
<author>
<name sortKey="Kirby, S" uniqKey="Kirby S">S Kirby</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Clark, N" uniqKey="Clark N">N Clark</name>
</author>
<author>
<name sortKey="Perlman, M" uniqKey="Perlman M">M Perlman</name>
</author>
<author>
<name sortKey="Johansson Falck, M" uniqKey="Johansson Falck M">M Johansson Falck</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Dolscheid, S" uniqKey="Dolscheid S">S Dolscheid</name>
</author>
<author>
<name sortKey="Shayan, S" uniqKey="Shayan S">S Shayan</name>
</author>
<author>
<name sortKey="Majid, A" uniqKey="Majid A">A Majid</name>
</author>
<author>
<name sortKey="Casasanto, D" uniqKey="Casasanto D">D Casasanto</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Zbikowski, Lm" uniqKey="Zbikowski L">LM Zbikowski</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Klima, E" uniqKey="Klima E">E Klima</name>
</author>
<author>
<name sortKey="Bellugi, U" uniqKey="Bellugi U">U Bellugi</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Nuckolls, Jb" uniqKey="Nuckolls J">JB Nuckolls</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Mcneill, D" uniqKey="Mcneill D">D McNeill</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kendon, A" uniqKey="Kendon A">A Kendon</name>
</author>
</analytic>
</biblStruct>
</listBibl>
</div1>
</back>
</TEI>
<pmc article-type="research-article">
<pmc-dir>properties open_access</pmc-dir>
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">R Soc Open Sci</journal-id>
<journal-id journal-id-type="iso-abbrev">R Soc Open Sci</journal-id>
<journal-id journal-id-type="publisher-id">RSOS</journal-id>
<journal-id journal-id-type="hwp">royopensci</journal-id>
<journal-title-group>
<journal-title>Royal Society Open Science</journal-title>
</journal-title-group>
<issn pub-type="epub">2054-5703</issn>
<publisher>
<publisher-name>The Royal Society Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="pmid">26361547</article-id>
<article-id pub-id-type="pmc">4555852</article-id>
<article-id pub-id-type="doi">10.1098/rsos.150152</article-id>
<article-id pub-id-type="publisher-id">rsos150152</article-id>
<article-categories>
<subj-group subj-group-type="hwp-journal-coll">
<subject>1001</subject>
<subject>14</subject>
<subject>42</subject>
<subject>70</subject>
</subj-group>
<subj-group subj-group-type="heading">
<subject>Psychology and Cognitive Neuroscience</subject>
</subj-group>
<subj-group subj-group-type="leader">
<subject>Research Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Iconicity can ground the creation of vocal symbols</article-title>
<alt-title alt-title-type="short">Iconcity Can Ground Vocal Symbols</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Perlman</surname>
<given-names>Marcus</given-names>
</name>
<xref ref-type="aff" rid="af1">1</xref>
<xref ref-type="corresp" rid="cor1"></xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Dale</surname>
<given-names>Rick</given-names>
</name>
<xref ref-type="aff" rid="af2">2</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Lupyan</surname>
<given-names>Gary</given-names>
</name>
<xref ref-type="aff" rid="af1">1</xref>
</contrib>
</contrib-group>
<aff id="af1">
<label>1</label>
<addr-line>Department of Psychology</addr-line>
,
<institution>University of Wisconsin-Madison</institution>
,
<addr-line>Madison, WI, USA</addr-line>
</aff>
<aff id="af2">
<label>2</label>
<addr-line>Cognitive and Information Sciences</addr-line>
,
<institution>University of California</institution>
,
<addr-line>Merced, Merced, CA, USA</addr-line>
</aff>
<author-notes>
<corresp id="cor1">Author for correspondence: Marcus Perlman e-mail:
<email>mperlman@wisc.edu</email>
</corresp>
</author-notes>
<pub-date pub-type="collection">
<month>8</month>
<year>2015</year>
</pub-date>
<pub-date pub-type="epub">
<day>5</day>
<month>8</month>
<year>2015</year>
</pub-date>
<pub-date pub-type="pmc-release">
<day>5</day>
<month>8</month>
<year>2015</year>
</pub-date>
<pmc-comment> PMC Release delay is 0 months and 0 days and was based on the . </pmc-comment>
<volume>2</volume>
<issue>8</issue>
<elocation-id>150152</elocation-id>
<history>
<date date-type="received">
<day>17</day>
<month>4</month>
<year>2015</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>7</month>
<year>2015</year>
</date>
</history>
<permissions>
<copyright-statement>© 2015 The Authors.</copyright-statement>
<copyright-year>2015</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Published by the Royal Society under the terms of the Creative Commons Attribution License
<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</ext-link>
, which permits unrestricted use, provided the original author and source are credited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="rsos150152.pdf"></self-uri>
<abstract>
<p>Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative ‘vocal’ charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic ‘meaning template’ we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.</p>
</abstract>
<kwd-group>
<kwd>experimental semiotics</kwd>
<kwd>iconicity</kwd>
<kwd>language evolution</kwd>
<kwd>vocalization</kwd>
</kwd-group>
<funding-group>
<award-group id="funding-1">
<funding-source>National Science Foundation
<named-content content-type="funder-id">http://dx.doi.org/10.13039/100000001</named-content>
</funding-source>
<award-id>1344279</award-id>
</award-group>
</funding-group>
<custom-meta-group>
<custom-meta>
<meta-name>cover-date</meta-name>
<meta-value>August, 2015</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1.</label>
<title>Introduction</title>
<p>In recent decades, scientists have learned a lot about how people create gestural communication systems and the circumstances under which this happens. Observations span the origins of highly sophisticated urban and rural signed languages used by deaf people and their communities [
<xref rid="RSOS150152C1" ref-type="bibr">1</xref>
<xref rid="RSOS150152C5" ref-type="bibr">5</xref>
], home sign systems devised by deaf children to communicate with hearing carers [
<xref rid="RSOS150152C6" ref-type="bibr">6</xref>
], and alternative sign languages used by hearing people under different conditions in which speech is disadvantaged or suppressed [
<xref rid="RSOS150152C7" ref-type="bibr">7</xref>
,
<xref rid="RSOS150152C8" ref-type="bibr">8</xref>
]. Experimental studies with hearing participants also show how, when prohibited from speaking, people quickly improvise more language-like means of gesturing [
<xref rid="RSOS150152C9" ref-type="bibr">9</xref>
]. In comparison, we know very little about how people create
<italic>vocal</italic>
communication systems, including the roughly 7000 languages currently spoken around the world [
<xref rid="RSOS150152C10" ref-type="bibr">10</xref>
].
<sup>
<xref ref-type="fn" rid="FN1">1</xref>
</sup>
Here, we examine the creation and development of a rudimentary vocal system in a laboratory setting. Our findings illuminate a mechanism by which iconic vocalizations can ground the creation of vocal symbols, analogous to the function of iconic gestures in gestural communication systems.</p>
<p>The study of gestural systems, including fully conventionalized signed languages, has concluded that they originate largely from spontaneously created
<italic>iconic</italic>
gestures [
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C13" ref-type="bibr">13</xref>
]. Iconicity refers to a quality of gestures and other signals that are characterized by a sense of resemblance or correspondence between their form and the meaning they express.
<sup>
<xref ref-type="fn" rid="FN2">2</xref>
</sup>
For example, in British Sign Language, the conventional sign ‘hammer’ is formed by simulating the manual action of hammering, and ‘bottle’ is articulated with a curved hand as if handling a bottle, using upward movement to show its shape [
<xref rid="RSOS150152C14" ref-type="bibr">14</xref>
]. These examples also reveal an element of metonymy that is common in iconic gestures: the option to represent concepts by their different characteristic properties, for instance, emphasizing function versus shape. Moreover, iconicity in signs and less conventionalized gestures is not limited to direct enactments and depictions. Iconicity can be highly schematic as in the American Sign Language (ASL) classifier that represents long skinny objects, often a singular person, with an upright index finger [
<xref rid="RSOS150152C15" ref-type="bibr">15</xref>
]. It can also be the source of metaphorical expressions, as in the ASL sign ‘think-penetrate’ (translated as ‘[the subject] finally got the point’), in which the signer uses her hand to depict an idea as an object, and shows it emerging from her forehead and penetrating a barrier represented by her second hand.</p>
<p>Notably, iconicity can be exhibited by signals in various modalities and media. For example, graphical media are often iconic, ranging from detailed drawings to schematic roadway signs [
<xref rid="RSOS150152C16" ref-type="bibr">16</xref>
]. Vocalizations can also be iconic in ways that are comparable to gesture and drawing. The vocal imitation of different environmental sounds is one obvious example of iconicity in vocalization (and presumably the source of conventionalized onomatopoeia in languages). But vocal iconicity can also be highly schematic, as in the elongation of the vowel in ‘a
<italic>loooong</italic>
time’ to depict the protracted temporal duration of an event [
<xref rid="RSOS150152C14" ref-type="bibr">14</xref>
]. Through metaphorical correspondence across modalities, temporal extension of the vowel can also be used to express spatial extension, as in describing the ‘
<italic>loooong</italic>
stem’ of a mushroom [
<xref rid="RSOS150152C17" ref-type="bibr">17</xref>
].</p>
<p>Iconicity is theorized to function in the creation of gestural systems by providing an initial connection between a gesture and its meaning [
<xref rid="RSOS150152C14" ref-type="bibr">14</xref>
,
<xref rid="RSOS150152C18" ref-type="bibr">18</xref>
,
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
]. Iconicity both facilitates the generation of novel forms, and it enables a perceiver to gain a sense meaning without (or at least, with less) previous learning. Building from an iconically grounded connection between form and meaning, gestures become conventionalized and systematic over the course of repeated use and transmission. For example, they become more stereotyped and categorical, and under the right circumstances, they can grow into grammatical systems of varying levels of complexity and conventionalization. For instance, in isolation from other signers, deaf children create iconic gestures that ground the development of symbolic home sign systems that they use (asymmetrically) with hearing adults [
<xref rid="RSOS150152C6" ref-type="bibr">6</xref>
,
<xref rid="RSOS150152C20" ref-type="bibr">20</xref>
]. These gestural systems develop hierarchical levels of organization, along with systematic contrast between gestures. In communities with multiple signers transmitting a growing system over multiple generations—such as in schools for the deaf or rural villages with high incidence of genetic deafness—fully conventionalized, grammatical signed languages can originate from iconic gestures in just a few generations [
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C21" ref-type="bibr">21</xref>
,
<xref rid="RSOS150152C22" ref-type="bibr">22</xref>
]. Iconic gestures can become systematic on shorter time scales too. Some fairly basic grammatical patterns, such as the use of gesture order to express thematic roles, can emerge quite quickly in the laboratory, such as when hearing adults are tasked to communicate with gestures instead of speech [
<xref rid="RSOS150152C8" ref-type="bibr">8</xref>
].</p>
<p>In addition to gestural systems, studies of communication in graphic media point to a more general process of
<italic>iconic creation</italic>
, according to which iconic signals generate the raw material for conventionalization, and thereby ground the generation of a communication system. For example, historical investigations of writing systems show that many of the world's written scripts (e.g. Sumerian, early Egyptian, ancient Chinese) originated in more detailed, iconic depictions which over history became conventionalized into an increasingly abstract code [
<xref rid="RSOS150152C23" ref-type="bibr">23</xref>
<xref rid="RSOS150152C25" ref-type="bibr">25</xref>
]. More recently, experimental evidence for iconic creation comes from laboratory studies of how graphic communication systems develop. In one set of experiments, participants produced drawings to communicate different items (e.g. Clint Eastwood, soap opera) over multiple blocks [
<xref rid="RSOS150152C16" ref-type="bibr">16</xref>
]. Participants originally produced detailed iconic drawings related to the items. Over repeated rounds, these drawings became increasingly conventionalized and symbolic, but only under conditions in which participants were able to interact with each other and provide feedback.</p>
<p>Given these basic findings in gesture and graphic media, an interesting question is the extent to which people can also generate vocal communication systems by this same process of iconic creation. On the face of it, many scholars have argued that they cannot [
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
,
<xref rid="RSOS150152C26" ref-type="bibr">26</xref>
<xref rid="RSOS150152C29" ref-type="bibr">29</xref>
]. Their arguments typically contend that gestures are special in their potential for iconicity, particularly when compared to vocalizations, because of the spatial nature of gestures and their resemblance to manual actions used for manipulating objects [
<xref rid="RSOS150152C11" ref-type="bibr">11</xref>
,
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C28" ref-type="bibr">28</xref>
]: hence, the pervasive iconicity evident in signed languages. By contrast, vocal communication systems—spoken languages in particular—are often characterized as essentially arbitrary, aside from a few marginal instances of onomatopoeia (i.e. words bearing sound–sound correspondences). For instance, Hockett reasoned that, ‘when a representation of some four-dimensional hunk of life has to be compressed into the single dimension of speech, most iconicity is necessarily squeezed out’ [[
<xref rid="RSOS150152C10" ref-type="bibr">10</xref>
], p. 275] (quoted in [
<xref rid="RSOS150152C28" ref-type="bibr">28</xref>
]). Some experimental support for this line of reasoning comes from a set of studies in which participants communicated an array of actions, objects and emotions using either gestures or non-linguistic vocalizations [
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
,
<xref rid="RSOS150152C30" ref-type="bibr">30</xref>
]. The results showed that participants performed much better with gestures, leading the authors to tout the superiority of gesture over vocalization for iconic representation. This rationale presents a challenge for scholars aiming to understand the origins of vocal communication systems and how vocal symbols come to be associated with their particular meanings (cf. [
<xref rid="RSOS150152C31" ref-type="bibr">31</xref>
]).</p>
<p>However, as hinted in the examples above, there is considerable evidence indicating that the vocal modality holds more potential for iconicity than is often realized [
<xref rid="RSOS150152C13" ref-type="bibr">13</xref>
,
<xref rid="RSOS150152C17" ref-type="bibr">17</xref>
]. It is widely recognized that vocalizations can be iconic of various environmental sounds in a directly imitative way (e.g. imitating the sounds of animals or cars), and through metonymy and metaphor, these imitative vocalizations can extend to new realms of meaning (e.g. the various senses of ‘splash’) [
<xref rid="RSOS150152C15" ref-type="bibr">15</xref>
]. In addition, the reach of iconicity in vocalization can be greatly expanded through more abstract and schematic connections to meaning. One example comes from research in cross-modal correspondences and synesthesia, which finds that people share deeply rooted associations between certain acoustic and visual dimensions, many of which arise as metaphors in spoken language [
<xref rid="RSOS150152C32" ref-type="bibr">32</xref>
<xref rid="RSOS150152C34" ref-type="bibr">34</xref>
]. For instance, people tend to represent dimensions of magnitude according to a common amodal or multimodal representation. These prothetic dimensions include acoustic properties like duration and loudness in association with visual dimensions like brightness and spatial dimensions like size and length. Thus, people are inclined to perceive equivalence between disparate stimuli as a long sound, a loud sound, a bright light, and a big or long object. Studies also show a deep association, detected in infants three to four months of age, between the acoustic dimension of pitch and both visual–spatial height (e.g. high pitch with high spatial position) and visual sharpness (e.g. high pitch with pointedness) [
<xref rid="RSOS150152C35" ref-type="bibr">35</xref>
].</p>
<p>The potential for iconicity in vocalization is also demonstrated by the many studies of sound symbolism, which find that people have a notable degree of agreement in the meanings they associate with different speech sounds, independently of their linguistic meaning [
<xref rid="RSOS150152C36" ref-type="bibr">36</xref>
<xref rid="RSOS150152C38" ref-type="bibr">38</xref>
]. One commonly attested example is the association of front vowels (with high-pitched second formants) with small size, and back vowels (with low-pitched second formants) with large size [
<xref rid="RSOS150152C39" ref-type="bibr">39</xref>
]. Another example is the so-called ‘kiki-bouba’ effect [
<xref rid="RSOS150152C34" ref-type="bibr">34</xref>
]: voiceless obstruent consonants alternated with front, closed and unrounded vowels evoke a sense of angular shape, whereas voiced sonorants combined with open, rounded vowels evoke a rounded shape [
<xref rid="RSOS150152C40" ref-type="bibr">40</xref>
]. Importantly, studies have found reliable sound symbolism effects related to a wide variety of semantic domains, including dimensions such as abrupt–continuous, liquid–solid, tight–loose, delicate–rugged, sharp–dull, thick–thin, falling–rising and fast–slow [
<xref rid="RSOS150152C41" ref-type="bibr">41</xref>
,
<xref rid="RSOS150152C42" ref-type="bibr">42</xref>
].</p>
<p>Further evidence for the iconic potential of vocalization comes from cross-linguistic research documenting the widespread existence of rich lexical classes of iconic words in many of the world's languages [
<xref rid="RSOS150152C36" ref-type="bibr">36</xref>
,
<xref rid="RSOS150152C43" ref-type="bibr">43</xref>
,
<xref rid="RSOS150152C44" ref-type="bibr">44</xref>
]. These ideophones are grammatically and phonologically distinct from other classes of words, and they typically convey meaning through the depiction of sensory and motor imagery. For example, in Japanese, the word ‘koron’ refers to a light object rolling once, ‘korokoro’ to a light object rolling repeatedly and ‘gorogoro’ to a heavy object rolling repeatedly. In these cases, reduplication is used for the iconic expression of repetition, and voiced compared to voiceless consonants iconically convey a more massive object. The schematic nature of this iconicity is characteristic of the lexical class, as are the systematic ways in which the words relate to each other. Ideophones are documented to span a range of meanings across different modalities, such as manner of motion, shape, psychological and mental states, texture, size, luminance, distance and temporal aspect. Although this synchronic research cannot uncover the origins of these words, the fact of their existence strongly hints at a process of iconic creation.</p>
<p>Thus, given this considerable potential for iconicity in vocalization, it appears possible that people could establish a vocal communication system through the process of iconic creation, comparable to gesture and drawing. To examine this possibility, we conducted a set of experiments in which we tested the ability of participants to create a simple system of vocal symbols in the laboratory.</p>
</sec>
<sec id="s2">
<label>2.</label>
<title>Vocal charades</title>
<p>Pairs of participants played a 10 round game of ‘vocal’ charades. Players took turns producing non-linguistic vocalizations to communicate different meanings as their partner tried to guess the meaning. The set of meanings included nine pairs of antonymic properties spanning a variety of semantic domains: speed, size, length, distance, (vertical) position, quantity, aesthetic and psychological valence. Our goal was to examine whether participants would be able to create iconic vocalizations to successfully communicate each meaning, and then in turn, use these vocalizations to establish a more conventionalized system of vocal symbols. Unavoidably, as in similar semiotics experiments, our participants already spoke a language and thus came well acquainted with symbolic communication. Many previous informative experiments have addressed this issue by tasking participants to communicate in a different modality or medium from speech (e.g. [
<xref rid="RSOS150152C9" ref-type="bibr">9</xref>
,
<xref rid="RSOS150152C16" ref-type="bibr">16</xref>
,
<xref rid="RSOS150152C45" ref-type="bibr">45</xref>
]); here, because of our specific interest in vocalization, we instructed participants to communicate without the words of their spoken language. Our analysis shows they succeeded in observing this constraint.</p>
<sec id="s2a" sec-type="methods">
<label>2.1</label>
<title>Methods</title>
<sec id="s2a1">
<label>2.1.1</label>
<title>Participants</title>
<p>Nine pairs of undergraduate students from the University of California, Santa Cruz participated in the study in exchange for course credit.</p>
</sec>
<sec id="s2a2">
<label>2.1.2</label>
<title>Materials</title>
<p>The game was played with nine pairs of antonyms spanning various semantic domains:
<italic>attractive, ugly, bad, good, big, small, down, up, far, near, fast, slow, few, many, long, short</italic>
,
<italic>rough</italic>
and
<italic>smooth</italic>
. The words were printed on individual index cards that were distributed to the players.</p>
</sec>
<sec id="s2a3">
<label>2.1.3</label>
<title>Design and procedure</title>
<p>Participants took turns producing non-linguistic vocalizations and oral sounds to express the meanings written on index cards as their partner tried to guess each one. No words or body movements were allowed. Players could see each other and were permitted to interact freely, as long as the vocalizer did not use words or gestures (see demonstration video in the electronic supplementary material).</p>
<p>Each player was given a set of 12 shuffled cards, which always included both items of an antonymic pair. Six meanings held by each player were possessed uniquely by that player, and the remaining six were possessed by both players.</p>
<p>The game was played for 10 rounds. In each round, the first player attempted to communicate the meanings contained on all of their cards, one at a time (i.e. a turn). The players then switched roles. Players shuffled their cards before each round. A turn lasted for up to 10 s, timed by the experimenter. During this time, players were permitted to make as many sounds and guesses as they wished. The experimenter monitored the game, and when necessary, reminded participants not to gesture if they began to move their hands, as well as not to use words if they produced a vocalization that sounded too word-like. The turn ended when the allotted time expired, or when a correct guess was made, which was immediately noted by the vocalizer. If no correct guess was made, then the experimenter called time at 10 s, and the vocalizer shared the correct word with their partner.</p>
<p>Vocalizations were recorded by a lapel microphone attached to the vocalizer and guesses by a flat boundary microphone placed in front of the guesser. Both microphones were plugged into stereo channels of a digital recorder, which recorded the experiment at a sampling rate of 44.1 kHz. The microphones were switched each time the players traded roles.</p>
</sec>
</sec>
<sec id="s2b">
<label>2.2</label>
<title>Analysis</title>
<sec id="s2b1">
<label>2.2.1</label>
<title>Statistical analyses</title>
<p>Statistical analyses with mixed-effect models were conducted using the lme4 package in R. Significance tests of continuous outcomes were calculated using
<italic>χ</italic>
<sup>2</sup>
-tests that compared the fit of mixed-effect models with and without the factor of interest on improvement in model fit. Significance tests of binomial outcomes used the
<italic>z</italic>
-values associated with the logistic mixed-effect models. Across all studies, the mixed-effect models included participant and meaning as random effects.</p>
</sec>
<sec id="s2b2">
<label>2.2.2</label>
<title>Acoustic measurements</title>
<p>Acoustic measurements were made with P
<sc>raat</sc>
phonetic analysis computer software [
<xref rid="RSOS150152C46" ref-type="bibr">46</xref>
]. The onset and offset of each sound was marked in a textgrid. The sound files and textgrids were then fed into a script that automated measurements of the acoustic variables: duration, harmonicity (i.e. harmonics to noise ratio), intensity, pitch and pitch change. For all analyses, the acoustic properties of multiple sounds produced during a single turn were averaged together (i.e. vocalizers contributed one, possibly averaged vocalization per meaning per round), and normalized for each individual speaker.</p>
</sec>
<sec id="s2b3">
<label>2.2.3</label>
<title>Measuring stability</title>
<p>We measured the stability of vocalizations across rounds by using the five acoustic measurements to compute the Euclidean distance of each vocalization to the vocalization produced by the same participant for the same meaning in the previous round (only computable for rounds 2–10). The variables were first normalized by participant, and the distance was calculated as the square root of the sum of the squared differences between each variable. To test whether this distance decreased over rounds, we constructed a linear mixed-effect model predicting distance between round
<italic>t</italic>
and round
<italic>t</italic>
−1 with round as a fixed effect and speaker and meaning as random factors.</p>
</sec>
<sec id="s2b4">
<label>2.2.4</label>
<title>Comparison to target words</title>
<p>To ensure that participants were not simply mimicking the prosodic properties of the corresponding target words with their vocalizations, we compared the vocalizations to the acoustic properties of the actual words. We asked three native English speakers to ‘pronounce the words in a neutral tone of voice’. The speakers—students from the University of Wisconsin-Madison—were naive to the hypotheses of the study. Speakers were recorded pronouncing the words as they were presented in randomized order in a Microsoft P
<sc>ower</sc>
P
<sc>oint</sc>
presentation. The duration, intensity, harmonicity, mean pitch and pitch change were measured for each pronunciation, normalized by speaker, and averaged for each word. This characteristic set of acoustic properties for each target word was then compared with the charades vocalizations.</p>
</sec>
</sec>
<sec id="s2c">
<label>2.3</label>
<title>Results</title>
<p>We first examined whether people became more accurate and efficient over the course of the game, i.e. from round 1 to 10. Overall accuracy as proportion of items correct was 82.2%.
<xref ref-type="fig" rid="RSOS150152F1">Figure 1</xref>
<italic>a</italic>
shows that accuracy increased over rounds and eventually exceeded 95% (
<italic>z</italic>
=17.77,
<italic>p</italic>
≪0.001,
<italic>b</italic>
=0.63, 95% CI=[0.57,0.71]). Shown in
<xref ref-type="fig" rid="RSOS150152F1">figure 1</xref>
<italic>b</italic>
,
<italic>c</italic>
, both turn time (
<italic>χ</italic>
<sub>1</sub>
<sup>2</sup>
=702.98,
<italic>p</italic>
≪0.001,
<italic>b</italic>
=−0.57, 95% CI=[−0.60,−0.52]) and the number of guesses per turn (
<italic>χ</italic>
<sub>1</sub>
<sup>2</sup>
=83.79,
<italic>p</italic>
≪0.001,
<italic>b</italic>
=−0.06, 95% CI=[−0.076,−0.050]) decreased over rounds.
<fig id="RSOS150152F1" orientation="portrait" position="float">
<label>Figure 1.</label>
<caption>
<p>Performance and change in the form of vocalizations over 10 round of vocal charades. The lines in each plot were fit by local regression. Error bars represent standard error between subjects. Round is displayed on the
<italic>x</italic>
-axis of each plot. (
<italic>a</italic>
) Mean accuracy increased across rounds. (
<italic>b</italic>
) The mean number of guesses per turn decreased across rounds. (
<italic>c</italic>
) Mean turn time (s) decreased across rounds. (
<italic>d</italic>
) Sound duration (s) decreased across rounds. (
<italic>e</italic>
) The mean Euclidean distance between the form of sounds produced at round
<italic>t</italic>
and round
<italic>t</italic>
−1. Sounds became increasingly similar across the first 5 rounds.</p>
</caption>
<graphic xlink:href="rsos150152-g1"></graphic>
</fig>
</p>
<p>We next examined how the vocalizations that participants produced changed over the rounds of the game, specifically whether they became abbreviated and more stable in form.
<xref ref-type="fig" rid="RSOS150152F1">Figure 1</xref>
<italic>d</italic>
shows that the duration of the sounds decreased across rounds (
<italic>χ</italic>
<sub>1</sub>
<sup>2</sup>
=24.84,
<italic>p</italic>
<0.001,
<italic>b</italic>
=−0.037, 95% CI=[−0.052,0.023]). The vocalizations also became more stable in form across the early rounds of the game, differing less from one round to the next, as shown in
<xref ref-type="fig" rid="RSOS150152F1">figure 1</xref>
<italic>e</italic>
(
<italic>χ</italic>
<sub>1</sub>
<sup>2</sup>
=26.86,
<italic>p</italic>
≪0.001,
<italic>b</italic>
=−0.050, 95% CI=[−0.069,−0.031]).</p>
<p>Were the vocalizations that participants produced iconic? To answer this question, we first looked for similarity in the acoustic properties of the vocalizations produced by different participants with respect to each meaning. We reasoned that if participants produced vocalizations according to a sense of iconicity—potentially highly abstract—between their voice and the meanings they expressed, then their vocalizations would take on a similar form for each meaning.
<xref ref-type="fig" rid="RSOS150152F2">Figure 2</xref>
shows profiles of the median acoustic characteristics (duration, harmonicity, intensity, pitch and pitch change) of the 18 meanings. The arrows in the figure show the results of logistic-regression models predicting each word versus its opposite with the five acoustic variables as predictors. Notably, each meaning varied reliably from its opposite in at least one acoustic property (with
<italic>α</italic>
=0.001). We further tested whether the acoustic profiles of each meaning reliably distinguished between all other meanings. We constructed 153 logistic-regression models for all of the meaning pairs (e.g.
<italic>up</italic>
versus
<italic>rough</italic>
,
<italic>short</italic>
versus
<italic>few</italic>
and so on), of which 132 could be fit with a convergent solution. At
<italic>α</italic>
=0.05, all but two pairs (98.5%) could be reliably distinguished by at least one significant variable (
<xref ref-type="table" rid="RSOS150152TB1">table 1</xref>
). In total, we tested five variables for each of the 132 comparisons for a total of 660 tests, and 376 (57.0%) of these differed reliably between meanings (compared to the null prediction of 0.05×660=33). These results indicate that each meaning tended to occupy a unique place in the acoustic space.
<fig id="RSOS150152F2" orientation="portrait" position="float">
<label>Figure 2.</label>
<caption>
<p>The plots show the acoustic characteristics of each of the 18 meanings. The five variables are represented on the
<italic>x</italic>
-axis: D, duration; H, harmonics to noise ratio; I, intensity; P, pitch; C, pitch change. All values are normalized (
<italic>z</italic>
-scored) for each of the five measures. The red line shows the median and the blue box spans the first and third quartiles. The up and down arrows indicate variables that differed reliably between antonymic meanings. For example, vocalizations for
<italic>bad</italic>
differed from those for
<italic>good</italic>
by having a lower harmonics to noise ratio and pitch. The variables marked with arrows were the basis for the iconic template of each meaning.</p>
</caption>
<graphic xlink:href="rsos150152-g2"></graphic>
</fig>
<table-wrap id="RSOS150152TB1" orientation="portrait" position="float">
<label>Table 1.</label>
<caption>
<p>Number of acoustic variables that reliably predicted one meaning versus another. The columns refer to the number of significant variables per comparison from 0 to 5 and the rows to different
<italic>α</italic>
-values. % significant indicates the percentage of 132 pairs×5 variables=660 tested variables that showed a reliable difference between word pairs. For example, at
<italic>α</italic>
=0.05, six pairs of meanings differed along all five variables, while only two did not exhibit reliable differences with respect to at least one variable. In total, 57.0% of the 660 tested variables were significant by this criterion.</p>
</caption>
<table frame="hsides" rules="groups">
<colgroup span="1">
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
<col align="left" span="1"></col>
</colgroup>
<thead valign="bottom">
<tr>
<th rowspan="1" colspan="1"></th>
<th align="left" colspan="6" rowspan="1">no. significant variables per comparison
<hr></hr>
</th>
<th rowspan="1" colspan="1"></th>
</tr>
<tr>
<th align="left" rowspan="1" colspan="1">
<italic>α</italic>
</th>
<th align="left" rowspan="1" colspan="1">0</th>
<th align="left" rowspan="1" colspan="1">1</th>
<th align="left" rowspan="1" colspan="1">2</th>
<th align="left" rowspan="1" colspan="1">3</th>
<th align="left" rowspan="1" colspan="1">4</th>
<th align="left" rowspan="1" colspan="1">5</th>
<th align="left" rowspan="1" colspan="1">% significant</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="1" colspan="1">0.05</td>
<td rowspan="1" colspan="1">2</td>
<td rowspan="1" colspan="1">13</td>
<td rowspan="1" colspan="1">33</td>
<td rowspan="1" colspan="1">45</td>
<td rowspan="1" colspan="1">33</td>
<td rowspan="1" colspan="1">6</td>
<td rowspan="1" colspan="1">57.0</td>
</tr>
<tr>
<td rowspan="1" colspan="1">0.01</td>
<td rowspan="1" colspan="1">7</td>
<td rowspan="1" colspan="1">31</td>
<td rowspan="1" colspan="1">43</td>
<td rowspan="1" colspan="1">35</td>
<td rowspan="1" colspan="1">15</td>
<td rowspan="1" colspan="1">1</td>
<td rowspan="1" colspan="1">43.5</td>
</tr>
<tr>
<td rowspan="1" colspan="1">0.001</td>
<td rowspan="1" colspan="1">18</td>
<td rowspan="1" colspan="1">46</td>
<td rowspan="1" colspan="1">47</td>
<td rowspan="1" colspan="1">17</td>
<td rowspan="1" colspan="1">4</td>
<td rowspan="1" colspan="1">0</td>
<td rowspan="1" colspan="1">31.4</td>
</tr>
</tbody>
</table>
</table-wrap>
</p>
<p>Finally, to determine that participants were not simply mimicking prosodic properties of the target words, we compared the acoustic profiles of each meaning to the profile of its target word. Across the 18 meanings, we tested the correlation between the average normalized acoustic properties of vocalizations and the properties of the corresponding word. None of the correlations for the five variables showed a reliable relationship between the vocalizations and corresponding words for each meaning (
<italic>r</italic>
's<0.40,
<italic>p</italic>
's>0.10), confirming our impression that participants were indeed generating non-linguistic vocalizations.</p>
</sec>
</sec>
<sec id="s3">
<label>3.</label>
<title>Charades playback experiments</title>
<p>Iconicity is theorized to ground the creation of a symbol system both by facilitating the generation of novel forms, and also by enabling a perceiver to gain a sense of meaning with limited previous learning. The results of the charades games suggest that people are able to generate iconic vocalizations. We next examined the iconicity of the vocalizations by testing the ability of naive listeners to infer their meaning (cf. [
<xref rid="RSOS150152C16" ref-type="bibr">16</xref>
]). Unlike the charade players whose accuracy was assisted by feedback, repetition and possibly by inadvertent cues like facial expressions and gestures, the only source of information for our naive listeners were the sounds themselves.</p>
<sec id="s3a">
<label>3.1</label>
<title>Methods</title>
<sec id="s3a1">
<label>3.1.1</label>
<title>Participants</title>
<p>Four hundred and forty participants were recruited through Amazon Mechanical Turk and they were restricted to be in the USA.</p>
</sec>
<sec id="s3a2">
<label>3.1.2</label>
<title>Stimuli</title>
<p>The stimuli were adapted from the recorded vocalizations produced during the vocal charades game. These included vocalizations for each of the 18 meanings from rounds 1, 5 and 10. A random subset of 6 to 9 vocalizations was used for each meaning from each round, for a total of 359 stimuli. Only the first vocalization produced for a trial was used. In some cases, the partner's guessing overlapped with the vocalization. This segment was trimmed off when possible, but if it occurred too early to be trimmed, the vocalization was not used.</p>
</sec>
<sec id="s3a3">
<label>3.1.3</label>
<title>Design and procedure</title>
<p>The experiment was conducted in two batches. Batch 1 (
<italic>n</italic>
=232) tested eight meanings (
<italic>big</italic>
,
<italic>small</italic>
,
<italic>down</italic>
,
<italic>up</italic>
,
<italic>fast</italic>
,
<italic>slow</italic>
,
<italic>long</italic>
,
<italic>short</italic>
), and batch 2 (
<italic>n</italic>
=208) tested the remaining 10 (
<italic>attractive</italic>
,
<italic>ugly</italic>
,
<italic>bad</italic>
,
<italic>good</italic>
,
<italic>far</italic>
,
<italic>near</italic>
,
<italic>few</italic>
,
<italic>many</italic>
,
<italic>rough</italic>
,
<italic>smooth</italic>
). In both, participants listened to 10 randomly drawn target sounds and selected from 10 options—the eight target meanings plus
<italic>attractive</italic>
and
<italic>ugly</italic>
in batch 1 and the 10 possible meanings in batch 2. Across trials, it was possible that participants heard multiple stimuli intended to express the same meaning, but no specific sound was repeated more than once.</p>
<p>
<xref ref-type="fig" rid="RSOS150152F3">Figure 3</xref>
shows the interface of the experiment. Participants could listen to the sound multiple times and made their choice by using the mouse to drag the selected word into the response box. They were forced to make at least one selection, but could make a total of three ranked selections if they were uncertain. We report only their first selection.
<fig id="RSOS150152F3" orientation="portrait" position="float">
<label>Figure 3.</label>
<caption>
<p>Screenshot of the interface from the playback experiments. Participants listened to a sound as many times as they wanted by clicking on the ‘play’ arrow. They then made their selection by dragging the chosen word into the empty box.</p>
</caption>
<graphic xlink:href="rsos150152-g3"></graphic>
</fig>
</p>
</sec>
</sec>
<sec id="s3b">
<label>3.2</label>
<title>Analysis</title>
<sec id="s3b1">
<label>3.2.1</label>
<title>Word similarity norming</title>
<p>While our main accuracy analyses code accuracy in a discrete way (correct/incorrect), not all errors are the same. For example, confusing ‘attractive’ and ‘good’ may be thought to be a smaller error than confusing ‘attractive’ and ‘far’. To obtain this more continuous error measure, we collected meaning-similarity ratings from a new group of participants. For each of the unique word-pairs, 10 participants indicated the degree to which they were ‘completely different in meaning’ versus ‘almost identical in meaning’ (1–7 Likert scale). Analyses with this continuous error measure mirrored all the results we obtained with the discrete measure. The results depicted in
<xref ref-type="fig" rid="RSOS150152F5">figure 5</xref>
rely on the continuous error measure to sort the rows and columns by rated semantic similarity.</p>
</sec>
<sec id="s3b2">
<label>3.2.2</label>
<title>Median, iconic and target word distance</title>
<p>To assess the role of iconicity in people's ability to accurately map the vocalization to its intended meaning, we computed two distance measures for each vocalization used as a stimulus (computable for 345 of the 359 stimuli; the remaining 14 lacked a measurable pitch). The first measure was a
<italic>median distance</italic>
baseline. We created a template for each meaning that consisted of the median values (across participants) for the five acoustic variables. The median distance of a stimulus was its Euclidean distance from these five median values. We measured
<italic>iconic distance</italic>
by creating a template for each meaning that included only the acoustic variables that differed reliably (
<italic>p</italic>
<0.01) between antonymic words (
<xref ref-type="fig" rid="RSOS150152F2">figure 2</xref>
). In this case, instead of the median values, the template consisted of either the maximum or the minimum value (corresponding to the direction of the comparison of the meaning to its antonym). The iconic distance between a stimulus and a given meaning was then computed as the Euclidean distance between the stimulus and the template with respect to only these particular variables.</p>
<p>We also wanted to determine whether the degree to which a vocalization resembled the acoustic properties of its corresponding target word would influence guessing accuracy. Using the template constructed for each target word for the charades game analysis, we calculated the absolute difference between the vocalization and the word for each of the five acoustic variables. We also computed a composite Euclidean distance between the vocalization and corresponding word template.</p>
</sec>
</sec>
<sec id="s3c">
<label>3.3</label>
<title>Results</title>
<p>Compared to a chance rate of 10%, overall accuracy at guessing the intended meaning was 35.6% in the first batch and 37.3% in the second (with no reliable difference between batches,
<italic>z</italic>
=−0.02, n.s.). Accuracy across meanings ranged from 10.4% (‘near’) to 73.6% (‘attractive’), and 15 of 18 words exceeded 20% (
<xref ref-type="fig" rid="RSOS150152F4">figure 4</xref>
). These accuracies were strongly correlated with accuracies in the charades game for the same meanings (
<italic>r</italic>
=0.57,
<italic>p</italic>
=0.01).
<xref ref-type="fig" rid="RSOS150152F5">Figure 5</xref>
<italic>a</italic>
,
<italic>b</italic>
shows confusion matrices associated with the 18 different meanings. Accuracy was higher for vocalizations made in rounds 5 (38.8%) and 10 (39.1%) compared to round 1 (32.0%;
<italic>z</italic>
=3.28,
<italic>p</italic>
=0.001,
<italic>b</italic>
=0.03, 95% CI=[0.013,0.052];
<xref ref-type="fig" rid="RSOS150152F6">figure 6</xref>
<italic>a</italic>
).
<fig id="RSOS150152F4" orientation="portrait" position="float">
<label>Figure 4.</label>
<caption>
<p>Accuracy of naive listeners at selecting the correct meaning in the playback experiments. The dark blue bars represent playback for stimuli with varied iconicity (10 choices for each meaning; chance=10%, indicated by the dark blue dashed line). The medium and light blue bars show performance from a follow-up experiment with more and less iconic instances of the vocalizations, respectively (18 choices for each meaning: chance=5.6% indicated by the medium blue dashed line). Accuracy was higher for the more iconic versus less iconic vocalizations for all 18 of the tested meanings.</p>
</caption>
<graphic xlink:href="rsos150152-g4"></graphic>
</fig>
<fig id="RSOS150152F5" orientation="portrait" position="float">
<label>Figure 5.</label>
<caption>
<p>Confusions made for each of the meanings during the playback experiments. The
<italic>y</italic>
-axis displays the meaning of the sound that was presented, and the
<italic>x</italic>
-axis displays the meaning that was selected. The ordering of meanings in each matrix was determined by a norming experiment that collected similarity ratings between each possible pair of meanings, with more similar words placed closer together. Warmer colours indicate more frequent choices. (
<italic>a</italic>
) Results from the first batch and (
<italic>b</italic>
) from the second batch in the playback experiment. Warmer colours gravitate along the diagonal, showing that listeners tended to select the correct response or a meaning that was similar to it. The results from the follow-up experiment with more and less iconic stimuli are shown in (
<italic>c</italic>
) and (
<italic>d</italic>
), respectively. Warmer colours gravitate along the diagonal for more iconic stimuli, but responses for less iconic stimuli appear more random.</p>
</caption>
<graphic xlink:href="rsos150152-g5"></graphic>
</fig>
<fig id="RSOS150152F6" orientation="portrait" position="float">
<label>Figure 6.</label>
<caption>
<p>(
<italic>a</italic>
) Accuracy in the playback experiment as a function of the round in which the stimulus was produced in the vocal charades game. Error bars display standard error between subjects. Accuracy is higher in rounds 5 and 10 compared to round 1. (
<italic>b</italic>
) Accuracy as a function of the normalized iconic distance of the stimulus. Iconic distance is the Euclidean distance between the vocalization and the iconic template for its intended meaning. Accuracy is higher for vocalizations that are closer to the template.</p>
</caption>
<graphic xlink:href="rsos150152-g6"></graphic>
</fig>
</p>
<p>The relatively high degree of accuracy by naive listeners indicates that they were able to extract information about meaning from the acoustic signal alone. But is it possible to identify the particular iconic properties that accounted for this ability? To determine this, we compared the
<italic>iconic distance</italic>
between the vocalization and the template to the non-iconic,
<italic>median distance</italic>
. Iconic distance was a reliable predictor of guessing accuracy (
<italic>z</italic>
=−5.71,
<italic>p</italic>
≪0.001,
<italic>b</italic>
=−0.20, 95% CI=[−0.27,−0.13]) when controlling for median distance; median distance was not a reliable predictor when controlling for iconic distance (
<italic>z</italic>
=−1.31, n.s.). As shown in
<xref ref-type="fig" rid="RSOS150152F6">figure 6</xref>
<italic>b</italic>
, listeners were more likely to guess the meaning of vocalizations that had forms closest to the idealized meaning template (i.e. were most iconic by this measure). The addition of
<italic>target word distance</italic>
as a predictor in the model did not alter these results in any significant way. Target word distance was not a reliable predictor of accuracy (
<italic>z</italic>
=1.62, n.s.), and iconic distance remained highly reliable (
<italic>z</italic>
=−3.22,
<italic>p</italic>
=0.001,
<italic>b</italic>
=−0.21, 95% CI=[−0.21,−0.05]). With one exception, the same pattern of results was also found for the absolute distance between vocalization and word with respect to each variable separately: iconic distance remained highly reliable (
<italic>z</italic>
's<−4.02,
<italic>p</italic>
's<0.001), and shorter variable distances did not predict greater accuracy (
<italic>z</italic>
's>0.59). The exception was that the distance between the intensity of the vocalization and the target word was a reliable predictor of accuracy, such that smaller distances lead to greater accuracy (
<italic>z</italic>
=−2.66,
<italic>p</italic>
<0.01,
<italic>b</italic>
=−0.15, 95% CI=[−0.26,−0.04]). However, iconic distance remained a highly reliably predictor (
<italic>z</italic>
=−6.07,
<italic>p</italic>
<0.001,
<italic>b</italic>
=−0.22, 95% CI=[−0.29,−0.15]). It is not clear what to make of this intensity result, particularly as words do not typically have a characteristic intensity. Overall, it appears that participants were making use of iconic properties of the vocalizations to infer their meanings, as the vocalizations did not resemble the corresponding English words.</p>
<p>Across all meanings, there was a moderate negative correlation between the iconic distance to that meaning and its probability of selection, regardless of its correctness (
<italic>r</italic>
=−0.30,
<italic>p</italic>
<0.001), and a weaker negative correlation between selection probability and median distance (
<italic>r</italic>
=−0.18,
<italic>p</italic>
<0.001). Stimuli with acoustic properties that were more similar to the iconic template for a particular meaning were more likely to elicit that meaning as a response. Similarity to the non-iconic median template was less predictive of participants' responses.
<xref ref-type="fig" rid="RSOS150152F7">Figure 7</xref>
shows the correlations between both iconic and median distance with the selection probability for each meaning. The iconic distance correlation was negative for 17 of 18 meanings and weaker than median distance for only two of the meanings.
<fig id="RSOS150152F7" orientation="portrait" position="float">
<label>Figure 7.</label>
<caption>
<p>Correlations between the two distance metrics to each particular meaning and probability of selecting that meaning. The correlation coefficient is plotted on the
<italic>x</italic>
-axis and meaning on the
<italic>y</italic>
-axis. Red represents iconic distance, and blue median distance. Negative correlations indicate that smaller distances to a particular meaning are associated with a higher likelihood of selecting that meaning. All meanings but
<italic>many</italic>
are associated with a negative correlation.</p>
</caption>
<graphic xlink:href="rsos150152-g7"></graphic>
</fig>
</p>
</sec>
</sec>
<sec id="s4">
<label>4.</label>
<title>Playback with more versus less iconic vocalizations</title>
<p>We further wondered whether we could identify exemplars of highly iconic vocalizations for each meaning and how accurately guessers would perform with these exemplary stimuli. To test this, we used iconic distance to identify the two most and the two least iconic stimuli for each meaning. Then, in a follow-up experiment, we played these to a new set of listeners in a between subjects design.</p>
<sec id="s4a">
<label>4.1</label>
<title>Methods</title>
<sec id="s4a1">
<label>4.1.1</label>
<title>Participants</title>
<p>Participants were recruited through Amazon Mechanical Turk and randomly assigned to a
<italic>more iconic</italic>
condition (
<italic>n</italic>
=26) or a
<italic>less iconic</italic>
condition (
<italic>n</italic>
=28).</p>
</sec>
<sec id="s4a2">
<label>4.1.2</label>
<title>Stimuli</title>
<p>The stimuli were drawn from the set used in the full charades playback experiment. For each meaning, we selected the two stimuli with the smallest iconic distance (i.e. highest levels of iconicity) and the two with the largest iconic distance (i.e. lowest levels of iconicity), resulting in 36 stimuli for each condition.</p>
</sec>
<sec id="s4a3">
<label>4.1.3</label>
<title>Design and procedure</title>
<p>The design and procedure were similar to the full charades playback experiment. The critical difference was that participants listened to stimuli expressing all 18 meanings and for each vocalization selected a meaning from the complete set of 18 possible meanings, i.e. an 18-alternative forced choice (chance level=5.6%).</p>
</sec>
</sec>
<sec id="s4b">
<label>4.2</label>
<title>Results</title>
<p>Overall accuracy was higher for more iconic (34.5%) compared with less iconic vocalizations (15.6%;
<italic>z</italic>
=−3.27,
<italic>p</italic>
=0.001,
<italic>b</italic>
=−1.57, 95% CI=[−2.56,−0.64]).
<xref ref-type="fig" rid="RSOS150152F4">Figure 4</xref>
shows the accuracy for the more (medium blue bars) and less iconic (light blue bars) vocalizations for each meaning.
<xref ref-type="fig" rid="RSOS150152F5">Figure 5</xref>
<italic>c</italic>
,
<italic>d</italic>
shows the corresponding confusion matrices.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s5">
<label>5.</label>
<title>Discussion</title>
<p>Our goal was to examine people's ability to generate a simple vocal symbol system through the process of iconic creation, comparable to the process by which many gestural communication systems are known to be generated. We found that participants were successful at communicating a set of 18 different meanings in a vocal charades game. Over 10 rounds, their performance became highly accurate and increasingly efficient. As participants’ vocalizations became more communicatively effective over rounds, they also became shorter and more stable in form. These findings show that people can create novel, non-linguistic vocalizations to express a range of meanings, and that even after relatively few interactions, these vocalizations begin to conventionalize into a system of symbols.</p>
<p>Although some theorists have made strong claims concerning the limited potential of iconicity in the vocal modality (e.g. [
<xref rid="RSOS150152C18" ref-type="bibr">18</xref>
,
<xref rid="RSOS150152C27" ref-type="bibr">27</xref>
,
<xref rid="RSOS150152C28" ref-type="bibr">28</xref>
]), our results demonstrate how people are able to make use of iconicity to ground successful vocal communication—for at least the present set of basic meanings. Tasked with communicating using novel, non-linguistic vocalizations, charades players were inclined to create vocalizations with highly consistent acoustic properties for each particular meaning. This shows that they shared intuitions for how the qualities of their voice mapped to each meaning. In turn, the naive listeners of the playback experiments displayed a considerable degree of accuracy in their ability to interpret the meanings of these vocalizations without any prior learning, showing that they shared intuitions for what vocalizations expressing different meanings
<italic>should</italic>
sound like. When we quantified the iconic properties of the vocalizations used in playback as a distance metric from iconic templates constructed for each meaning, we found that this measure (iconic distance) predicted accuracy and also the specific meanings that listeners selected in response to the vocalization. In a subsequent experiment, when participants were presented with more iconic stimuli, they were much better at inferring the meanings of sounds than when presented with stimuli having lower iconicity.</p>
<p>Strengthening the case for iconicity, we showed that the vocalizations that participants produced were generally not influenced by the corresponding English words. The forms of the vocalizations were not correlated with the forms of the words for any of the measured variables, and listeners' responses were also not influenced by the degree of similarity between the vocalization and its corresponding word in the playback experiments. It is worth emphasizing that our analysis focused on prosodic properties of the vocalizations, and with the exception of duration, and to some extent pitch, these properties are not characteristically associated with lexical forms in English (e.g. words are not distinguished by the intensity at which they are spoken). Thus, the evidence indicates that participants were able to successfully perform the task of creating non-linguistic vocalizations, which has been a point of concern in other semiotic experiments [
<xref rid="RSOS150152C45" ref-type="bibr">45</xref>
].</p>
<p>Even though participants were able to generate non-linguistic vocalizations, the constraint to communicate vocally without using English may nevertheless have affected the vocalizations they produced. For example, participants may have been pushed to make greater use of gross prosodic features rather than more detailed ‘phonetic’ articulatory gestures that they would associate with speaking English. This reduction of the articulatory space could have driven participants to produce more similar vocalizations than they would have with the full space at their disposal. At the same time, the reduction of articulatory space also limits some potential for iconicity, for instance, reducing the potential for sound symbolism associated with English phonological segments.</p>
<p>According to the process of iconic creation, communicators create iconic signals, which then become conventionalized and moulded into a grammatical system over repeated use and transmission.
<sup>
<xref ref-type="fn" rid="FN3">3</xref>
</sup>
Early researchers of signed languages noted the ways in which the iconicity of signs faded over time as the forms became more regularized and systematic, leading to the idea that iconicity appears in the original creation of signs and then decays in a unidirectional process as systematicity sets in [
<xref rid="RSOS150152C4" ref-type="bibr">4</xref>
,
<xref rid="RSOS150152C21" ref-type="bibr">21</xref>
]. Yet, although signed languages exhibit more arbitrariness as they mature, they nevertheless continue to maintain high levels of iconicity [
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C15" ref-type="bibr">15</xref>
]. We do not know what would happen to the iconicity of signed languages over the much longer time periods over which spoken languages have developed, i.e. tens or hundreds of years compared to tens of thousands of years or more. Additionally, studies of other communication systems suggest that the transition from iconic to more arbitrary signals varies between different kinds of media and interactional contexts [
<xref rid="RSOS150152C16" ref-type="bibr">16</xref>
,
<xref rid="RSOS150152C30" ref-type="bibr">30</xref>
].</p>
<p>Here we found that naive listeners were more accurate with vocalizations generated during later rounds of the charades game, suggesting that the vocalizations became more iconic over rounds, even as they showed evidence of increasing conventionalization. This finding may appear somewhat contrary to the theory of iconic creation and the idea that iconicity facilitates the initial generation of signals. One possible explanation relates to the point that communication involves negotiating an effective signal between interlocutors. The vocalizer's intuition of an iconic vocalization may differ somewhat from that of her partner, and a more widely interpretable iconic vocalization may be established between them over the course of multiple interactions. A second explanation relates to the possibility that the particular iconic properties of vocalizations might have become enhanced over rounds, as less relevant properties were de-emphasized. Such a process of maintaining properties of iconic contrast while relaxing other properties could also facilitate contrasts within the wider system of items—promoting better guessing accuracy between the alternatives by a naive listener.</p>
<p>An additional factor that is likely to affect the balance between iconicity and arbitrariness is the crowding of the semantic space in the communication system. Arbitrariness may increase as larger lexicons put pressure on signal systems to be efficient while maintaining discriminability between meanings [
<xref rid="RSOS150152C48" ref-type="bibr">48</xref>
]. If true, tasking people to differentiate just these 18 meanings may not have posed sufficient crowding of the semantic space to disrupt the iconicity that was originally introduced into the system. It is also possible that if we had run the game for more rounds, the iconicity of the vocalizations and their interpretability to naive listeners may have eventually begun to decline.</p>
<p>One notable qualification of our findings is that, while participants produced
<italic>non-linguistic</italic>
vocalizations, it is unavoidable that their vocalizations were influenced by other sound–meaning correspondences shared in their culture. In a few cases, they may have been influenced by iconic vocal emblems: for example, a high-pitched whistle for attractive, or a disgusted ‘eww’ for ugly. Another case, recognized as a common conceptual metaphor of American-English culture, is the use of high pitch for up and low pitch for down [
<xref rid="RSOS150152C49" ref-type="bibr">49</xref>
,
<xref rid="RSOS150152C50" ref-type="bibr">50</xref>
]. On the one hand, evidence indicates that verticality–pitch correspondence arises very early in development [
<xref rid="RSOS150152C35" ref-type="bibr">35</xref>
], and scholars have posed plausible accounts of iconic motivation to explain it. For example, small things that make high-pitched sounds tend to be located upward, whereas big things that make low-pitched sounds tend to be down on the ground, and additionally in production, low-pitched vocalizations resonate in the chest, whereas high-pitched vocalizations resonate higher in the head [
<xref rid="RSOS150152C51" ref-type="bibr">51</xref>
]. However, on the other hand, despite these possible explanations, there is considerable variation across languages in how speakers describe and conceptualize ‘high’ and ‘low’ pitches [
<xref rid="RSOS150152C50" ref-type="bibr">50</xref>
]. Farsi speakers use words translating to ‘thin’ and ‘thick’, for instance. These examples illustrate how conventionalized vocal expressions can reflect iconic motivations without being determined by them. While it is not possible to isolate the role of culture here, future work might conduct similar experiments with people across different cultural and linguistic backgrounds to better understand its role in the generation and interpretation of iconic vocalizations.</p>
<p>Another point of qualification concerns the extent to which these findings might generalize to other meanings beyond the present set. Indeed, we found that for a few of the 18 items examined here—particularly
<italic>few</italic>
,
<italic>near</italic>
and
<italic>far</italic>
—the ability of naive listeners to guess their meaning was quite limited and did not exceed chance performance. (For comparison, however, it is worth noting that when non-signing English speakers viewed undisputedly iconic ASL signs and attempted to guess their meanings in a five-alternative multiple choice format, they were no better than chance for the great majority of signs [
<xref rid="RSOS150152C52" ref-type="bibr">52</xref>
].) Although guessing accuracy for playback was low for these meanings, on the production end, charades players were consistent in using the duration of their vocalizations to distinguish
<italic>few</italic>
from
<italic>many</italic>
. Specifically, they tended to produce contrasting sequences of repeated sounds—a transparently iconic strategy also used to a lesser extent for
<italic>fast</italic>
and
<italic>slow</italic>
. In the case of
<italic>near</italic>
and
<italic>far</italic>
, the difficulties may have resulted from the spatial nature of the concepts, which may not easily map to iconic actions of the vocal tract. Yet again, from the perspective of production, charades players' use of longer duration for
<italic>far</italic>
is intuitive, mapping on to the relative extent of the distance. So it does not appear to be the simple case that these particular meanings are intractable to representation with iconic vocalizations.</p>
<p>Nevertheless, it is important for future studies to consider more broadly the kinds of meanings and conceptual domains that are most readily translated into iconic vocalizations, especially in comparison to gestures. Studies of ideophones across different spoken languages find that they typically express adjectival and adverbial properties [
<xref rid="RSOS150152C53" ref-type="bibr">53</xref>
], and experimental studies have found that people are better at using novel vocalizations to communicate emotion words (e.g.
<italic>tired, pain</italic>
) compared to actions (
<italic>fleeing</italic>
,
<italic>sleeping</italic>
) and objects (
<italic>rock, fruit</italic>
) [
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
,
<xref rid="RSOS150152C30" ref-type="bibr">30</xref>
]. Participants' success with the set of meanings tested here—concentrating on various adjectival and adverbial properties—is in line with these previous findings, as is participants' relative success with meanings bearing emotional qualities (e.g.
<italic>attractive, bad</italic>
). Clearly, some meanings lend themselves more to iconic vocalizations than others, and not just any meaning at all is easily expressed through an iconic vocalization. If participants had not been limited to vocal communication, it is also likely that they would have opted to express some items with gestures or with multimodal combinations of vocalization and gesture [
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
]. In combination with gestures, iconic vocalizations could function effectively to distinguish different meanings without needing to be as distinctive from each other, since contrasts between the forms of confusable meanings could be distributed across the modalities.</p>
<p>This differential affordance for iconic representation has important implications for understanding the creation and development of unbridled communication systems, like spoken languages. Communication with spoken language involves a coordinated mix of vocalizations with manual and other bodily gestures [
<xref rid="RSOS150152C54" ref-type="bibr">54</xref>
,
<xref rid="RSOS150152C55" ref-type="bibr">55</xref>
]. Given the traditional linguistic principle of the arbitrariness of the sign, many scholars have maintained that, in these systems, the vocal channel primarily functions to carry the arbitrary linguistic components of a message, while the visible gestural channel conveys the iconic, non-linguistic components [
<xref rid="RSOS150152C55" ref-type="bibr">55</xref>
]. Stemming from this idea, some have proposed that spoken languages must have originated as iconically grounded systems of manual and other visible bodily gestures, and at some point in human history, arbitrary vocalizations were bootstrapped on inherently iconic manual gestures [
<xref rid="RSOS150152C12" ref-type="bibr">12</xref>
,
<xref rid="RSOS150152C19" ref-type="bibr">19</xref>
,
<xref rid="RSOS150152C27" ref-type="bibr">27</xref>
,
<xref rid="RSOS150152C28" ref-type="bibr">28</xref>
].</p>
<p>Alternatively, we argue for a more balanced hypothesis of how spoken languages are created from iconic origins, taking into account the combined evidence of sound symbolism, ideophones and other iconic vocal phenomena [
<xref rid="RSOS150152C14" ref-type="bibr">14</xref>
,
<xref rid="RSOS150152C18" ref-type="bibr">18</xref>
,
<xref rid="RSOS150152C34" ref-type="bibr">34</xref>
], as well as this work showing that vocal symbols can be formed through iconic creation. We observe that when humans are free to communicate across modalities, they are inclined to embed iconic form–meaning correspondences in vocalizations as well as gestures. Thus, when facing a naturally occurring challenge to devise a communication system, people are likely to take advantage of the strengths of iconic representation in each modality. As a simple example, they may rely predominantly on gestures for representing actions and spatial relationships, and on vocalizations for representing objects and events that are associated with distinctive sounds. The ability to construct reinforcing representations across modalities—for instance, combining a gesture and vocalization to distinguish a particular predator—could also be advantageous, enabling robust communication over different perceptual conditions (e.g. low lighting, background noise, occluded line of sight). An interesting line of research will be to examine how people develop symbol systems when they are permitted to vocalize and gesture freely, including how their use of each modality changes over repetition and transmission.</p>
</sec>
<sec sec-type="conclusions" id="s6">
<label>6.</label>
<title>Conclusion</title>
<p>In a variation of a classic thought experiment, Tomasello [
<xref rid="RSOS150152C28" ref-type="bibr">28</xref>
] asked readers to imagine two groups of children, well cared for, but lacking a language model: one group is constrained to communicate with gesture, the other with vocalization. Citing evidence from signed languages and home sign systems, he suggested that the children would naturally use gestures to create a language because of their iconic potential to ground meaning by pointing at things in the world, pantomiming actions, depicting shapes and spatial layouts, and so on. By contrast, he noted that the iconic potential of vocalization is trivial. It is difficult to imagine the children ‘inventing on their own vocalizations to refer the attention or imagination of others to the world in meaningful ways—beyond perhaps a few vocalizations tied to emotional situations and/or a few instances of vocal mimicry…and so the issue of conventionalizing already meaningful communicative acts never arises' [[
<xref rid="RSOS150152C27" ref-type="bibr">27</xref>
], p. 228]. It is unlikely that we will ever have the opportunity to witness the creation of a spoken language, or that we will succeed in the reconstruction of the original spoken languages, which are speculated to have emerged at the start of the Upper Paleolithic (approx. 50 000 years ago), if not long before [
<xref rid="RSOS150152C10" ref-type="bibr">10</xref>
]. But here, by bringing a version of this thought experiment into the laboratory (albeit with UCSC undergraduates in place of language-less infants), we find that vocal symbol systems can, at least in principle, originate from the same process of iconic creation and conventionalization that has been observed in the formation of signed languages and other communication systems.</p>
</sec>
</body>
<back>
<ack>
<title>Acknowledgements</title>
<p>We are grateful to Ashley Cain for running the experiment.</p>
</ack>
<fn-group>
<fn id="FN1">
<label>1</label>
<p>A compelling case can be made that spoken languages are multimodal systems (e.g. [
<xref rid="RSOS150152C10" ref-type="bibr">10</xref>
,
<xref rid="RSOS150152C11" ref-type="bibr">11</xref>
]), and we consider this important point in Discussion of how vocalizations and gestures might function together in the creation of multimodal systems.</p>
</fn>
<fn id="FN2">
<label>2</label>
<p>We use ‘iconic’ as a modifier of ‘gesture’ independently of whether the gesture co-occurs with speech, or whether the gesture is conventionalized (e.g. an ASL sign). That is, ‘iconic’ refers to the motivated quality of a signal, and ‘gesture’ to its visible modality.</p>
</fn>
<fn id="FN3">
<label>3</label>
<p>Arbitrarily generated signals also become systematic over repeated use and transmission [
<xref rid="RSOS150152C45" ref-type="bibr">45</xref>
,
<xref rid="RSOS150152C47" ref-type="bibr">47</xref>
]; this part of the process is not particular to iconic signals.</p>
</fn>
</fn-group>
<sec id="s7">
<title>Ethics</title>
<p>The vocal charades experiment was approved by the Institutional Review Board of the University of California, Santa Cruz, and all participants provided informed, written consent. The playback experiments were approved by the Institutional Review Board of the University of Wisconsin-Madison. The individuals included in the supplementary video gave informed consent for the publication of their identifying information.</p>
</sec>
<sec id="s8">
<title>Data accessibility</title>
<p>We cannot share the original vocalizations because they contain potentially identifying information. We have made our analysis and code, including the measurements of the vocalizations, available through the Dryad Digital Repository at the following link:
<uri xlink:href="https://datadryad.org/resource/doi:10.5061/dryad.vq33n">https://datadryad.org/resource/doi:10.5061/dryad.vq33n</uri>
.</p>
</sec>
<sec id="s9">
<title>Authors' contributions</title>
<p>M.P. and G.L. devised the experiments. M.P., G.L. and R.D. analysed the data and wrote the manuscript.</p>
</sec>
<sec id="s10">
<title>Competing interests</title>
<p>We declare have no competing interests.</p>
</sec>
<sec id="s11">
<title>Funding</title>
<p>This work was funded in part by an NSF INSPIRE Award 1344279 to G.L. and R.D.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="RSOS150152C1">
<label>1</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>De Vos</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Pfau</surname>
<given-names>R</given-names>
</name>
</person-group>
<year>2015</year>
<article-title>Sign language typology: the contribution of rural sign languages</article-title>
.
<source>Annu. Rev. Linguist.</source>
<volume>1</volume>
,
<fpage>265</fpage>
<lpage>288</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1146/annurev-linguist-030514-124958">doi:10.1146/annurev-linguist-030514-124958</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C2">
<label>2</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sandler</surname>
<given-names>W</given-names>
</name>
,
<name>
<surname>Meir</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Padden</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Aronoff</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2005</year>
<article-title>The emergence of grammar: systematic structure in a new language</article-title>
.
<source>Proc. Natl Acad. Sci. USA</source>
<volume>102</volume>
,
<fpage>2661</fpage>
<lpage>2665</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1073/pnas.0405448102">doi:10.1073/pnas.0405448102</ext-link>
)
<pub-id pub-id-type="pmid">15699343</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C3">
<label>3</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Senghas</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Kita</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Özyürek</surname>
<given-names>A</given-names>
</name>
</person-group>
<year>2004</year>
<article-title>Children creating core properties of language: evidence from an emerging sign language in Nicaragua</article-title>
.
<source>Science</source>
<volume>305</volume>
,
<fpage>1779</fpage>
<lpage>1782</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1126/science.1100199">doi:10.1126/science.1100199</ext-link>
)
<pub-id pub-id-type="pmid">15375269</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C4">
<label>4</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Frishberg</surname>
<given-names>N</given-names>
</name>
</person-group>
<year>1975</year>
<article-title>Arbitrariness and iconicity: historical change in American Sign Language</article-title>
.
<source>Language</source>
<volume>51</volume>
,
<fpage>696</fpage>
<lpage>719</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.2307/412894">doi:10.2307/412894</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C5">
<label>5</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meir</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Israel</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Sandler</surname>
<given-names>W</given-names>
</name>
,
<name>
<surname>Padden</surname>
<given-names>CA</given-names>
</name>
,
<name>
<surname>Aronoff</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2012</year>
<article-title>The influence of community on language structure: evidence from two young sign languages</article-title>
.
<source>Linguist. Var.</source>
<volume>12</volume>
,
<fpage>247</fpage>
<lpage>291</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1075/lv.12.2.04mei">doi:10.1075/lv.12.2.04mei</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C6">
<label>6</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goldin-Meadow</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Feldman</surname>
<given-names>H</given-names>
</name>
</person-group>
<year>1977</year>
<article-title>The development of language-like communication without a language model</article-title>
.
<source>Science</source>
<volume>197</volume>
,
<fpage>401</fpage>
<lpage>403</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1126/science.877567">doi:10.1126/science.877567</ext-link>
)
<pub-id pub-id-type="pmid">877567</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C7">
<label>7</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Umiker-Sebeok</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Sebeok</surname>
<given-names>TA</given-names>
</name>
</person-group>
<year>1987</year>
<source>Monastic sign languages</source>
.
<publisher-loc>Berlin, Germany</publisher-loc>
:
<publisher-name>Walter de Gruyter</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C8">
<label>8</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kendon</surname>
<given-names>A</given-names>
</name>
</person-group>
<year>1988</year>
<source>Sign languages of aboriginal Australia: cultural, semiotic and communicative perspectives</source>
.
<publisher-loc>Cambridge, UK</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C9">
<label>9</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goldin-Meadow</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>McNeill</surname>
<given-names>D</given-names>
</name>
,
<name>
<surname>Singleton</surname>
<given-names>J</given-names>
</name>
</person-group>
<year>1996</year>
<article-title>Silence is liberating: removing the handcuffs on grammatical expression in the manual modality</article-title>
.
<source>Psychol. Rev.</source>
<volume>103</volume>
,
<fpage>34</fpage>
<lpage>55</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1037/0033-295X.103.1.34">doi:10.1037/0033-295X.103.1.34</ext-link>
)
<pub-id pub-id-type="pmid">8650298</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C10">
<label>10</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dediu</surname>
<given-names>D</given-names>
</name>
,
<name>
<surname>Levinson</surname>
<given-names>SC</given-names>
</name>
</person-group>
<year>2013</year>
<article-title>On the antiquity of language: the reinterpretation of Neandertal linguistic capacities and its consequences</article-title>
.
<source>Lang. Sci.</source>
<volume>4</volume>
,
<fpage>397</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.3389/fpsyg.2013.00397">doi:10.3389/fpsyg.2013.00397</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C11">
<label>11</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kendon</surname>
<given-names>A</given-names>
</name>
</person-group>
<year>2014</year>
<article-title>Semiotic diversity in utterance production and the concept of ‘language’</article-title>
.
<source>Phil. Trans. R. Soc. B</source>
<volume>369</volume>
,
<fpage>20130293</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1098/rstb.2013.0293">doi:10.1098/rstb.2013.0293</ext-link>
)
<pub-id pub-id-type="pmid">25092661</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C12">
<label>12</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Armstrong</surname>
<given-names>DF</given-names>
</name>
,
<name>
<surname>Wilcox</surname>
<given-names>SE</given-names>
</name>
</person-group>
<year>2007</year>
<source>The gestural origin of language</source>
,
<edition>1 edn</edition>
<publisher-loc>Oxford, UK</publisher-loc>
:
<publisher-name>Oxford University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C13">
<label>13</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meir</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Padden</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Aronoff</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Sandler</surname>
<given-names>W</given-names>
</name>
</person-group>
<year>2013</year>
<article-title>Competing iconicities in the structure of languages</article-title>
.
<source>Cogn. Linguist.</source>
<volume>24</volume>
,
<fpage>309</fpage>
<lpage>343</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1515/cog-2013-0010">doi:10.1515/cog-2013-0010</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C14">
<label>14</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perniss</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Vigliocco</surname>
<given-names>G</given-names>
</name>
</person-group>
<year>2014</year>
<article-title>The bridge of iconicity: from a world of experience to the experience of language</article-title>
.
<source>Phil. Trans. R. Soc. B</source>
<volume>369</volume>
,
<fpage>20130300</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1098/rstb.2013.0300">doi:10.1098/rstb.2013.0300</ext-link>
)
<pub-id pub-id-type="pmid">25092668</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C15">
<label>15</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Taub</surname>
<given-names>SF</given-names>
</name>
</person-group>
<year>2001</year>
<source>Language from the body: iconicity and metaphor in American sign language</source>
.
<publisher-loc>Cambridge, UK</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C16">
<label>16</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Garrod</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Fay</surname>
<given-names>N</given-names>
</name>
,
<name>
<surname>Lee</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Oberlander</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>MacLeod</surname>
<given-names>T</given-names>
</name>
</person-group>
<year>2007</year>
<article-title>Foundations of representation: where might graphical symbol systems come from</article-title>
?
<source>Cogn. Sci.</source>
<volume>31</volume>
,
<fpage>961</fpage>
<lpage>987</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1080/03640210701703659">doi:10.1080/03640210701703659</ext-link>
)
<pub-id pub-id-type="pmid">21635324</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C17">
<label>17</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perlman</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Clark</surname>
<given-names>N</given-names>
</name>
,
<name>
<surname>Johansson Falck</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2015</year>
<article-title>Iconic prosody in story reading</article-title>
.
<source>Cogn. Sci.</source>
<volume>39</volume>
,
<fpage>1348</fpage>
<lpage>1368</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1111/cogs.12190">doi:10.1111/cogs.12190</ext-link>
)
<pub-id pub-id-type="pmid">25351919</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C18">
<label>18</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Imai</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Kita</surname>
<given-names>S</given-names>
</name>
</person-group>
<year>2014</year>
<article-title>The sound symbolism bootstrapping hypothesis for language acquisition and language evolution</article-title>
.
<source>Phil. Trans. R. Soc. B</source>
<volume>369</volume>
,
<fpage>20130298</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1098/rstb.2013.0298">doi:10.1098/rstb.2013.0298</ext-link>
)
<pub-id pub-id-type="pmid">25092666</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C19">
<label>19</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fay</surname>
<given-names>N</given-names>
</name>
,
<name>
<surname>Arbib</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Garrod</surname>
<given-names>S</given-names>
</name>
</person-group>
<year>2013</year>
<article-title>How to bootstrap a human communication system</article-title>
.
<source>Cogn. Sci.</source>
<volume>37</volume>
,
<fpage>1356</fpage>
<lpage>1367</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1111/cogs.12048">doi:10.1111/cogs.12048</ext-link>
)
<pub-id pub-id-type="pmid">23763661</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C20">
<label>20</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Goldin-Meadow</surname>
<given-names>S</given-names>
</name>
</person-group>
<year>2003</year>
<source>The resilience of language: what gesture creation in deaf children can tell us about how all children learn language</source>
.
<publisher-loc>New York, NY</publisher-loc>
:
<publisher-name>Psychology Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C21">
<label>21</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aronoff</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Meir</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Sandler</surname>
<given-names>W</given-names>
</name>
</person-group>
<year>2005</year>
<article-title>The paradox of sign language morphology</article-title>
.
<source>Language</source>
<volume>81</volume>
,
<fpage>301</fpage>
<lpage>344</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1353/lan.2005.0043">doi:10.1353/lan.2005.0043</ext-link>
)
<pub-id pub-id-type="pmid">22223926</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C22">
<label>22</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sandler</surname>
<given-names>W</given-names>
</name>
,
<name>
<surname>Aronoff</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Meir</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Padden</surname>
<given-names>C</given-names>
</name>
</person-group>
<year>2011</year>
<article-title>The gradual emergence of phonological form in a new language</article-title>
.
<source>Nat. Lang. Linguist. Theory</source>
<volume>29</volume>
,
<fpage>503</fpage>
<lpage>543</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1007/s11049-011-9128-2">doi:10.1007/s11049-011-9128-2</ext-link>
)
<pub-id pub-id-type="pmid">22223927</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C23">
<label>23</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Gelb</surname>
<given-names>IJ</given-names>
</name>
</person-group>
<year>1952</year>
<source>A study of writing: the foundations of grammatology</source>
.
<publisher-loc>Chicago, IL</publisher-loc>
:
<publisher-name>University of Chicago Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C24">
<label>24</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Sampson</surname>
<given-names>G</given-names>
</name>
</person-group>
<year>1985</year>
<source>Writing systems: a linguistic introduction</source>
.
<publisher-loc>Palo Alto, CA</publisher-loc>
:
<publisher-name>Stanford University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C25">
<label>25</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Vaccari</surname>
<given-names>O</given-names>
</name>
,
<name>
<surname>Vaccari</surname>
<given-names>EE</given-names>
</name>
</person-group>
<year>1964</year>
<source>Pictorial Chinese–Japanese characters: a new and fascinating method to learn ideographs</source>
.
<publisher-loc>Tokyo, Japan</publisher-loc>
:
<publisher-name>Vaccari's Language Institute</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C26">
<label>26</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arbib</surname>
<given-names>MA</given-names>
</name>
</person-group>
<year>2005</year>
<article-title>From monkey-like action recognition to human language: an evolutionary framework for neurolinguistics</article-title>
.
<source>Behav. Brain Sci.</source>
<volume>28</volume>
,
<fpage>105</fpage>
<lpage>124</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1017/S0140525X05000038">doi:10.1017/S0140525X05000038</ext-link>
)
<pub-id pub-id-type="pmid">16201457</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C27">
<label>27</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Corballis</surname>
<given-names>MC</given-names>
</name>
</person-group>
<year>2003</year>
<source>From hand to mouth: the origins of language</source>
.
<publisher-loc>Princeton, NJ</publisher-loc>
:
<publisher-name>Princeton University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C28">
<label>28</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Tomasello</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2008</year>
<source>Origins of human communication</source>
.
<publisher-loc>Cambridge, MA</publisher-loc>
:
<publisher-name>The MIT Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C29">
<label>29</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hockett</surname>
<given-names>CF</given-names>
</name>
</person-group>
<year>1978</year>
<article-title>In search of Jove's brow</article-title>
.
<source>Am. Speech</source>
<volume>53</volume>
,
<fpage>243</fpage>
<lpage>313</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.2307/455140">doi:10.2307/455140</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C30">
<label>30</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fay</surname>
<given-names>N</given-names>
</name>
,
<name>
<surname>Lister</surname>
<given-names>CJ</given-names>
</name>
,
<name>
<surname>Ellison</surname>
<given-names>TM</given-names>
</name>
,
<name>
<surname>Goldin-Meadow</surname>
<given-names>S</given-names>
</name>
</person-group>
<year>2014</year>
<article-title>Creating a communication system from scratch: gesture beats vocalization hands down</article-title>
.
<source>Front. Psychol.</source>
<volume>5</volume>
,
<fpage>354</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.3389/fpsyg.2014.00354">doi:10.3389/fpsyg.2014.00354</ext-link>
)
<pub-id pub-id-type="pmid">24808874</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C31">
<label>31</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Quine</surname>
<given-names>WV</given-names>
</name>
</person-group>
<year>1960</year>
<source>Word and object</source>
.
<publisher-loc>Cambridge, MA</publisher-loc>
:
<publisher-name>Technology Press of the Massachusetts Institute of Technology</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C32">
<label>32</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Spence</surname>
<given-names>C</given-names>
</name>
</person-group>
<year>2011</year>
<article-title>Crossmodal correspondences: a tutorial review</article-title>
.
<source>Atten. Percept. Psychophys.</source>
<volume>73</volume>
,
<fpage>971</fpage>
<lpage>995</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.3758/s13414-010-0073-7">doi:10.3758/s13414-010-0073-7</ext-link>
)
<pub-id pub-id-type="pmid">21264748</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C33">
<label>33</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Winter</surname>
<given-names>B</given-names>
</name>
,
<name>
<surname>Marghetis</surname>
<given-names>T</given-names>
</name>
,
<name>
<surname>Matlock</surname>
<given-names>T</given-names>
</name>
</person-group>
<year>2015</year>
<article-title>Of magnitudes and metaphors: explaining cognitive interactions between space, time, and number</article-title>
.
<source>Cortex</source>
<volume>64</volume>
,
<fpage>209</fpage>
<lpage>224</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1016/j.cortex.2014.10.015">doi:10.1016/j.cortex.2014.10.015</ext-link>
)
<pub-id pub-id-type="pmid">25437376</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C34">
<label>34</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ramachandran</surname>
<given-names>VS</given-names>
</name>
,
<name>
<surname>Hubbard</surname>
<given-names>EM</given-names>
</name>
</person-group>
<year>2001</year>
<article-title>Synaesthesia—a window into perception, thought and language</article-title>
.
<source>J. Conscious. Stud.</source>
<volume>8</volume>
,
<fpage>3</fpage>
<lpage>34</lpage>
.</mixed-citation>
</ref>
<ref id="RSOS150152C35">
<label>35</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Walker</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Bremner</surname>
<given-names>JG</given-names>
</name>
,
<name>
<surname>Mason</surname>
<given-names>U</given-names>
</name>
,
<name>
<surname>Spring</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Mattock</surname>
<given-names>K</given-names>
</name>
,
<name>
<surname>Slater</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Johnson</surname>
<given-names>SP</given-names>
</name>
</person-group>
<year>2010</year>
<article-title>Preverbal infants' sensitivity to synaesthetic cross-modality correspondences</article-title>
.
<source>Psychol. Sci.</source>
<volume>21</volume>
,
<fpage>21</fpage>
<lpage>25</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1177/0956797609354734">doi:10.1177/0956797609354734</ext-link>
)
<pub-id pub-id-type="pmid">20424017</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C36">
<label>36</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Hinton</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Nichols</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Ohala</surname>
<given-names>JJ</given-names>
</name>
</person-group>
<year>1994</year>
<source>Sound symbolism</source>
.
<publisher-loc>Cambridge, UK</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C37">
<label>37</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nuckolls</surname>
<given-names>JB</given-names>
</name>
</person-group>
<year>1999</year>
<article-title>The case for sound symbolism</article-title>
.
<source>Annu. Rev. Anthropol.</source>
<volume>28</volume>
,
<fpage>225</fpage>
<lpage>252</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1146/annurev.anthro.28.1.225">doi:10.1146/annurev.anthro.28.1.225</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C38">
<label>38</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perniss</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Thompson</surname>
<given-names>RL</given-names>
</name>
,
<name>
<surname>Vigliocco</surname>
<given-names>G</given-names>
</name>
</person-group>
<year>2010</year>
<article-title>Iconicity as a general property of language: evidence from spoken and signed languages</article-title>
.
<source>Front. Psychol.</source>
<volume>1</volume>
,
<fpage>227</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.3389/fpsyg.2010.00227">doi:10.3389/fpsyg.2010.00227</ext-link>
)
<pub-id pub-id-type="pmid">21833282</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C39">
<label>39</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ohala</surname>
<given-names>JJ</given-names>
</name>
</person-group>
<year>1994</year>
<article-title>The frequency code underlies the sound-symbolic use of voice pitch</article-title>
. In
<source>Sound symbolism</source>
(eds
<person-group person-group-type="editor">
<name>
<surname>Hinton</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Nichols</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Ohala</surname>
<given-names>JJ</given-names>
</name>
</person-group>
), pp.
<fpage>325</fpage>
<lpage>347</lpage>
.
<publisher-loc>Cambridge, UK</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C40">
<label>40</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ahlner</surname>
<given-names>F</given-names>
</name>
,
<name>
<surname>Zlatev</surname>
<given-names>J</given-names>
</name>
</person-group>
<year>2010</year>
<article-title>Cross-modal iconicity: a cognitive semiotic approach to sound symbolism</article-title>
.
<source>Sign Syst. Stud.</source>
<volume>38</volume>
,
<fpage>298</fpage>
<lpage>348</lpage>
.</mixed-citation>
</ref>
<ref id="RSOS150152C41">
<label>41</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Greenberg</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Jenkins</surname>
<given-names>J</given-names>
</name>
</person-group>
<year>1964</year>
<article-title>Studies in the psychological correlates of the sound system of American English</article-title>
.
<source>Word</source>
<volume>20</volume>
,
<fpage>157</fpage>
<lpage>177</lpage>
.</mixed-citation>
</ref>
<ref id="RSOS150152C42">
<label>42</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Klink</surname>
<given-names>RR</given-names>
</name>
</person-group>
<year>2000</year>
<article-title>Creating brand names with meaning: the use of sound symbolism</article-title>
.
<source>Mark. Lett.</source>
<volume>11</volume>
,
<fpage>5</fpage>
<lpage>20</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1023/A:1008184423824">doi:10.1023/A:1008184423824</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C43">
<label>43</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dingemanse</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2012</year>
<article-title>Advances in the cross-linguistic study of ideophones</article-title>
.
<source>Lang. Linguist. Compass</source>
<volume>6</volume>
,
<fpage>654</fpage>
<lpage>672</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1002/lnc3.361">doi:10.1002/lnc3.361</ext-link>
)</mixed-citation>
</ref>
<ref id="RSOS150152C44">
<label>44</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Voeltz</surname>
<given-names>EFK</given-names>
</name>
,
<name>
<surname>Kilian-Hatz</surname>
<given-names>C</given-names>
</name>
</person-group>
<year>2001</year>
<source>Ideophones</source>
.
<publisher-loc>Amsterdam, The Netherlands</publisher-loc>
:
<publisher-name>John Benjamins Publishing</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C45">
<label>45</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Verhoef</surname>
<given-names>T</given-names>
</name>
</person-group>
<year>2012</year>
<article-title>The origins of duality of patterning in artificial whistled languages</article-title>
.
<source>Lang. Cogn.</source>
<volume>4</volume>
,
<fpage>357</fpage>
<lpage>380</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1515/langcog-2012-0019">doi:10.1515/langcog-2012-0019</ext-link>
)
<pub-id pub-id-type="pmid">23637710</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C46">
<label>46</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boersma</surname>
<given-names>P</given-names>
</name>
</person-group>
<year>2001</year>
<article-title>Praat, a system for doing phonetics by computer</article-title>
.
<source>Glot Int.</source>
<volume>5</volume>
,
<fpage>341</fpage>
<lpage>345</lpage>
.</mixed-citation>
</ref>
<ref id="RSOS150152C47">
<label>47</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kirby</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Cornish</surname>
<given-names>H</given-names>
</name>
,
<name>
<surname>Smith</surname>
<given-names>K</given-names>
</name>
</person-group>
<year>2008</year>
<article-title>Cumulative cultural evolution in the laboratory: an experimental approach to the origins of structure in human language</article-title>
.
<source>Proc. Natl Acad. Sci. USA</source>
<volume>105</volume>
,
<fpage>10681</fpage>
<lpage>10686</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1073/pnas.0707835105">doi:10.1073/pnas.0707835105</ext-link>
)
<pub-id pub-id-type="pmid">18667697</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C48">
<label>48</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Monaghan</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Shillcock</surname>
<given-names>RC</given-names>
</name>
,
<name>
<surname>Christiansen</surname>
<given-names>MH</given-names>
</name>
,
<name>
<surname>Kirby</surname>
<given-names>S</given-names>
</name>
</person-group>
<year>2014</year>
<article-title>How arbitrary is language</article-title>
?
<source>Phil. Trans. R. Soc. B</source>
<volume>369</volume>
,
<fpage>20130299</fpage>
(
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1098/rstb.2013.0299">doi:10.1098/rstb.2013.0299</ext-link>
)
<pub-id pub-id-type="pmid">25092667</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C49">
<label>49</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Clark</surname>
<given-names>N</given-names>
</name>
,
<name>
<surname>Perlman</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Johansson Falck</surname>
<given-names>M</given-names>
</name>
</person-group>
<year>2013</year>
<article-title>Iconic pitch expresses vertical space</article-title>
. In
<source>Language and the creative mind</source>
(eds
<person-group person-group-type="editor">
<name>
<surname>Dancygier</surname>
<given-names>B</given-names>
</name>
,
<name>
<surname>Borkent</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Hinnell</surname>
<given-names>J</given-names>
</name>
</person-group>
), pp.
<fpage>393</fpage>
<lpage>410</lpage>
.
<publisher-loc>Stanford, CA</publisher-loc>
:
<publisher-name>SCLI Publications</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C50">
<label>50</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dolscheid</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Shayan</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Majid</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Casasanto</surname>
<given-names>D</given-names>
</name>
</person-group>
<year>2013</year>
<article-title>The thickness of musical pitch: psychophysical evidence for linguistic relativity</article-title>
.
<source>Psychol. Sci.</source>
<volume>24</volume>
,
<fpage>613</fpage>
<lpage>621</lpage>
. (
<ext-link ext-link-type="uri" xlink:href="http://dx.doi.org/10.1177/0956797612457374">doi:10.1177/0956797612457374</ext-link>
)
<pub-id pub-id-type="pmid">23538914</pub-id>
</mixed-citation>
</ref>
<ref id="RSOS150152C51">
<label>51</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zbikowski</surname>
<given-names>LM</given-names>
</name>
</person-group>
<year>1998</year>
<article-title>Metaphor and music theory: reflections from cognitive science</article-title>
.
<source>Music Theory Online</source>
<volume>4</volume>
<comment>(
<uri xlink:href="http://www.mtosmt.org/issues/mto.98.4.1/mto.98.4.1.zbikowski.html">http://www.mtosmt.org/issues/mto.98.4.1/mto.98.4.1.zbikowski.html</uri>
)</comment>
</mixed-citation>
</ref>
<ref id="RSOS150152C52">
<label>52</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Klima</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Bellugi</surname>
<given-names>U</given-names>
</name>
</person-group>
<year>1988</year>
<source>The signs of language</source>
.
<publisher-loc>Cambridge, MA</publisher-loc>
:
<publisher-name>Harvard University Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C53">
<label>53</label>
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Nuckolls</surname>
<given-names>JB</given-names>
</name>
</person-group>
<year>2003</year>
<article-title>To be or not to be ideophonically impoverished</article-title>
. In
<conf-name>Proc. 11th Annual Symp. about Language and Society</conf-name>
,
<conf-loc>Austin, TX</conf-loc>
,
<conf-date>11–13 April 2003</conf-date>
(eds
<person-group person-group-type="editor">
<name>
<surname>Chiang</surname>
<given-names>WF</given-names>
</name>
,
<name>
<surname>Chun</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Mahalingappa</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Mehus</surname>
<given-names>S</given-names>
</name>
</person-group>
), pp.
<fpage>131</fpage>
<lpage>142</lpage>
.</mixed-citation>
</ref>
<ref id="RSOS150152C54">
<label>54</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>McNeill</surname>
<given-names>D</given-names>
</name>
</person-group>
<year>1992</year>
<source>Hand and mind: what gestures reveal about thought</source>
.
<publisher-loc>Chicago, IL</publisher-loc>
:
<publisher-name>University of Chicago Press</publisher-name>
.</mixed-citation>
</ref>
<ref id="RSOS150152C55">
<label>55</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kendon</surname>
<given-names>A</given-names>
</name>
</person-group>
<year>2004</year>
<source>Gesture: visible action as utterance</source>
.
<publisher-loc>Cambridge, UK</publisher-loc>
:
<publisher-name>Cambridge University Press</publisher-name>
.</mixed-citation>
</ref>
</ref-list>
</back>
</pmc>
<affiliations>
<list>
<country>
<li>États-Unis</li>
</country>
<region>
<li>Californie</li>
<li>Wisconsin</li>
</region>
</list>
<tree>
<country name="États-Unis">
<region name="Wisconsin">
<name sortKey="Perlman, Marcus" sort="Perlman, Marcus" uniqKey="Perlman M" first="Marcus" last="Perlman">Marcus Perlman</name>
</region>
<name sortKey="Dale, Rick" sort="Dale, Rick" uniqKey="Dale R" first="Rick" last="Dale">Rick Dale</name>
<name sortKey="Lupyan, Gary" sort="Lupyan, Gary" uniqKey="Lupyan G" first="Gary" last="Lupyan">Gary Lupyan</name>
</country>
</tree>
</affiliations>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Wicri/Musique/explor/OperaV1/Data/Ncbi/Merge
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000E00 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Ncbi/Merge/biblio.hfd -nk 000E00 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Wicri/Musique
   |area=    OperaV1
   |flux=    Ncbi
   |étape=   Merge
   |type=    RBID
   |clé=     PMC:4555852
   |texte=   Iconicity can ground the creation of vocal symbols
}}

Pour générer des pages wiki

HfdIndexSelect -h $EXPLOR_AREA/Data/Ncbi/Merge/RBID.i   -Sk "pubmed:26361547" \
       | HfdSelect -Kh $EXPLOR_AREA/Data/Ncbi/Merge/biblio.hfd   \
       | NlmPubMed2Wicri -a OperaV1 

Wicri

This area was generated with Dilib version V0.6.21.
Data generation: Thu Apr 14 14:59:05 2016. Site generation: Thu Jan 4 23:09:23 2024