Serveur d'exploration sur les relations entre la France et l'Australie

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.
***** Acces problem to record *****\

Identifieur interne : 0003688 ( Pmc/Corpus ); précédent : 0003687; suivant : 0003689 ***** probable Xml problem with record *****

Links to Exploration step


Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Motherese by Eye and Ear: Infants Perceive Visual Prosody in Point-Line Displays of Talking Heads</title>
<author>
<name sortKey="Kitamura, Christine" sort="Kitamura, Christine" uniqKey="Kitamura C" first="Christine" last="Kitamura">Christine Kitamura</name>
<affiliation>
<nlm:aff id="aff1">
<addr-line>School of Social Science and Psychology, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
<affiliation>
<nlm:aff id="aff2">
<addr-line>MARCS Institute, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
</author>
<author>
<name sortKey="Guellai, Bahia" sort="Guellai, Bahia" uniqKey="Guellai B" first="Bahia" last="Guellaï">Bahia Guellaï</name>
<affiliation>
<nlm:aff id="aff3">
<addr-line>Laboratoire Ethologie, Cognition, Développement, Université Paris Ouest Nanterre La Défense, Paris, France</addr-line>
</nlm:aff>
</affiliation>
</author>
<author>
<name sortKey="Kim, Jeesun" sort="Kim, Jeesun" uniqKey="Kim J" first="Jeesun" last="Kim">Jeesun Kim</name>
<affiliation>
<nlm:aff id="aff2">
<addr-line>MARCS Institute, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">PMC</idno>
<idno type="pmid">25353978</idno>
<idno type="pmc">4213016</idno>
<idno type="url">http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4213016</idno>
<idno type="RBID">PMC:4213016</idno>
<idno type="doi">10.1371/journal.pone.0111467</idno>
<date when="2014">2014</date>
<idno type="wicri:Area/Pmc/Corpus">000368</idno>
<idno type="wicri:explorRef" wicri:stream="Pmc" wicri:step="Corpus" wicri:corpus="PMC">000368</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en" level="a" type="main">Motherese by Eye and Ear: Infants Perceive Visual Prosody in Point-Line Displays of Talking Heads</title>
<author>
<name sortKey="Kitamura, Christine" sort="Kitamura, Christine" uniqKey="Kitamura C" first="Christine" last="Kitamura">Christine Kitamura</name>
<affiliation>
<nlm:aff id="aff1">
<addr-line>School of Social Science and Psychology, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
<affiliation>
<nlm:aff id="aff2">
<addr-line>MARCS Institute, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
</author>
<author>
<name sortKey="Guellai, Bahia" sort="Guellai, Bahia" uniqKey="Guellai B" first="Bahia" last="Guellaï">Bahia Guellaï</name>
<affiliation>
<nlm:aff id="aff3">
<addr-line>Laboratoire Ethologie, Cognition, Développement, Université Paris Ouest Nanterre La Défense, Paris, France</addr-line>
</nlm:aff>
</affiliation>
</author>
<author>
<name sortKey="Kim, Jeesun" sort="Kim, Jeesun" uniqKey="Kim J" first="Jeesun" last="Kim">Jeesun Kim</name>
<affiliation>
<nlm:aff id="aff2">
<addr-line>MARCS Institute, University of Western Sydney, Sydney, Australia</addr-line>
</nlm:aff>
</affiliation>
</author>
</analytic>
<series>
<title level="j">PLoS ONE</title>
<idno type="eISSN">1932-6203</idno>
<imprint>
<date when="2014">2014</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass></textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">
<p>Infant-directed (ID) speech provides exaggerated auditory and visual prosodic cues. Here we investigated if infants were sensitive to the match between the auditory and visual correlates of ID speech prosody. We presented 8-month-old infants with two silent line-joined point-light displays of faces speaking different ID sentences, and a single vocal-only sentence matched to one of the displays. Infants looked longer to the matched than mismatched visual signal when full-spectrum speech was presented; and when the vocal signals contained speech low-pass filtered at 400 Hz. When the visual display was separated into rigid (head only) and non-rigid (face only) motion, the infants looked longer to the visual match in the rigid condition; and to the visual mismatch in the non-rigid condition. Overall, the results suggest 8-month-olds can extract information about the prosodic structure of speech from voice and head kinematics, and are sensitive to their match; and that they are less sensitive to the match between lip and voice information in connected speech.</p>
</div>
</front>
<back>
<div1 type="bibliography">
<listBibl>
<biblStruct>
<analytic>
<author>
<name sortKey="Sumby, Wh" uniqKey="Sumby W">WH Sumby</name>
</author>
<author>
<name sortKey="Pollack, I" uniqKey="Pollack I">I Pollack</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Davis, C" uniqKey="Davis C">C Davis</name>
</author>
<author>
<name sortKey="Kim, J" uniqKey="Kim J">J Kim</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Cvejic, E" uniqKey="Cvejic E">E Cvejic</name>
</author>
<author>
<name sortKey="Kim, J" uniqKey="Kim J">J Kim</name>
</author>
<author>
<name sortKey="Davis, C" uniqKey="Davis C">C Davis</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Davis, C" uniqKey="Davis C">C Davis</name>
</author>
<author>
<name sortKey="Kim, J" uniqKey="Kim J">J Kim</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Swerts, M" uniqKey="Swerts M">M Swerts</name>
</author>
<author>
<name sortKey="Krahmer, E" uniqKey="Krahmer E">E Krahmer</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Aldridge, Ma" uniqKey="Aldridge M">MA Aldridge</name>
</author>
<author>
<name sortKey="Braga, Es" uniqKey="Braga E">ES Braga</name>
</author>
<author>
<name sortKey="Walton, Ge" uniqKey="Walton G">GE Walton</name>
</author>
<author>
<name sortKey="Bower, Tgr" uniqKey="Bower T">TGR Bower</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Patterson, Ml" uniqKey="Patterson M">ML Patterson</name>
</author>
<author>
<name sortKey="Werker, Jf" uniqKey="Werker J">JF Werker</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kuhl, P" uniqKey="Kuhl P">P Kuhl</name>
</author>
<author>
<name sortKey="Meltzoff, A" uniqKey="Meltzoff A">A Meltzoff</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kuhl, P" uniqKey="Kuhl P">P Kuhl</name>
</author>
<author>
<name sortKey="Meltzoff, A" uniqKey="Meltzoff A">A Meltzoff</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Patterson, Ml" uniqKey="Patterson M">ML Patterson</name>
</author>
<author>
<name sortKey="Werker, Jf" uniqKey="Werker J">JF Werker</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Patterson, Ml" uniqKey="Patterson M">ML Patterson</name>
</author>
<author>
<name sortKey="Werker, Jf" uniqKey="Werker J">JF Werker</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kuhl, P" uniqKey="Kuhl P">P Kuhl</name>
</author>
<author>
<name sortKey="Williams, Ka" uniqKey="Williams K">KA Williams</name>
</author>
<author>
<name sortKey="Meltzoff, An" uniqKey="Meltzoff A">AN Meltzoff</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Soken, Nh" uniqKey="Soken N">NH Soken</name>
</author>
<author>
<name sortKey="Pick, Ad" uniqKey="Pick A">AD Pick</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Walker Andrews, As" uniqKey="Walker Andrews A">AS Walker-Andrews</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Dodd, B" uniqKey="Dodd B">B Dodd</name>
</author>
</analytic>
</biblStruct>
<biblStruct></biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kitamura, C" uniqKey="Kitamura C">C Kitamura</name>
</author>
<author>
<name sortKey="Burnham, D" uniqKey="Burnham D">D Burnham</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Kitamura, C" uniqKey="Kitamura C">C Kitamura</name>
</author>
<author>
<name sortKey="Thanavisuth, C" uniqKey="Thanavisuth C">C Thanavisuth</name>
</author>
<author>
<name sortKey="Luksaneeyanawin, S" uniqKey="Luksaneeyanawin S">S Luksaneeyanawin</name>
</author>
<author>
<name sortKey="Burnham, D" uniqKey="Burnham D">D Burnham</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fernald, A" uniqKey="Fernald A">A Fernald</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Panneton Cooper, R" uniqKey="Panneton Cooper R">R Panneton Cooper</name>
</author>
<author>
<name sortKey="Aslin, Rn" uniqKey="Aslin R">RN Aslin</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Theissen, E" uniqKey="Theissen E">E Theissen</name>
</author>
<author>
<name sortKey="Hill, A" uniqKey="Hill A">A Hill</name>
</author>
<author>
<name sortKey="Saffran, Jr" uniqKey="Saffran J">JR Saffran</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Fernald, A" uniqKey="Fernald A">A Fernald</name>
</author>
<author>
<name sortKey="Mazzie, C" uniqKey="Mazzie C">C Mazzie</name>
</author>
</analytic>
</biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Green, J" uniqKey="Green J">J Green</name>
</author>
<author>
<name sortKey="Nip, I" uniqKey="Nip I">I Nip</name>
</author>
<author>
<name sortKey="Wilson, E" uniqKey="Wilson E">E Wilson</name>
</author>
<author>
<name sortKey="Mefferd, A" uniqKey="Mefferd A">A Mefferd</name>
</author>
<author>
<name sortKey="Yunusova, Y" uniqKey="Yunusova Y">Y Yunusova</name>
</author>
</analytic>
</biblStruct>
<biblStruct></biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Chong, Scf" uniqKey="Chong S">SCF Chong</name>
</author>
<author>
<name sortKey="Werker, Jf" uniqKey="Werker J">JF Werker</name>
</author>
<author>
<name sortKey="Russell, Ja" uniqKey="Russell J">JA Russell</name>
</author>
<author>
<name sortKey="Carroll, Jm" uniqKey="Carroll J">JM Carroll</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Arterberry, Me" uniqKey="Arterberry M">ME Arterberry</name>
</author>
<author>
<name sortKey="Bornstein, Mh" uniqKey="Bornstein M">MH Bornstein</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Bardi, L" uniqKey="Bardi L">L Bardi</name>
</author>
<author>
<name sortKey="Regolin, L" uniqKey="Regolin L">L Regolin</name>
</author>
<author>
<name sortKey="Simion, F" uniqKey="Simion F">F Simion</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Bidet Ildei, C" uniqKey="Bidet Ildei C">C Bidet-Ildei</name>
</author>
<author>
<name sortKey="Kitromilides, E" uniqKey="Kitromilides E">E Kitromilides</name>
</author>
<author>
<name sortKey="Orliaguet, J P" uniqKey="Orliaguet J">J-P Orliaguet</name>
</author>
<author>
<name sortKey="Pavlova, M" uniqKey="Pavlova M">M Pavlova</name>
</author>
<author>
<name sortKey="Gentaz, E" uniqKey="Gentaz E">E Gentaz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Simion, F" uniqKey="Simion F">F Simion</name>
</author>
<author>
<name sortKey="Regolin, L" uniqKey="Regolin L">L Regolin</name>
</author>
<author>
<name sortKey="Bulf, H" uniqKey="Bulf H">H Bulf</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Bertoncini, J" uniqKey="Bertoncini J">J Bertoncini</name>
</author>
<author>
<name sortKey="Mehler, J" uniqKey="Mehler J">J Mehler</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Bijeljac Babic, R" uniqKey="Bijeljac Babic R">R Bijeljac-Babic</name>
</author>
<author>
<name sortKey="Bertoncini, J" uniqKey="Bertoncini J">J Bertoncini</name>
</author>
<author>
<name sortKey="Mehler, J" uniqKey="Mehler J">J Mehler</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Moon, C" uniqKey="Moon C">C Moon</name>
</author>
<author>
<name sortKey="Bever, T" uniqKey="Bever T">T Bever</name>
</author>
<author>
<name sortKey="Fifer, W" uniqKey="Fifer W">W Fifer</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Cutler, A" uniqKey="Cutler A">A Cutler</name>
</author>
<author>
<name sortKey="Mehler, J" uniqKey="Mehler J">J Mehler</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Munhall, Kg" uniqKey="Munhall K">KG Munhall</name>
</author>
<author>
<name sortKey="Jones, Ja" uniqKey="Jones J">JA Jones</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Grant, Kw" uniqKey="Grant K">KW Grant</name>
</author>
<author>
<name sortKey="Van Wassenhove, V" uniqKey="Van Wassenhove V">V van Wassenhove</name>
</author>
<author>
<name sortKey="Poeppel, D" uniqKey="Poeppel D">D Poeppel</name>
</author>
</analytic>
</biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Lewkowicz, D" uniqKey="Lewkowicz D">D Lewkowicz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Lewkowicz, D" uniqKey="Lewkowicz D">D Lewkowicz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Munhall, Kg" uniqKey="Munhall K">KG Munhall</name>
</author>
<author>
<name sortKey="Jones, Ja" uniqKey="Jones J">JA Jones</name>
</author>
<author>
<name sortKey="Callan, De" uniqKey="Callan D">DE Callan</name>
</author>
<author>
<name sortKey="Kuratate, T" uniqKey="Kuratate T">T Kuratate</name>
</author>
<author>
<name sortKey="Vatikiotis Bateson, E" uniqKey="Vatikiotis Bateson E">E Vatikiotis-Bateson</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Houston Price, C" uniqKey="Houston Price C">C Houston-Price</name>
</author>
<author>
<name sortKey="Nakai, S" uniqKey="Nakai S">S Nakai</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Golinkoff, R" uniqKey="Golinkoff R">R Golinkoff</name>
</author>
<author>
<name sortKey="Hirsh Pasek, K" uniqKey="Hirsh Pasek K">K Hirsh-Pasek</name>
</author>
<author>
<name sortKey="Cauley, K" uniqKey="Cauley K">K Cauley</name>
</author>
<author>
<name sortKey="Gordon, L" uniqKey="Gordon L">L Gordon</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Lewkowicz, D" uniqKey="Lewkowicz D">D Lewkowicz</name>
</author>
<author>
<name sortKey="Hansen Tift, A" uniqKey="Hansen Tift A">A Hansen-Tift</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Lewkowicz, D" uniqKey="Lewkowicz D">D Lewkowicz</name>
</author>
</analytic>
</biblStruct>
<biblStruct>
<analytic>
<author>
<name sortKey="Desjardins, Rn" uniqKey="Desjardins R">RN Desjardins</name>
</author>
<author>
<name sortKey="Werker, Jf" uniqKey="Werker J">JF Werker</name>
</author>
</analytic>
</biblStruct>
</listBibl>
</div1>
</back>
</TEI>
<pmc article-type="research-article">
<pmc-dir>properties open_access</pmc-dir>
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">PLoS One</journal-id>
<journal-id journal-id-type="iso-abbrev">PLoS ONE</journal-id>
<journal-id journal-id-type="publisher-id">plos</journal-id>
<journal-id journal-id-type="pmc">plosone</journal-id>
<journal-title-group>
<journal-title>PLoS ONE</journal-title>
</journal-title-group>
<issn pub-type="epub">1932-6203</issn>
<publisher>
<publisher-name>Public Library of Science</publisher-name>
<publisher-loc>San Francisco, USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="pmid">25353978</article-id>
<article-id pub-id-type="pmc">4213016</article-id>
<article-id pub-id-type="publisher-id">PONE-D-14-23223</article-id>
<article-id pub-id-type="doi">10.1371/journal.pone.0111467</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Research Article</subject>
</subj-group>
<subj-group subj-group-type="Discipline-v2">
<subject>Biology and Life Sciences</subject>
<subj-group>
<subject>Neuroscience</subject>
<subj-group>
<subject>Cognitive Science</subject>
<subj-group>
<subject>Cognitive Psychology</subject>
<subj-group>
<subject>Language</subject>
<subject>Perception</subject>
</subj-group>
</subj-group>
</subj-group>
<subj-group>
<subject>Sensory Systems</subject>
</subj-group>
</subj-group>
<subj-group>
<subject>Psychology</subject>
<subj-group>
<subject>Behavior</subject>
<subj-group>
<subject>Verbal Behavior</subject>
</subj-group>
</subj-group>
<subj-group>
<subject>Developmental Psychology</subject>
</subj-group>
</subj-group>
</subj-group>
<subj-group subj-group-type="Discipline-v2">
<subject>Social Sciences</subject>
<subj-group>
<subject>Linguistics</subject>
<subj-group>
<subject>Language Acquisition</subject>
<subject>Speech</subject>
</subj-group>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Motherese by Eye and Ear: Infants Perceive Visual Prosody in Point-Line Displays of Talking Heads</article-title>
<alt-title alt-title-type="running-head">Motherese by Eye and Ear</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Kitamura</surname>
<given-names>Christine</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="cor1">
<sup>*</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Guellaï</surname>
<given-names>Bahia</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Kim</surname>
<given-names>Jeesun</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<addr-line>School of Social Science and Psychology, University of Western Sydney, Sydney, Australia</addr-line>
</aff>
<aff id="aff2">
<label>2</label>
<addr-line>MARCS Institute, University of Western Sydney, Sydney, Australia</addr-line>
</aff>
<aff id="aff3">
<label>3</label>
<addr-line>Laboratoire Ethologie, Cognition, Développement, Université Paris Ouest Nanterre La Défense, Paris, France</addr-line>
</aff>
<contrib-group>
<contrib contrib-type="editor">
<name>
<surname>Pavlova</surname>
<given-names>Marina A.</given-names>
</name>
<role>Editor</role>
<xref ref-type="aff" rid="edit1"></xref>
</contrib>
</contrib-group>
<aff id="edit1">
<addr-line>University of Tuebingen Medical School, Germany</addr-line>
</aff>
<author-notes>
<corresp id="cor1">* E-mail:
<email>c.kitamura@uws.edu.au</email>
</corresp>
<fn fn-type="conflict">
<p>
<bold>Competing Interests: </bold>
The authors have declared that no competing interests exist.</p>
</fn>
<fn fn-type="con">
<p>Conceived and designed the experiments: CK JK BG. Performed the experiments: BG CK. Analyzed the data: CK BG. Contributed reagents/materials/analysis tools: CK JK. Contributed to the writing of the manuscript: CK JK BG.</p>
</fn>
</author-notes>
<pub-date pub-type="collection">
<year>2014</year>
</pub-date>
<pub-date pub-type="epub">
<day>29</day>
<month>10</month>
<year>2014</year>
</pub-date>
<volume>9</volume>
<issue>10</issue>
<elocation-id>e111467</elocation-id>
<history>
<date date-type="received">
<day>28</day>
<month>5</month>
<year>2014</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>9</month>
<year>2014</year>
</date>
</history>
<permissions>
<copyright-year>2014</copyright-year>
<copyright-holder>Kitamura et al</copyright-holder>
<license>
<license-p>This is an open-access article distributed under the terms of the
<ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License</ext-link>
, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<abstract>
<p>Infant-directed (ID) speech provides exaggerated auditory and visual prosodic cues. Here we investigated if infants were sensitive to the match between the auditory and visual correlates of ID speech prosody. We presented 8-month-old infants with two silent line-joined point-light displays of faces speaking different ID sentences, and a single vocal-only sentence matched to one of the displays. Infants looked longer to the matched than mismatched visual signal when full-spectrum speech was presented; and when the vocal signals contained speech low-pass filtered at 400 Hz. When the visual display was separated into rigid (head only) and non-rigid (face only) motion, the infants looked longer to the visual match in the rigid condition; and to the visual mismatch in the non-rigid condition. Overall, the results suggest 8-month-olds can extract information about the prosodic structure of speech from voice and head kinematics, and are sensitive to their match; and that they are less sensitive to the match between lip and voice information in connected speech.</p>
</abstract>
<funding-group>
<funding-statement>This work was supported by the Australian Research Council (ARC): ARC DP120104298 [CK JK] and ARC DP0666857 [JK].
<ext-link ext-link-type="uri" xlink:href="http://www.arc.gov.au/ncgp/default.htm">http://www.arc.gov.au/ncgp/default.htm</ext-link>
. The funders had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript.</funding-statement>
</funding-group>
<counts>
<page-count count="8"></page-count>
</counts>
<custom-meta-group>
<custom-meta id="data-availability">
<meta-name>Data Availability</meta-name>
<meta-value>The authors confirm that all data underlying the findings are fully available without restriction. All data are contained within the paper.</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
<notes>
<title>Data Availability</title>
<p>The authors confirm that all data underlying the findings are fully available without restriction. All data are contained within the paper.</p>
</notes>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Young infants need to make sense of an array of multisensory stimulation that pervades their everyday experience. This includes the mother's talking face usually positioned facing the infant ensuring that her speech is both seen and heard. It is well known that visual information influences the perception of speech. For adults, watching a talker's moving face assists them to understand speech in noisy
<xref rid="pone.0111467-Sumby1" ref-type="bibr">[1]</xref>
and quiet environments
<xref rid="pone.0111467-Davis1" ref-type="bibr">[2]</xref>
. A growing body of research shows adults also use visual prosodic cues to stress, rhythm and intonation to enhance their understanding of expressive speech and they can do this without access to articulatory cues of the lips, jaw etc.
<xref rid="pone.0111467-Cvejic1" ref-type="bibr">[3]</xref>
,
<xref rid="pone.0111467-Davis2" ref-type="bibr">[4]</xref>
,
<xref rid="pone.0111467-Swerts1" ref-type="bibr">[5]</xref>
. These studies highlight the role peri-oral cues play in the perception of visual speech.</p>
<p>Much auditory-visual speech research concentrates on infants’ ability to relate phonetic information in the lips and the voice. Infants are typically tested facing side-by-side displays of two faces articulating two vowels, e.g.,/i/versus/a/, while hearing one vowel. They are deemed to detect congruency if they look longer to the matching side. Remarkably, there is evidence that from birth, infants detect equivalent phonetic information in the lips and voice
<xref rid="pone.0111467-Aldridge1" ref-type="bibr">[6]</xref>
but auditory-visual phonetic matching is also shown at 2 months
<xref rid="pone.0111467-Patterson1" ref-type="bibr">[7]</xref>
, at 4.5 months
<xref rid="pone.0111467-Kuhl1" ref-type="bibr">[8]</xref>
,
<xref rid="pone.0111467-Kuhl2" ref-type="bibr">[9]</xref>
,
<xref rid="pone.0111467-Patterson2" ref-type="bibr">[10]</xref>
; and at 8 months based on the gender of the talker
<xref rid="pone.0111467-Patterson3" ref-type="bibr">[11]</xref>
. However, it seems infants require the full spectral complexity in the voice to detect congruency, because when the vowels are reduced to, for example, sine-wave analogues or simple tones, they are no longer sensitive to audiovisual vowel correspondences
<xref rid="pone.0111467-Kuhl3" ref-type="bibr">[12]</xref>
. Taken together, these studies suggest young infants already have the primitives of lip reading for single speech sounds.</p>
<p>At the syllable level, it is clear infants are sensitive to correspondences between lip and voice information. We are interested, however, in longer stretches of speech and whether infants can detect the prosodic commonalities in the multi-word utterances from their sight and sound. Evidence using continuous speech typically shows that young infants rely on the congruence between auditory emotions (happy, angry) and the appropriate facial gestures
<xref rid="pone.0111467-Soken1" ref-type="bibr">[13]</xref>
,
<xref rid="pone.0111467-WalkerAndrews1" ref-type="bibr">[14]</xref>
. Infants are also sensitive to temporal synchrony in continuous speech as they look longer to nursery rhymes that are presented with face and voice in synchrony than out of synchrony
<xref rid="pone.0111467-Dodd1" ref-type="bibr">[15]</xref>
. More recently, Blossom and Morgan examined sensitivity to the temporal synchrony of visual prosody using continuous infant-directed (ID) speech
<xref rid="pone.0111467-Blossom1" ref-type="bibr">[16]</xref>
. Similar to others, they found infants aged 10–11 months intermodally matched synchronous faces and voices, but in a second visual prosody only experiment (where the mouth was obscured) infants could not detect synchrony even when the delay was increased to 20 frames. A third experiment tested whether infants could match visual prosody (mouths obscured) when each of the two silent faces was speaking different passages. They found that with in-phase voices and faces, infants looked longer to the congruent face. This study indicates firstly, that visual prosody may be subject to different constraints than visual phonetic information, and secondly, that at 10–11 months, infants use visual prosody, especially as it is exaggerated in ID speech, to extract information about the structure of language.</p>
<p>Inherent in mother infant interactions is the exaggerated prosodic characteristics found in the mother's speech
<xref rid="pone.0111467-Kitamura1" ref-type="bibr">[17]</xref>
,
<xref rid="pone.0111467-Kitamura2" ref-type="bibr">[18]</xref>
. The perceptual salience of ID over adult-directed (AD) speech has been well established in the auditory domain
<xref rid="pone.0111467-Fernald1" ref-type="bibr">[19]</xref>
,
<xref rid="pone.0111467-PannetonCooper1" ref-type="bibr">[20]</xref>
with its exaggerated prosody assisting infants in many of the challenges of early speech perception, e.g., to segment the speech stream
<xref rid="pone.0111467-Theissen1" ref-type="bibr">[21]</xref>
, to find stressed words in sentences
<xref rid="pone.0111467-Fernald2" ref-type="bibr">[22]</xref>
to discriminate multi-syllabic words
<xref rid="pone.0111467-Karzon1" ref-type="bibr">[23]</xref>
, and parse clauses
<xref rid="pone.0111467-KemlerNelson1" ref-type="bibr">[24]</xref>
. ID speech is less studied in the visual domain, but does contain similar exaggerations to facial expressions and articulatory lip gestures for corner vowels
<xref rid="pone.0111467-Green1" ref-type="bibr">[25]</xref>
,
<xref rid="pone.0111467-Kim1" ref-type="bibr">[26]</xref>
,
<xref rid="pone.0111467-Chong1" ref-type="bibr">[27]</xref>
.</p>
<p>The aim of the present study is to examine whether infants are able to intermodally match multi-syllabic ID utterances from their auditory and visual signals. Evidence that the perceptual salience of ID speech prosody translates to the visual domain would suggest that infants’ speech perception is assisted by not only auditory but also visual speech prosody. The current study follows in the same vein to Blossom and Morgan, but we tested younger infants, 8-month-olds, on their ability to match the intermodal properties of auditory and visual prosody. As the visual counterpart, we used specially constructed line-joined point-light displays derived from the head and face motion of three females recorded reading short sentences to their babies. We used point-
<italic>line</italic>
rather than point-
<italic>light</italic>
displays because we were concerned that point-light displays might not provide infants with enough coherent information given the relatively sparse number of motion detectors used in our displays (see
<xref ref-type="fig" rid="pone-0111467-g001">Figure 1</xref>
). It is also important to note that placement of the mouth markers on the outer edges of the lips de-emphasizes the contrastive details of consonant articulation and tends to emphasize syllabic modulations. In other words, the lip displays do not provide fine articulatory detail. Our main objective was to test whether infants would be able to perceive the prosodic relations of auditory and visual speech by using stimuli that have been tailored to provide relatively more suprasegmental than segmental speech cues.</p>
<fig id="pone-0111467-g001" orientation="portrait" position="float">
<object-id pub-id-type="doi">10.1371/journal.pone.0111467.g001</object-id>
<label>Figure 1</label>
<caption>
<title>A static image of the side-by-side the point-line talking faces presented to the infants.</title>
</caption>
<graphic xlink:href="pone.0111467.g001"></graphic>
</fig>
<p>Importantly, these kinds of point-light displays permit face and head kinematics to be presented without other potentially distracting information such as form, texture, and other features. Point-light displays have been used in the past to show that infants can, for example, categorize displays of animals and vehicles
<xref rid="pone.0111467-Arterberry1" ref-type="bibr">[28]</xref>
; extract emotional information
<xref rid="pone.0111467-Soken1" ref-type="bibr">[13]</xref>
and recognise biological motion
<xref rid="pone.0111467-Bardi1" ref-type="bibr">[29]</xref>
,
<xref rid="pone.0111467-BidetIldei1" ref-type="bibr">[30]</xref>
,
<xref rid="pone.0111467-Simion1" ref-type="bibr">[31]</xref>
but not to examine visual speech motion. Another benefit of using point-line displays derived from Optotrak data is that the motion produced during speech can be separated into rigid and non-rigid motion. Rigid motion refers to movements in which the head translates up-and-down, side-to-side, front-to-back and rotates (pitch, yaw, roll) whereas non-rigid motion refers to face-internal movements that change the spatial configuration of the facial surface over time, e.g., articulatory gestures of the lips and jaw, and expressions provided by eye and cheek region.</p>
<p>The auditory-visual stimuli were developed taking into account properties to which infants are sensitive, that is, the global prosodic motion signaled by rigid (head) movement, and the modulation of locally produced syllables in non-rigid motion (face internal). For the mouth movement, we were trying to match the “grain-size” of information across the two modalities. For example, in auditory perception infants pay particular attention to the syllable
<xref rid="pone.0111467-Bertoncini1" ref-type="bibr">[32]</xref>
,
<xref rid="pone.0111467-BijeljacBabic1" ref-type="bibr">[33]</xref>
,
<xref rid="pone.0111467-Moon1" ref-type="bibr">[34]</xref>
. The bias to syllables is consistent with infants’ predilection to attend to the periodic portions of the speech signal
<xref rid="pone.0111467-Cutler1" ref-type="bibr">[35]</xref>
which carries prosodic information. Here, we use the term visual prosody to refer to all aspects of speech production that are used to visually convey the acoustic correlates of prosody (e.g. stress, rhythm, intonation). This assumes that peri-oral cues are mainly used in the perception of visual prosody but even with lip movement, there is a certain rhythm to the cyclic gesturing of the mouth opening and closing to signal the onset and offset of syllables
<xref rid="pone.0111467-Munhall1" ref-type="bibr">[36]</xref>
. Here we call this syllabic rhythm with the caveat that this is not necessarily reflective of the rhythmic structure of the language as the stimuli consist of mainly single syllable ID sentences. Indeed, we used pairs of sentences that differed in the number of syllables but were matched for duration to allow infants to focus on syllable differences.</p>
<p>At some point in development infants should become sensitive, as adults are, to the sequencing of visual syllable-sized events, that is, events that occur in relatively quick succession and are conveyed by the rhythmic sequencing of syllabic information in the mouth region. In this study the sentences presented to each infant varied in the syllable difference between each pair of sentences (see
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
). For infants, we categorize these events as occurring at a
<italic>local rhythmic</italic>
level because exaggerated mouth openings should act as a syllabic timing device before acting as a cue to phonetic grouping in an utterance
<xref rid="pone.0111467-Grant1" ref-type="bibr">[37]</xref>
. Using sentence pairs with unequal syllables should make it easier for infants to match face and voice when the syllable difference between sentences is larger (see
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
). This is especially pertinent if infants are paying attention to the information carried in non-rigid face internal motion. Over and above the local syllabic cues, infants should derive meaning from
<italic>global intonation</italic>
, that is, exaggerated intonational contours that extend over larger temporal windows. These larger intonational groupings are conveyed movements of the head (e.g. nodding and tilting) and unfold more slowly over time. Given a talking face/head produces auditory-visual signals that consist of multiple events that occur at different timescales, we expect infants to be more sensitive to global intonational groupings than to local rhythmic ones.</p>
<table-wrap id="pone-0111467-t001" orientation="portrait" position="float">
<object-id pub-id-type="doi">10.1371/journal.pone.0111467.t001</object-id>
<label>Table 1</label>
<caption>
<title>Sentence pairs from female talker 1, 2 and 3 showing the sentence pairs (A and B) and the number of syllables in each sentence in parenthesis.</title>
</caption>
<alternatives>
<graphic id="pone-0111467-t001-1" xlink:href="pone.0111467.t001"></graphic>
<table frame="hsides" rules="groups">
<colgroup span="1">
<col align="left" span="1"></col>
<col align="center" span="1"></col>
<col align="center" span="1"></col>
<col align="center" span="1"></col>
<col align="center" span="1"></col>
</colgroup>
<thead>
<tr>
<td align="left" rowspan="1" colspan="1">Female Talker</td>
<td align="left" rowspan="1" colspan="1">Sentence A (syllable no)</td>
<td align="left" rowspan="1" colspan="1">Sentence B (syllable no)</td>
<td align="left" rowspan="1" colspan="1">Syllable Diff</td>
<td align="left" rowspan="1" colspan="1">Duration (secs)</td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" rowspan="1" colspan="1">1</td>
<td align="left" rowspan="1" colspan="1">Sheep (8)</td>
<td align="left" rowspan="1" colspan="1">Apple (10)</td>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">4</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">2</td>
<td align="left" rowspan="1" colspan="1">Clams (9)</td>
<td align="left" rowspan="1" colspan="1">Mouse (12)</td>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">6</td>
</tr>
<tr>
<td align="left" rowspan="1" colspan="1">3</td>
<td align="left" rowspan="1" colspan="1">Sheep (8)</td>
<td align="left" rowspan="1" colspan="1">Mouse (12)</td>
<td align="left" rowspan="1" colspan="1">4</td>
<td align="left" rowspan="1" colspan="1">5</td>
</tr>
</tbody>
</table>
</alternatives>
<table-wrap-foot>
<fn id="nt101">
<label></label>
<p>The difference in syllable number between A and B, and duration of sentence pairs is shown in the last two columns. The sentences were: “Did you know, Woolly is a sheep” (Sheep); “Yum, clams are round, small, soft and tasty” (Clams); “Look, the big red apple fell to the ground” (Apple) and “It's tea time, let's feed the white mouse some flower seeds” (Mouse).</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>In this study we conducted two experiments. The aim of Experiment 1 was to establish whether 8-month-olds could match a talking face and voice, when (i) the auditory speech signal contained the full spectrum of speech frequencies; and (ii) the auditory signal was low-pass filtered at 400 Hz to provide mainly intonation in the auditory signal. Here, our question was whether there would be a difference in the degree to which infants could match full-spectrum versus low-pass filtered intonation-only speech to a talking face/head. The aim of Experiment 2 was to determine whether infants could match auditory and visual displays when one group was presented with (i) rigid-only head motion only; and another group with (ii) non-rigid face-internal motion only. As already stated, we expected infants would find it easier to match a talking face and voice using rigid head than non-rigid face motion in an intermodal preference procedure.</p>
</sec>
<sec id="s2">
<title>Experiment 1: Full Spectrum versus Intonation-Only Auditory Speech</title>
<p>In Experiment 1 we tested two groups of 8-month-old infants to examine the role visual prosody plays in speech perception. We chose this younger age, as it has been established that 10 to 11-month-olds can match the intermodal properties of auditory and visual prosody in continuous speech. Both groups saw side-by-side silent visual displays and heard the auditory counterpart of one of the displays. The auditory counterpart for the first infant group was the full speech spectrum of multi-word utterances; and the auditory counterpart for the second group of infants was an intonation version of the utterances (low-pass filtered at 400 Hz), that is, all segmental information above 400 Hz is removed from the auditory speech signal.</p>
<sec id="s2a" sec-type="method|materials">
<title>Method and Materials</title>
<p>All research was approved by the Human Research Ethics Committee (HREC) at University of Western Sydney. In accordance with ethics approval, written and oral consent was obtained from all mothers in the study.</p>
<sec id="s2a1">
<title>Participants</title>
<p>Forty infants participated in
<xref ref-type="sec" rid="s2">Experiment 1</xref>
<bold>.</bold>
For the full-spectrum auditory speech group, the final sample consisted of twenty 8-month-old infants (
<italic>M</italic>
age = 34.5 weeks;
<italic>range = </italic>
33.0−37.6 weeks; 13 males). Only infants who looked to each side for at least 2 seconds in at least 4 trials, and paid attention to both visual displays for at least 50% of the test period were included in the analysis. On this criterion, 5 additional infants were excluded due to lack of attention. For the intonation-only auditory speech group, the final sample consisted of twenty 8-month-old infants (
<italic>M</italic>
age = 35 weeks;
<italic>range = </italic>
32.7−38.6 weeks; 12 males). Six additional infants were excluded because they looked less than 50% of the time. Parental reports indicated that all infants were born full-term, were healthy at the time of testing, had no history of chronic or acute ear infections, and passed newborn hearing screening.</p>
</sec>
<sec id="s2a2">
<title>Speech stimuli</title>
<p>The speech stimuli consisted of a three pairs of sentences, one pair from each of the three female talkers (see
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
). The sentences were adapted from the IEEE list of phonetically balanced sentences
<xref rid="pone.0111467-Measurements1" ref-type="bibr">[38]</xref>
, and were “Yum, clams are round, small, soft and tasty” (9 syllables); “Look, the big red apple fell to the ground” (10 syllables); “It's tea time, let's feed the white mouse some flower seeds” (12 syllables) and the utterance “Did you know, Woolly is a sheep” (8 syllables). Adaptations to facilitate a more infant-directed style to the sentences were for instance, adding the word “Yum” and “Look” to the beginning of the utterances. From the list of sentences for each of the mother, pairs of sentences were selected that matched in duration but differed in syllable number. For some of the original sentence pairs there were small differences in duration (<500 ms). In these situations, the longer utterance was slowed, and the shorter utterance made slightly faster using Adobe Premier. The sentences and syllable differences for each female are shown in
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
.</p>
<p>In Experiment 1, the point-line displays remained unchanged for both groups of infants, but the auditory component of each sentence was either presented as full spectrum speech or was low-pass filtered at 400 Hz in Praat using a Hann filter. This removes auditory segmental information but leaves prosodic cues such as intonation, and duration relatively intact.</p>
</sec>
<sec id="s2a3">
<title>Speech Recording Procedure</title>
<p>Three females were recorded using a Northern Digital Optotrak 3020 system and a Sennheiser MKH416P48U-3 floor microphone. Each female was a mother who sat facing her infant during the recording session. She was instructed to speak a list of sentences as she would if she were reading to her infant at home. Optotrak markers were attached to the female’s face-internal features and to a head rig to measure rigid head movement. There were 21 non-rigid face internal markers on the eyebrows (4), cheeks (6), mouth (8) and jaw (3); and 3 markers on the head rig. Each frame of speech movement was extracted from the raw marker positions, and represented in terms of its displacement from the first frame. Data reduction was achieved by using principle components (PC) analysis; the PC’s coefficients were used to generate point-line visual speech motion. An animated schema of the face was generated by displaying the PC’s coefficients generated in Matlab using a marker visualization tool. Rendering was achieved in Togl and OpenGL
<xref rid="pone.0111467-Vignali1" ref-type="bibr">[39]</xref>
, and each frame was written to an image file and compiled to a video sequence at 25 fps. This line-joined version of visual speech did not capture fine detail because the motion capture markers were placed on the outside of the lips and so closure was never complete (see
<xref ref-type="supplementary-material" rid="pone.0111467.s001">Movies S1</xref>
and
<xref ref-type="supplementary-material" rid="pone.0111467.s002">S2</xref>
). Such a sparse representation was chosen to emphasize rhythmic speech motion and the perception of simple onsets and offsets. Nevertheless, the motion of the display did correlate with auditory properties as measured over a rendition. So for example, we measured how changes in speech amplitude (RMS energy) and mean F0 correlated with the rigid motion of the display (in the X and Y axes) and with the motion of the eyebrows, mouth (lip height & width) and jaw. For each of the sentences, at least one correlation coefficient was 0.45 or greater indicating that there was always a significant relationship between some of the auditory and visual features.</p>
</sec>
<sec id="s2a4">
<title>Testing Procedure</title>
<p>Each infant was tested with one of the three talkers' sentence pairs (see
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
). Infants were tested sitting on their parents’ lap in a darkened sound-attenuated room facing three computer monitors - a left, right and middle screen placed on a shelf in front of them. The infant’s gaze was centered at the start of all trials with an attention-getting looming video presented on the middle screen. Because the visual stimuli were unfamiliar point-line displays, we initially played the infants four repetitions of one sentence on the left monitor, and four repetitions of the other sentence on the right monitor. In the test trials, the visual sentence displays were played either on the same side or the opposite side to presentation in the initial phase. There were six test trials in which the auditory-only target sentence played through a central speaker with the two silent point-line displays presented simultaneously on the left and right monitors. We counterbalanced the target sentence so that half the infants heard one sentence in the pair and the other half heard the other sentence in the pair. The target sentence was also counterbalanced so that it was displayed on the left or right side equally. Each of the six test trials lasted for seven repetitions of each sentence. A digital video camera placed midline facing the infant, was connected to a video monitor in the control room and used by the experimenter to judge the infant's head and eye movements (left, right or no monitor) in real time. The experimenter used a computer keyboard to input the timing of the infant’s looking behavior: The left arrow key was pressed when the infant looked to the left monitor; the right arrow key when the infant looked to the right monitor, and the down arrow key when the infant was fixating the central monitor between trials. No key was pressed when the infant looked elsewhere. The experimenter was blind to the experimental manipulations, including the side of the matching visual display. To diminish parental influence, the same utterances with all the silences removed were played to parents using a Panasonic portable CD player and AKG studio headphones.</p>
</sec>
</sec>
<sec id="s2b">
<title>Results and Discussion</title>
<p>The extent to which infants fixated the matching visual display was used to index the degree to which 8-month-olds detected congruency between the auditory and visual speech signals. Due to each sentence pair (1, 2 and 3) having different durations (see
<xref ref-type="table" rid="pone-0111467-t001">Table 1</xref>
), the dependent variable was fixation duration as a percentage of total test trial duration for each sentence pair. Assumptions of normality were tested and met, both Shapiro-Wilk and Kolmogorov-Smirnov tests proved non-significant (
<italic>p</italic>
s>0.1). For both the full-spectrum and intonation-only auditory speech groups, the preliminary ANOVAs testing the effects of the three counterbalancing variables, [(i) same or different side for initial trial and test trials, (ii) left or right presentation of visual target sentence and (iii) target auditory sentence 1 or 2 in pair] proved non-significant. We use partial eta
<sup>2</sup>
to report effect size.</p>
<p>A 2 × 3 × ANOVA testing group (full-spectrum vs. intonation-only) and syllable difference (2, 3, or 4 syllables) between participants and trial (matching vs. mismatching) within participants revealed significant main effects for trial
<italic>F</italic>
(1,34) = 14.6,
<italic>p</italic>
<0.001;
<italic>η
<sub>p</sub>
<sup>2</sup>
</italic>
 = 0.3; groups
<italic>F</italic>
(1,34) = 17.8,
<italic>p</italic>
<0.0001;
<italic>η
<sub>p</sub>
<sup>2</sup>
</italic>
 = 0.3 and syllable number
<italic>F</italic>
(1,34) = 3.9,
<italic>p</italic>
<0.03,
<italic>η
<sub>p</sub>
<sup>2</sup>
</italic>
 = 0.19. There were no significant 2- or 3-way interactions (
<italic>p</italic>
s>0.23). Raw looking time scores (seconds) were also analyzed (ANOVA) with the same main effects and interactions. Irrespective of whether infants heard full-spectrum or intonation-only speech, 8-month-olds looked significantly longer to the matching than mismatching display. In addition, infants found low-pass filtered speech more interesting than full-spectrum speech in the context of this task. Thus, despite the preference for matching displays by both groups, overall percentage looking time in the intonation-only group (80%) was greater than in the full-spectrum group (61%).</p>
<p>The main effect for syllable difference showed that, irrespective of visual preference, infants looked less overall in the 3-syllable condition than the 2- and 4-syllable conditions. No firm conclusions can be drawn from this finding, as syllable difference was conflated with female talker, that is, the female talker who had the lowest percentage looking time was also the talker who displayed the least motion when talking to her infant.</p>
<p>Importantly, infants spent a greater percentage of trial duration looking to the matching (
<italic>M</italic>
 = 38.6%) than the mismatching display (
<italic>M</italic>
 = 22.8%) in the full spectrum group and to the matching (
<italic>M</italic>
 = 48.4%) than mismatching display (
<italic>M</italic>
 = 31.5%) in the intonation-only group (see
<xref ref-type="fig" rid="pone-0111467-g002">Figure 2</xref>
). Of the 20 infants in the full-spectrum group, 17 infants looked longer at the matching than mismatching point-line display (
<italic>binomial p</italic>
 = .001). In the intonation-only group, 15 of the 20 infants, looked longer at the matching than mismatching visual display (
<italic>binomial p</italic>
 = .01).</p>
<fig id="pone-0111467-g002" orientation="portrait" position="float">
<object-id pub-id-type="doi">10.1371/journal.pone.0111467.g002</object-id>
<label>Figure 2</label>
<caption>
<title>Mean percentage fixation to the matching and mismatching point-line displays for the full spectrum and prosody-only groups in
<xref ref-type="sec" rid="s2">Experiment 1</xref>
.</title>
<p>Error bars = 1 standard error of the mean.</p>
</caption>
<graphic xlink:href="pone.0111467.g002"></graphic>
</fig>
<p>In sum, the results demonstrate that under conditions in which prosody is exaggerated in both auditory and visual domains, infants have no difficulty perceiving congruent temporal and spatial speech patterns across modalities, even when the visual signal consists only of animated markers of a talking face, and the speech is low-pass filtered. That infants perceive congruence even without the benefit of fine-grain phonetic information in the visual and/or auditory displays, suggests that the AV match occurred at the prosodic level (e.g., phrasal intonation and/or syllabic rhythm). Here, it should be noted that the filtered auditory speech maintained prosodic information such as intonation, syllable rhythm, word stress etc. The degraded visual speech also included information in mouth/jaw movements at a local rhythmic level, and other global prosodic groupings in the head movements. So, it is not clear from these results whether it is global prosodic movement in the head or local syllabic rhythm from the mouth that is recognized and matched across modalities. It is possible that coarse grain articulatory information was recovered from the mouth/jaw motion. To resolve this issue, it is necessary to separate this information in the visual domain, and examine whether infants can match auditory and visual speech signals.</p>
</sec>
</sec>
<sec id="s3">
<title>Experiment 2: Rigid versus Non-Rigid Motion</title>
<p>Experiment 2 presented one group of infants with (i) rigid-only head motion only holding non-rigid motion constant; and another group with (ii) non-rigid face-internal motion only holding rigid motion constant. The aim was to examine separately, the contribution that (i) global prosody and (ii) the local property of syllable sequencing may make to the infant's ability to match auditory and visual connected speech. The rigid-only motion condition tested whether infants could detect the congruence between global prosodic properties, such as intonation, word stress and/or phrasal rhythm conveyed by head motion and the voice. The non-rigid condition, on the other hand, tested whether infants were sensitive to the temporal sequencing of the seen and heard syllables found in the non-rigid movements of the mouth/jaw. The effect of temporal synchrony is powerful, as it coordinates both temporal and spatial aspects of seen and heard speech events, and evidence shows that infants detect the congruence between lip movements/shape and speech sounds at the syllable level
<xref rid="pone.0111467-Kuhl1" ref-type="bibr">[8]</xref>
,
<xref rid="pone.0111467-Lewkowicz1" ref-type="bibr">[40]</xref>
; and have a proclivity for the synchrony between voice and lips in longer stretches of speech
<xref rid="pone.0111467-Dodd1" ref-type="bibr">[15]</xref>
.</p>
<sec id="s3a">
<title>Method and Materials</title>
<sec id="s3a1">
<title>Participants</title>
<p>The final sample consisted of 40 8-month-old infants. None of these infants participated in Experiment 1. In the rigid-only condition there were 20 infants (
<italic>M</italic>
age = 34 weeks;
<italic>range = </italic>
32.3–36.8 weeks; 9 males). Six additional infants were excluded due to lack of attention (4) and fixation to one side (2). In the non-rigid condition, there were 20 infants (
<italic>M</italic>
age = 34.4 weeks;
<italic>range = </italic>
33.1−38 weeks; 12 males). Six additional infants were excluded due to lack of attention (5) and mother interference (1). Parental reports indicated that all infants were born full-term, were healthy at the time of testing, had no history of chronic or acute ear infections, and passed newborn hearing screening.</p>
</sec>
<sec id="s3a2">
<title>Stimuli</title>
<p>The same sentence stimuli were used as in
<xref ref-type="sec" rid="s2">Experiment 1</xref>
. The visual displays were designed to assess the contribution of rigid and non-rigid movement to perceptual matching by independently testing the effects of global head motion and local face cues, respectively
<xref rid="pone.0111467-Vignali1" ref-type="bibr">[39]</xref>
. Rigid-only head motion was presented to one group of infants, and non-rigid face-internal movement was presented to the other group (see
<xref ref-type="supplementary-material" rid="pone.0111467.s003">Movies S3</xref>
and
<xref ref-type="supplementary-material" rid="pone.0111467.s004">S4</xref>
).</p>
</sec>
<sec id="s3a3">
<title>Procedure</title>
<p>The same infant procedure described in
<xref ref-type="sec" rid="s2">Experiment 1</xref>
was used.</p>
</sec>
</sec>
<sec id="s3b">
<title>Results and Discussion</title>
<p>As in
<xref ref-type="sec" rid="s2">Experiment 1</xref>
, the percentage of time infants spent fixating the matching visual display as proportion of total trial length was used to index the degree to which infants matched auditory visual ID speech in rigid and non-rigid conditions. Partial eta
<sup>2</sup>
is used to report effect size. Again, assumptions of normality were met prior to conducting ANOVAs. The preliminary ANOVAs testing counterbalancing effects were non-significant.</p>
<p>A 2 × 3 × (2) ANOVA testing group (rigid vs. non-rigid) and syllable difference (2, 3, or 4 syllables) between participants and trial (matching vs. mismatching) within participants revealed no significant main effects for trial, group or syllable number (
<italic>p</italic>
s>0.17). However, there was significant trial x group interaction
<italic>F</italic>
(1,34) = 7.1,
<italic>p</italic>
<0.01;
<italic>η
<sub>p</sub>
<sup>2</sup>
</italic>
 = 0.18. No other interactions were significant (
<italic>p</italic>
s>0.32). In the rigid group, infants spent a greater percentage of trial duration looking to the matching (
<italic>M</italic>
 = 46.23%) than the mismatching display (
<italic>M</italic>
 = 31.9%) while in the non-rigid group infants preferred the mismatching (
<italic>M</italic>
 = 48.7%) over the matching display (
<italic>M</italic>
 = 30.1%) (see
<xref ref-type="fig" rid="pone-0111467-g003">Figure 3</xref>
). Raw looking time scores (seconds) were also analysed using ANOVA with the same results. Binomial tests were significant. In the rigid condition 17 of the 20 infants looked longer to the matching than the mismatching display (
<italic>binomial p</italic>
 = .001) and in the non-rigid condition 14 of the 20 infants preferred the mismatching over the matching display (
<italic>binomial p</italic>
 = .04).</p>
<fig id="pone-0111467-g003" orientation="portrait" position="float">
<object-id pub-id-type="doi">10.1371/journal.pone.0111467.g003</object-id>
<label>Figure 3</label>
<caption>
<title>Mean percentage fixation to the matching and mismatching point-line displays in the rigid only and the non-rigid only groups in
<xref ref-type="sec" rid="s3">Experiment 2</xref>
.</title>
<p>Error bars = 1 standard error of the mean.</p>
</caption>
<graphic xlink:href="pone.0111467.g003"></graphic>
</fig>
<p>Importantly the results revealed that in the rigid group, infants looked longer to the matched head and voice, whereas the non-rigid group showed a preference for the mismatched face and voice (see
<xref ref-type="fig" rid="pone-0111467-g003">Figure 3</xref>
). The results of the rigid-only condition were similar to the results of
<xref ref-type="sec" rid="s2">Experiment 1</xref>
, indicating that infants were able to match the seen and heard speech based on global prosodic information without the benefit of articulatory information. The absence of mouth motion in the rigid-only condition strongly suggests that these results are not based on low-level synchrony detection driven by the perception of synchronous energy onsets and offsets (see
<xref rid="pone.0111467-Lewkowicz2" ref-type="bibr">[41]</xref>
), but by the correlated acoustic and visual dynamics of a talking head.</p>
</sec>
</sec>
<sec id="s4">
<title>General Discussion</title>
<p>The results show that 8-month-old infants preferentially fixate matching over mismatching talking faces whether the auditory speech is presented as full spectrum or low-pass filtered at 400 Hz (
<xref ref-type="sec" rid="s2">Experiment 1</xref>
). Infants did this from the kinematics of point-line displays of talking faces/heads and without the benefit of fine phonetic detail in the visual displays or phonetic detail in the auditory speech, suggesting that performance was based upon matching coarse-grain prosodic information from both modalities. Moreover, the same preference for matched displays was found when rigid head motion alone was presented with facial motion held constant (
<xref ref-type="sec" rid="s3">Experiment 2</xref>
), indicating that detecting an auditory-visual speech match does not necessarily depend on the temporal relationship of lip movements and voice in the two modalities. Indeed, when restricted to talking faces displaying lip/jaw motion, the infants in the non-rigid group preferred the mismatched to the matched display. Eight-month-olds could clearly discriminate the non-rigid visual displays, and were evidently sensitive to correlated auditory and visual attributes of local (syllable) rhythmic cues in connected speech. Nonetheless, as we predicted, infants were more sensitive when dependent on global prosodic cues than the auditory-visual correlates occurring at the local rhythmic level.</p>
<p>Overall, the results support the proposition that infants perceptually prioritize intersensory signals that share common properties, particularly the prosodic gestures from head movement. The paradox is, however, that with face-only information, infants prefer incongruent syllable sequences across auditory and visual modalities. Typically infants can recognise auditory and visual correspondences in single syllable stimuli
<xref rid="pone.0111467-Kuhl1" ref-type="bibr">[8]</xref>
,
<xref rid="pone.0111467-Kuhl2" ref-type="bibr">[9]</xref>
,
<xref rid="pone.0111467-Patterson2" ref-type="bibr">[10]</xref>
or the same sentence presented in and out of synchrony
<xref rid="pone.0111467-Dodd1" ref-type="bibr">[15]</xref>
,
<xref rid="pone.0111467-Lewkowicz2" ref-type="bibr">[41]</xref>
. However, when presented with two point-line talking faces and a sentence, they preferentially attend to the mismatched face. It is unclear why infants looked longer to the mismatched lips and voice displays in the non-rigid condition. It is possible that with a lack of rigid head motion, the motion of the joined segments (e.g., mouth, jaw) were no longer perceived as a gestalt talking head, and the separation of motion acted to increase the difficulty of the non-rigid matching task. It seems that adults can identify speech better when non-rigid motion is accompanied by the appropriate rigid movements than when it is not
<xref rid="pone.0111467-Munhall2" ref-type="bibr">[42]</xref>
. Even for adults, it seems that the visual prosody signaled by head motion matters.</p>
<p>Compared to the non-rigid stimuli, the rigid-only stimuli provided a task with low cognitive demand, as it was more visually expressive, making it easier for infants to match the acoustic and visual correlates of sentences when the dominant feature was prosodic variation, and not smaller, more rapidly realised sequences of localised syllable size events. Not only does prosody evolve more slowly over time, but it is a relatively primitive mode of communication, and as such easier to perceive and to be decoded non-linguistically. Thus, it is in the non-rigid condition, that the effects of uncoupling rigid and non-rigid elements into separate motions are most marked, and shown in the infants' attention being drawn to the mismatching display. We suggest that our findings implicate firstly, the rate of prosodic change and secondly, the type of prosody. Prosody can be either linguistic as it conveys linguistic distinctions, such as word boundaries, syntactic structure and lexical meaning; or it can be emotional, where it serves to mark intent or attitude involved in the emotional or social aspects of language.</p>
<p>One possibility is that this significant looking preference to the unexpected pairing represents an instance of a shift from a familiarity to a novelty preference
<xref rid="pone.0111467-HoustonPrice1" ref-type="bibr">[43]</xref>
. That is, typically infants will prefer to fixate on stimuli that are familiar to them. This preference also operates with cross-modal stimuli, i.e., infants tend to fixate a visual stimulus that matches an auditory one
<xref rid="pone.0111467-Golinkoff1" ref-type="bibr">[44]</xref>
. In the current study, we suggest that familiarity for the infants is defined at the level of rigid head motion for the visual stimuli and rhythm (prosody) for the auditory ones. This is because, for infants these properties are the most salient; exaggerated rigid head motion characterizes infant-directed speech
<xref rid="pone.0111467-Kim1" ref-type="bibr">[26]</xref>
as does exaggerated prosody for auditory speech. However, when rigid head motion is removed and face-only motion presented, the familiarity of the visual stimulus is reduced. In this case, a preference for the more novel of the cross-modal pairs (mismatch) may arise. This tendency to pay attention to a mismatch between face (mouth) motion and speech when the salient head motion signal has been removed may be because infants between 4 and 8 months tend to pay more attention to the mouth region when viewing AV speech stimuli
<xref rid="pone.0111467-Lewkowicz3" ref-type="bibr">[45]</xref>
. That is, finding an AV mismatch may be particularly compelling. Whatever is the correct interpretation of the unexpected looking preference, it is clear that this shows that infants can detect the mismatch between non-rigid face motion and speech.</p>
<p>Our findings for
<xref ref-type="sec" rid="s3">Experiment 2</xref>
in which infants match global prosody cues (slowly fluctuating rigid head motion plus auditory change) and mismatch local syllable cues (more rapid non-rigid motion and auditory change), give rise to two questions. First, when do infants acquire the ability to match auditory syllable sequences to the congruent talking face without the benefit of rigid motion? To ensure that adults could perform the task, we tested 24 adults using the same stimuli and experimental set up. Adults were found to easily perceive the match in non-rigid condition and indeed performed better in the non-rigid than the rigid condition. In this study, we asked adults (i) to judge which of the two visual displays matched the auditory sentence; and (ii) how confident they were with their choice. Adults correctly identified the matching face in both rigid and non-rigid conditions, and were more confident with their non-rigid than rigid choice (
<italic>p</italic>
 = .01).</p>
<p>For 8-month-old infants, sensitivity to visual prosody conveyed by rigid motion is well developed, but sensitivity to locally produced syllable sequences, is subject to developmental improvement, and should emerge at a later age. By what age is unclear but it would be expected that this should occur in the months following, and perhaps co-occur with the development of language-specific speech perception. Second, by what mechanism do infants detect congruence between auditory and visual displays? It has been suggested that infants' detection of redundancy is based on intersensory responsiveness to temporal synchrony, and that this emerges first, followed by the emergence of responsiveness to duration, rate, and rhythm
<xref rid="pone.0111467-Lewkowicz4" ref-type="bibr">[46]</xref>
. It seems that 10- to 11-month-olds cannot detect visual prosody based on temporal synchrony. Blossom and Morgan found infants could only match face/head and voice (with the mouth was obscured) when different sentences were used, but not when temporal synchrony was disrupted
<xref rid="pone.0111467-Blossom1" ref-type="bibr">[16]</xref>
. Therefore the mechanism for detecting congruency between auditory and visual prosody may involve perceiving amodal relations that include multiple cues, not just temporal synchrony but also relating to differences in rhythm and rate. That is, infants rely on a combination of cues. Central to this we suggest that the primary cue is the slower rate of global prosody over local rhythm.</p>
<p>The capacity to perceive the prosodic alignment of auditory and visual components of speech in this study suggests that infants use the voice and head kinematics of prosody to extract information about the structure of speech. One of the basic requirements for learning language is to segment fluent speech into meaningful units, initially clauses and phrases and then words. The development of such abilities may evolve from an initial ability to parse the wave form using coarse grain prosodic features carried at the global prosodic level across modalities. Furthermore, such capacity might assist infants in establishing a perception-production link - one that starts at the global level of intonation in rigid head motion, and then incorporating the rhythmic motion of the mouth and jaw to gradually develop into something more fine-grained that involves phonetic distinctions in continuous speech. The future direction of this work is to examine the developmental trajectory of auditory visual speech perception. In sum, the evidence is accumulating for a sophistication in the perception of auditory-visual speech information during the first year of postnatal life - at the level of speech segments
<xref rid="pone.0111467-Patterson1" ref-type="bibr">[7]</xref>
,
<xref rid="pone.0111467-Lewkowicz4" ref-type="bibr">[46]</xref>
,
<xref rid="pone.0111467-Desjardins1" ref-type="bibr">[47]</xref>
; and at the level of prosody in continuous speech, another cornerstone in the scaffold that underpins speech development.</p>
</sec>
<sec sec-type="supplementary-material" id="s5">
<title>Supporting Information</title>
<supplementary-material content-type="local-data" id="pone.0111467.s001">
<label>Movie S1</label>
<caption>
<p>
<bold>Side by side displays of sentence pairs for full spectrum condition, labeled ‘normal'.</bold>
</p>
<p>(AVI)</p>
</caption>
<media xlink:href="pone.0111467.s001.avi">
<caption>
<p>Click here for additional data file.</p>
</caption>
</media>
</supplementary-material>
<supplementary-material content-type="local-data" id="pone.0111467.s002">
<label>Movie S2</label>
<caption>
<p>
<bold>Side by side displays of sentence pairs for low pass filtered condition, labeled ‘prosody'.</bold>
</p>
<p>(AVI)</p>
</caption>
<media xlink:href="pone.0111467.s002.avi">
<caption>
<p>Click here for additional data file.</p>
</caption>
</media>
</supplementary-material>
<supplementary-material content-type="local-data" id="pone.0111467.s003">
<label>Movie S3</label>
<caption>
<p>
<bold>Side by side displays of sentence pairs for rigid condition, labeled ‘rigid'.</bold>
</p>
<p>(AVI)</p>
</caption>
<media xlink:href="pone.0111467.s003.avi">
<caption>
<p>Click here for additional data file.</p>
</caption>
</media>
</supplementary-material>
<supplementary-material content-type="local-data" id="pone.0111467.s004">
<label>Movie S4</label>
<caption>
<p>
<bold>Side by side displays of sentence pairs for non-rigid condition, labeled ‘non-rigid'.</bold>
</p>
<p>(AVI)</p>
</caption>
<media xlink:href="pone.0111467.s004.avi">
<caption>
<p>Click here for additional data file.</p>
</caption>
</media>
</supplementary-material>
</sec>
</body>
<back>
<ack>
<p>The authors would like to thank Christopher Davis for his helpful comments on the manuscript and technical assistance with the stimuli. The authors also acknowledge support from the Australian Research Council (DP120104298, DP0666857).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="pone.0111467-Sumby1">
<label>1</label>
<mixed-citation publication-type="journal">
<name>
<surname>Sumby</surname>
<given-names>WH</given-names>
</name>
,
<name>
<surname>Pollack</surname>
<given-names>I</given-names>
</name>
(
<year>1954</year>
)
<article-title>Visual contribution to speech intelligibility in noise</article-title>
.
<source>The Journal of the Acoustical Society of America</source>
<volume>26</volume>
:
<fpage>212</fpage>
<lpage>215</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Davis1">
<label>2</label>
<mixed-citation publication-type="journal">
<name>
<surname>Davis</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Kim</surname>
<given-names>J</given-names>
</name>
(
<year>2004</year>
)
<article-title>Audiovisual interactions with intact clearly audible speech</article-title>
.
<source>The Quarterly Journal of Experimental Psychology</source>
<volume>57A</volume>
:
<fpage>1103</fpage>
<lpage>1121</lpage>
.
<pub-id pub-id-type="pmid">15370518</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Cvejic1">
<label>3</label>
<mixed-citation publication-type="journal">
<name>
<surname>Cvejic</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Kim</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Davis</surname>
<given-names>C</given-names>
</name>
(
<year>2010</year>
)
<article-title>Prosody off the top of the head: Prosodic contrasts can be discriminated by head motion</article-title>
.
<source>Speech Communication</source>
<volume>52</volume>
:
<fpage>555</fpage>
<lpage>564</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Davis2">
<label>4</label>
<mixed-citation publication-type="journal">
<name>
<surname>Davis</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Kim</surname>
<given-names>J</given-names>
</name>
(
<year>2006</year>
)
<article-title>Audio-visual speech perception off the top of the head</article-title>
.
<source>Cognition</source>
<volume>100</volume>
:
<fpage>B21</fpage>
<lpage>B31</lpage>
.
<pub-id pub-id-type="pmid">16289070</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Swerts1">
<label>5</label>
<mixed-citation publication-type="journal">
<name>
<surname>Swerts</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Krahmer</surname>
<given-names>E</given-names>
</name>
(
<year>2008</year>
)
<article-title>Facial expression and prosodic prominence: Effects of modality and facial area</article-title>
.
<source>Journal of Phonetics</source>
<volume>36</volume>
:
<fpage>219</fpage>
<lpage>238</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Aldridge1">
<label>6</label>
<mixed-citation publication-type="journal">
<name>
<surname>Aldridge</surname>
<given-names>MA</given-names>
</name>
,
<name>
<surname>Braga</surname>
<given-names>ES</given-names>
</name>
,
<name>
<surname>Walton</surname>
<given-names>GE</given-names>
</name>
,
<name>
<surname>Bower</surname>
<given-names>TGR</given-names>
</name>
(
<year>1999</year>
)
<article-title>The intermodal representation of speech in newborns</article-title>
.
<source>Developmental Science</source>
<volume>2</volume>
:
<fpage>42</fpage>
<lpage>46</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Patterson1">
<label>7</label>
<mixed-citation publication-type="journal">
<name>
<surname>Patterson</surname>
<given-names>ML</given-names>
</name>
,
<name>
<surname>Werker</surname>
<given-names>JF</given-names>
</name>
(
<year>2003</year>
)
<article-title>Two-month-old infants match phonemic information in lips and voice</article-title>
.
<source>Developmental Science</source>
<volume>6</volume>
:
<fpage>191</fpage>
<lpage>196</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Kuhl1">
<label>8</label>
<mixed-citation publication-type="journal">
<name>
<surname>Kuhl</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Meltzoff</surname>
<given-names>A</given-names>
</name>
(
<year>1982</year>
)
<article-title>The bimodal perception of speech in infancy</article-title>
.
<source>Science</source>
<volume>218</volume>
:
<fpage>1138</fpage>
<lpage>1141</lpage>
.
<pub-id pub-id-type="pmid">7146899</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Kuhl2">
<label>9</label>
<mixed-citation publication-type="journal">
<name>
<surname>Kuhl</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Meltzoff</surname>
<given-names>A</given-names>
</name>
(
<year>1984</year>
)
<article-title>The intermodal representation of speech in infants</article-title>
.
<source>Infant Behavior and Development</source>
<volume>7</volume>
:
<fpage>361</fpage>
<lpage>381</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Patterson2">
<label>10</label>
<mixed-citation publication-type="journal">
<name>
<surname>Patterson</surname>
<given-names>ML</given-names>
</name>
,
<name>
<surname>Werker</surname>
<given-names>JF</given-names>
</name>
(
<year>1999</year>
)
<article-title>Matching phonetic information in lips and voice is robust in 4.5-month-old infants</article-title>
.
<source>Infant Behavior & Development</source>
<volume>22</volume>
:
<fpage>237</fpage>
<lpage>247</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Patterson3">
<label>11</label>
<mixed-citation publication-type="journal">
<name>
<surname>Patterson</surname>
<given-names>ML</given-names>
</name>
,
<name>
<surname>Werker</surname>
<given-names>JF</given-names>
</name>
(
<year>2002</year>
)
<article-title>Infants' ability to match dynamic phonetic and gender information in the face and voice</article-title>
.
<source>Journal of Experimental Child Psychology</source>
<volume>81</volume>
:
<fpage>93</fpage>
<lpage>115</lpage>
.
<pub-id pub-id-type="pmid">11741376</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Kuhl3">
<label>12</label>
<mixed-citation publication-type="journal">
<name>
<surname>Kuhl</surname>
<given-names>P</given-names>
</name>
,
<name>
<surname>Williams</surname>
<given-names>KA</given-names>
</name>
,
<name>
<surname>Meltzoff</surname>
<given-names>AN</given-names>
</name>
(
<year>1991</year>
)
<article-title>Cross-modal speech perception in adults and infants using nonspeech auditory stimulus</article-title>
.
<source>Journal of Experimental Psychology</source>
<volume>17</volume>
:
<fpage>829</fpage>
<lpage>840</lpage>
.
<pub-id pub-id-type="pmid">1834794</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Soken1">
<label>13</label>
<mixed-citation publication-type="journal">
<name>
<surname>Soken</surname>
<given-names>NH</given-names>
</name>
,
<name>
<surname>Pick</surname>
<given-names>AD</given-names>
</name>
(
<year>1992</year>
)
<article-title>Intermodal perception of happy and angry expressive behaviors by seven-month-old infants</article-title>
.
<source>Child Development</source>
<volume>63</volume>
:
<fpage>787</fpage>
<lpage>795</lpage>
.
<pub-id pub-id-type="pmid">1505240</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-WalkerAndrews1">
<label>14</label>
<mixed-citation publication-type="journal">
<name>
<surname>Walker-Andrews</surname>
<given-names>AS</given-names>
</name>
(
<year>1997</year>
)
<article-title>Infants' perception of expressive behaviours: Differentiation of multimodal information</article-title>
.
<source>Psychological Bulletin</source>
<volume>121</volume>
:
<fpage>437</fpage>
<lpage>456</lpage>
.
<pub-id pub-id-type="pmid">9136644</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Dodd1">
<label>15</label>
<mixed-citation publication-type="journal">
<name>
<surname>Dodd</surname>
<given-names>B</given-names>
</name>
(
<year>1979</year>
)
<article-title>Lip reading in infants: Attention to speech presented in- and out-of-synchrony</article-title>
.
<source>Cognitive Psychology</source>
<volume>11</volume>
:
<fpage>478</fpage>
<lpage>484</lpage>
.
<pub-id pub-id-type="pmid">487747</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Blossom1">
<label>16</label>
<mixed-citation publication-type="other">Blossom M, Morgan J (2006) Does the face say what the mouth says? A study of infants’ sensitivity to visual prosody. In: In D. Bamman TM, & C. Zaller (Eds.) editor. 30th Annual Boston University Conference on Language Development. Boston: Somerville, MA: Cascadilla Press.</mixed-citation>
</ref>
<ref id="pone.0111467-Kitamura1">
<label>17</label>
<mixed-citation publication-type="journal">
<name>
<surname>Kitamura</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Burnham</surname>
<given-names>D</given-names>
</name>
(
<year>2003</year>
)
<article-title>Pitch and communicative intent in mothers' speech: Adjustments for age and sex in the first year</article-title>
.
<source>Infancy</source>
<volume>4</volume>
:
<fpage>85</fpage>
<lpage>110</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Kitamura2">
<label>18</label>
<mixed-citation publication-type="journal">
<name>
<surname>Kitamura</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Thanavisuth</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Luksaneeyanawin</surname>
<given-names>S</given-names>
</name>
,
<name>
<surname>Burnham</surname>
<given-names>D</given-names>
</name>
(
<year>2002</year>
)
<article-title>Universality and specificity in infant-directed speech: Pitch modifications as a function of infant age and sex in a tonal and non-tonal language</article-title>
.
<source>Infant Behavior and Development</source>
<volume>24</volume>
:
<fpage>372</fpage>
<lpage>392</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Fernald1">
<label>19</label>
<mixed-citation publication-type="journal">
<name>
<surname>Fernald</surname>
<given-names>A</given-names>
</name>
(
<year>1985</year>
)
<article-title>Four-month-old infants prefer to listen to motherese</article-title>
.
<source>Infant Behavior and Development</source>
<volume>8</volume>
:
<fpage>181</fpage>
<lpage>195</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-PannetonCooper1">
<label>20</label>
<mixed-citation publication-type="journal">
<name>
<surname>Panneton Cooper</surname>
<given-names>R</given-names>
</name>
,
<name>
<surname>Aslin</surname>
<given-names>RN</given-names>
</name>
(
<year>1990</year>
)
<article-title>Preference for infant-directed speech in the first month after birth</article-title>
.
<source>Child Development</source>
<volume>61</volume>
:
<fpage>1584</fpage>
<lpage>1595</lpage>
.
<pub-id pub-id-type="pmid">2245748</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Theissen1">
<label>21</label>
<mixed-citation publication-type="journal">
<name>
<surname>Theissen</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Hill</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Saffran</surname>
<given-names>JR</given-names>
</name>
(
<year>2005</year>
)
<article-title>Infant directed speech facilitates word segmentation</article-title>
.
<source>Infancy</source>
<volume>7</volume>
:
<fpage>53</fpage>
<lpage>71</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Fernald2">
<label>22</label>
<mixed-citation publication-type="journal">
<name>
<surname>Fernald</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Mazzie</surname>
<given-names>C</given-names>
</name>
(
<year>1991</year>
)
<article-title>Prosody and focus in speech to infants and adults</article-title>
.
<source>Developmental Psychology</source>
<volume>27</volume>
:
<fpage>209</fpage>
<lpage>221</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Karzon1">
<label>23</label>
<mixed-citation publication-type="other">Karzon R (1985) Discrimination of polysyllabic sequences by one to four month old infants. Journal of Experimental Child Psychology 39.</mixed-citation>
</ref>
<ref id="pone.0111467-KemlerNelson1">
<label>24</label>
<mixed-citation publication-type="other">Kemler Nelson DG, Hirsh-Pasek K, Jusczyk PW, Wright Cassidy K (1989) How the prosodic cues in motherese might assist language learning. Journal of Child Language 16.</mixed-citation>
</ref>
<ref id="pone.0111467-Green1">
<label>25</label>
<mixed-citation publication-type="journal">
<name>
<surname>Green</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Nip</surname>
<given-names>I</given-names>
</name>
,
<name>
<surname>Wilson</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Mefferd</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Yunusova</surname>
<given-names>Y</given-names>
</name>
(
<year>2010</year>
)
<article-title>Lip movement exaggerations during infant-directed speech</article-title>
.
<source>Journal of Speech, Language and Hearing Research</source>
<volume>53</volume>
:
<fpage>1529</fpage>
<lpage>1542</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Kim1">
<label>26</label>
<mixed-citation publication-type="other">Kim J, Davis C, Kitamura C (2012) Auditory-visual speech to infants and adults: signals and correlations. Proceedings of Interspeech: 1119–1122.</mixed-citation>
</ref>
<ref id="pone.0111467-Chong1">
<label>27</label>
<mixed-citation publication-type="journal">
<name>
<surname>Chong</surname>
<given-names>SCF</given-names>
</name>
,
<name>
<surname>Werker</surname>
<given-names>JF</given-names>
</name>
,
<name>
<surname>Russell</surname>
<given-names>JA</given-names>
</name>
,
<name>
<surname>Carroll</surname>
<given-names>JM</given-names>
</name>
(
<year>2003</year>
)
<article-title>Three facial expressions mothers direct to their infants</article-title>
.
<source>Infant and Child Development</source>
<volume>12</volume>
:
<fpage>211</fpage>
<lpage>232</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Arterberry1">
<label>28</label>
<mixed-citation publication-type="journal">
<name>
<surname>Arterberry</surname>
<given-names>ME</given-names>
</name>
,
<name>
<surname>Bornstein</surname>
<given-names>MH</given-names>
</name>
(
<year>2002</year>
)
<article-title>Infant perceptual and conceptual categorization: The roles of static and dynamic stimulus attributes</article-title>
.
<source>Cognition</source>
<volume>86</volume>
:
<fpage>1</fpage>
<lpage>24</lpage>
.
<pub-id pub-id-type="pmid">12208649</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Bardi1">
<label>29</label>
<mixed-citation publication-type="journal">
<name>
<surname>Bardi</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Regolin</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Simion</surname>
<given-names>F</given-names>
</name>
(
<year>2011</year>
)
<article-title>Biological motion preference in humans at birth: role of dynamic and configural properties</article-title>
.
<source>Developmental Science</source>
<volume>14</volume>
:
<fpage>353</fpage>
<lpage>359</lpage>
.
<pub-id pub-id-type="pmid">22213905</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-BidetIldei1">
<label>30</label>
<mixed-citation publication-type="journal">
<name>
<surname>Bidet-Ildei</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Kitromilides</surname>
<given-names>E</given-names>
</name>
,
<name>
<surname>Orliaguet</surname>
<given-names>J-P</given-names>
</name>
,
<name>
<surname>Pavlova</surname>
<given-names>M</given-names>
</name>
,
<name>
<surname>Gentaz</surname>
<given-names>E</given-names>
</name>
(
<year>2014</year>
)
<article-title>Preference for point-light human biological motion in newborns: contribution of translational displacement</article-title>
.
<source>Developmental Psychology</source>
<volume>50</volume>
:
<fpage>113</fpage>
<lpage>120</lpage>
.
<pub-id pub-id-type="pmid">23668800</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Simion1">
<label>31</label>
<mixed-citation publication-type="journal">
<name>
<surname>Simion</surname>
<given-names>F</given-names>
</name>
,
<name>
<surname>Regolin</surname>
<given-names>L</given-names>
</name>
,
<name>
<surname>Bulf</surname>
<given-names>H</given-names>
</name>
(
<year>2008</year>
)
<article-title>A predisposition for biological motion in the newborn baby</article-title>
.
<source>Proceedings of the National Academy of Science</source>
<volume>105</volume>
:
<fpage>809</fpage>
<lpage>813</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Bertoncini1">
<label>32</label>
<mixed-citation publication-type="journal">
<name>
<surname>Bertoncini</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Mehler</surname>
<given-names>J</given-names>
</name>
(
<year>1981</year>
)
<article-title>Syllables as units in infant speech perception</article-title>
.
<source>Infant Behavior and Development</source>
<volume>4</volume>
:
<fpage>247</fpage>
<lpage>260</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-BijeljacBabic1">
<label>33</label>
<mixed-citation publication-type="journal">
<name>
<surname>Bijeljac-Babic</surname>
<given-names>R</given-names>
</name>
,
<name>
<surname>Bertoncini</surname>
<given-names>J</given-names>
</name>
,
<name>
<surname>Mehler</surname>
<given-names>J</given-names>
</name>
(
<year>1993</year>
)
<article-title>How do 4-day-old infants categorise multisyllabic utterances?</article-title>
<source>Developmental Psychology</source>
<volume>29</volume>
:
<fpage>711</fpage>
<lpage>721</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Moon1">
<label>34</label>
<mixed-citation publication-type="journal">
<name>
<surname>Moon</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Bever</surname>
<given-names>T</given-names>
</name>
,
<name>
<surname>Fifer</surname>
<given-names>W</given-names>
</name>
(
<year>1992</year>
)
<article-title>Canonical and non-canonical syllable discrimination by 2-day-old infants</article-title>
.
<source>Journal of Child Language</source>
<volume>19</volume>
:
<fpage>1</fpage>
<lpage>17</lpage>
.
<pub-id pub-id-type="pmid">1551926</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Cutler1">
<label>35</label>
<mixed-citation publication-type="journal">
<name>
<surname>Cutler</surname>
<given-names>A</given-names>
</name>
,
<name>
<surname>Mehler</surname>
<given-names>J</given-names>
</name>
(
<year>1993</year>
)
<article-title>The periodicity bias</article-title>
.
<source>Journal of Phonetics</source>
<volume>21</volume>
:
<fpage>103</fpage>
<lpage>108</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Munhall1">
<label>36</label>
<mixed-citation publication-type="journal">
<name>
<surname>Munhall</surname>
<given-names>KG</given-names>
</name>
,
<name>
<surname>Jones</surname>
<given-names>JA</given-names>
</name>
(
<year>1998</year>
)
<article-title>Articulatory evidence for syllable structure</article-title>
.
<source>Brain and Behavioural Sciences</source>
<volume>21</volume>
:
<fpage>524</fpage>
<lpage>525</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Grant1">
<label>37</label>
<mixed-citation publication-type="journal">
<name>
<surname>Grant</surname>
<given-names>KW</given-names>
</name>
,
<name>
<surname>van Wassenhove</surname>
<given-names>V</given-names>
</name>
,
<name>
<surname>Poeppel</surname>
<given-names>D</given-names>
</name>
(
<year>2004</year>
)
<article-title>Detection of auditory (cross-spectral) and auditory-visual (cross-modal) synchrony</article-title>
.
<source>Speech Communication</source>
<volume>44</volume>
:
<fpage>43</fpage>
<lpage>53</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Measurements1">
<label>38</label>
<mixed-citation publication-type="other">Measurements IsoS (1969) IEEE recommended practices for speech quality measurements. 227–246. p.</mixed-citation>
</ref>
<ref id="pone.0111467-Vignali1">
<label>39</label>
<mixed-citation publication-type="other">Vignali G (2005) Analysis of 3D multivariable data of expressive speech motion; Japan. ATR.</mixed-citation>
</ref>
<ref id="pone.0111467-Lewkowicz1">
<label>40</label>
<mixed-citation publication-type="journal">
<name>
<surname>Lewkowicz</surname>
<given-names>D</given-names>
</name>
(
<year>2000</year>
)
<article-title>Infants' perception of the audible, visible, and bimodal atributes of multimodal syllables</article-title>
.
<source>Child Development</source>
<volume>71</volume>
:
<fpage>1241</fpage>
.
<pub-id pub-id-type="pmid">11108094</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Lewkowicz2">
<label>41</label>
<mixed-citation publication-type="journal">
<name>
<surname>Lewkowicz</surname>
<given-names>D</given-names>
</name>
(
<year>2010</year>
)
<article-title>Infant perception of audio-visual speech synchrony</article-title>
.
<source>Developmental Psychology</source>
<volume>46</volume>
:
<fpage>66</fpage>
<lpage>77</lpage>
.
<pub-id pub-id-type="pmid">20053007</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Munhall2">
<label>42</label>
<mixed-citation publication-type="journal">
<name>
<surname>Munhall</surname>
<given-names>KG</given-names>
</name>
,
<name>
<surname>Jones</surname>
<given-names>JA</given-names>
</name>
,
<name>
<surname>Callan</surname>
<given-names>DE</given-names>
</name>
,
<name>
<surname>Kuratate</surname>
<given-names>T</given-names>
</name>
,
<name>
<surname>Vatikiotis-Bateson</surname>
<given-names>E</given-names>
</name>
(
<year>2004</year>
)
<article-title>Visual prosody and speech intelligibility head movement improves auditory speech perception</article-title>
.
<source>Psychological Science</source>
<volume>15</volume>
:
<fpage>133</fpage>
<lpage>137</lpage>
.
<pub-id pub-id-type="pmid">14738521</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-HoustonPrice1">
<label>43</label>
<mixed-citation publication-type="journal">
<name>
<surname>Houston-Price</surname>
<given-names>C</given-names>
</name>
,
<name>
<surname>Nakai</surname>
<given-names>S</given-names>
</name>
(
<year>2004</year>
)
<article-title>Distinguishing novelty and familiarity effects in infant preference procedures</article-title>
.
<source>Infant and Child Development</source>
<volume>13</volume>
:
<fpage>341</fpage>
<lpage>348</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Golinkoff1">
<label>44</label>
<mixed-citation publication-type="journal">
<name>
<surname>Golinkoff</surname>
<given-names>R</given-names>
</name>
,
<name>
<surname>Hirsh-Pasek</surname>
<given-names>K</given-names>
</name>
,
<name>
<surname>Cauley</surname>
<given-names>K</given-names>
</name>
,
<name>
<surname>Gordon</surname>
<given-names>L</given-names>
</name>
(
<year>1987</year>
)
<article-title>The eyes have it: lexical and syntactic comprehension in a new paradigm</article-title>
.
<source>Journal of Child Language</source>
<volume>14</volume>
:
<fpage>23</fpage>
<lpage>45</lpage>
.
<pub-id pub-id-type="pmid">3558524</pub-id>
</mixed-citation>
</ref>
<ref id="pone.0111467-Lewkowicz3">
<label>45</label>
<mixed-citation publication-type="journal">
<name>
<surname>Lewkowicz</surname>
<given-names>D</given-names>
</name>
,
<name>
<surname>Hansen-Tift</surname>
<given-names>A</given-names>
</name>
(
<year>2012</year>
)
<article-title>Infant deploy selective attention to the mouth of a talking face when learning speech</article-title>
.
<source>Proceedings of the National Academy of Science</source>
<volume>109</volume>
:
<fpage>1431</fpage>
<lpage>1436</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Lewkowicz4">
<label>46</label>
<mixed-citation publication-type="journal">
<name>
<surname>Lewkowicz</surname>
<given-names>D</given-names>
</name>
(
<year>2000</year>
)
<article-title>The development of intersensory temporal perception: An epigenetic systems/limitations view</article-title>
.
<source>Pschological Bulletin</source>
<volume>126</volume>
:
<fpage>281</fpage>
<lpage>308</lpage>
.</mixed-citation>
</ref>
<ref id="pone.0111467-Desjardins1">
<label>47</label>
<mixed-citation publication-type="journal">
<name>
<surname>Desjardins</surname>
<given-names>RN</given-names>
</name>
,
<name>
<surname>Werker</surname>
<given-names>JF</given-names>
</name>
(
<year>2004</year>
)
<article-title>Is the integration of heard and seen speech mandatory for infants?</article-title>
<source>Developmental Psychobiology</source>
<volume>45</volume>
:
<fpage>187</fpage>
<lpage>203</lpage>
.
<pub-id pub-id-type="pmid">15549681</pub-id>
</mixed-citation>
</ref>
</ref-list>
</back>
</pmc>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Wicri/Asie/explor/AustralieFrV1/Data/Pmc/Corpus
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 0003688 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Pmc/Corpus/biblio.hfd -nk 0003688 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Wicri/Asie
   |area=    AustralieFrV1
   |flux=    Pmc
   |étape=   Corpus
   |type=    RBID
   |clé=     
   |texte=   
}}

Wicri

This area was generated with Dilib version V0.6.33.
Data generation: Tue Dec 5 10:43:12 2017. Site generation: Tue Mar 5 14:07:20 2024