Serveur d'exploration sur l'opéra

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Automatic Rush Generation with Application to Theatre Performances

Identifieur interne : 000032 ( Hal/Curation ); précédent : 000031; suivant : 000033

Automatic Rush Generation with Application to Theatre Performances

Auteurs : Vineet Gandhi [France]

Source :

RBID : Hal:tel-01119207

Descripteurs français

English descriptors

Abstract

Professional quality videos of live staged performances are created by recording them fromdifferent appropriate viewpoints. These are then edited together to portray an eloquent storyreplete with the ability to draw out the intended emotion from the viewers. Creating such competentvideos typically requires a team of skilled camera operators to capture the scene frommultiple viewpoints. In this thesis, we explore an alternative approach where we automaticallycompute camera movements in post-production using specially designed computer visionmethods.A high resolution static camera replaces the plural camera crew and their efficient cameramovements are then simulated by virtually panning - tilting - zooming within the originalrecordings. We show that multiple virtual cameras can be simulated by choosing different trajectoriesof cropping windows inside the original recording. One of the key novelties of thiswork is an optimization framework for computing the virtual camera trajectories using the informationextracted from the original video based on computer vision techniques.The actors present on stage are considered as the most important elements of the scene.For the task of localizing and naming actors, we introduce generative models for learning viewindependent person and costume specific detectors from a set of labeled examples. We explainhow to learn the models from a small number of labeled keyframes or video tracks, and how todetect novel appearances of the actors in a maximum likelihood framework. We demonstratethat such actor specific models can accurately localize actors despite changes in view point andocclusions, and significantly improve the detection recall rates over generic detectors.The thesis then proposes an offline algorithm for tracking objects and actors in long videosequences using these actor specific models. Detections are first performed to independentlyselect candidate locations of the actor/object in each frame of the video. The candidate detectionsare then combined into smooth trajectories by minimizing a cost function accounting forfalse detections and occlusions.Using the actor tracks, we then describe a method for automatically generating multipleclips suitable for video editing by simulating pan-tilt-zoom camera movements within theframe of a single static camera. Our method requires only minimal user input to define thesubject matter of each sub-clip. The composition of each sub-clip is automatically computedin a novel convex optimization framework. Our approach encodes several common cinematographicpractices into a single convex cost function minimization problem, resulting inaesthetically-pleasing sub-clips which can easily be edited together using off-the-shelf multiclipvideo editing software.The proposed methods have been tested and validated on a challenging corpus of theatrerecordings. They open the way to novel applications of computer vision methods for costeffectivevideo production of live performances including, but not restricted to, theatre, musicand opera.

Url:

Links toward previous steps (curation, corpus...)


Links to Exploration step

Hal:tel-01119207

Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Automatic Rush Generation with Application to Theatre Performances</title>
<title xml:lang="fr">Généation Automatique de Prises de Vues Cinématographiques avec Applications aux Captations de Théâtre</title>
<author>
<name sortKey="Gandhi, Vineet" sort="Gandhi, Vineet" uniqKey="Gandhi V" first="Vineet" last="Gandhi">Vineet Gandhi</name>
<affiliation wicri:level="1">
<hal:affiliation type="researchteam" xml:id="struct-174814" status="VALID">
<orgName>Intuitive Modeling and Animation for Interactive Graphics & Narrative Environments</orgName>
<orgName type="acronym">IMAGINE</orgName>
<desc>
<address>
<addrLine>655 avenue de l'Europe 38 334 Saint Ismier cedex</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/equipes/imagine</ref>
</desc>
<listRelation>
<relation active="#struct-24474" type="direct"></relation>
<relation active="#struct-3886" type="indirect"></relation>
<relation active="#struct-51016" type="indirect"></relation>
<relation active="#struct-300339" type="indirect"></relation>
<relation name="UMR5224" active="#struct-441569" type="indirect"></relation>
<relation active="#struct-300275" type="direct"></relation>
<relation active="#struct-2497" type="direct"></relation>
<relation active="#struct-300009" type="indirect"></relation>
</listRelation>
<tutelles>
<tutelle active="#struct-24474" type="direct">
<org type="laboratory" xml:id="struct-24474" status="VALID">
<orgName>Laboratoire Jean Kuntzmann</orgName>
<orgName type="acronym">LJK</orgName>
<desc>
<address>
<addrLine>Tour IRMA 51 rue des Mathématiques - 53 38041 GRENOBLE CEDEX 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://ljk.imag.fr</ref>
</desc>
<listRelation>
<relation active="#struct-3886" type="direct"></relation>
<relation active="#struct-51016" type="direct"></relation>
<relation active="#struct-300339" type="direct"></relation>
<relation name="UMR5224" active="#struct-441569" type="direct"></relation>
</listRelation>
</org>
</tutelle>
<tutelle active="#struct-3886" type="indirect">
<org type="institution" xml:id="struct-3886" status="OLD">
<orgName>Université Pierre Mendès France</orgName>
<orgName type="acronym">Grenoble 2 UPMF</orgName>
<date type="end">2015-12-31</date>
<desc>
<address>
<addrLine>BP 47 - 38040 Grenoble Cedex 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.upmf-grenoble.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-51016" type="indirect">
<org type="institution" xml:id="struct-51016" status="OLD">
<orgName>Université Joseph Fourier</orgName>
<orgName type="acronym">UJF</orgName>
<date type="end">2015-12-31</date>
<desc>
<address>
<addrLine>BP 53 - 38041 Grenoble Cedex 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.ujf-grenoble.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-300339" type="indirect">
<org type="institution" xml:id="struct-300339" status="VALID">
<orgName>Institut Polytechnique de Grenoble - Grenoble Institute of Technology</orgName>
<desc>
<address>
<country key="FR"></country>
</address>
</desc>
</org>
</tutelle>
<tutelle name="UMR5224" active="#struct-441569" type="indirect">
<org type="institution" xml:id="struct-441569" status="VALID">
<idno type="IdRef">02636817X</idno>
<idno type="ISNI">0000000122597504</idno>
<orgName>Centre National de la Recherche Scientifique</orgName>
<orgName type="acronym">CNRS</orgName>
<date type="start">1939-10-19</date>
<desc>
<address>
<country key="FR"></country>
</address>
<ref type="url">http://www.cnrs.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-300275" type="direct">
<org type="institution" xml:id="struct-300275" status="VALID">
<orgName>Institut National Polytechnique de Grenoble (INPG)</orgName>
<desc>
<address>
<country key="FR"></country>
</address>
</desc>
</org>
</tutelle>
<tutelle active="#struct-2497" type="direct">
<org type="laboratory" xml:id="struct-2497" status="VALID">
<orgName>Inria Grenoble - Rhône-Alpes</orgName>
<desc>
<address>
<addrLine>Inovallée655 avenue de l'Europe38330 Montbonnot</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/centre/grenoble</ref>
</desc>
<listRelation>
<relation active="#struct-300009" type="direct"></relation>
</listRelation>
</org>
</tutelle>
<tutelle active="#struct-300009" type="indirect">
<org type="institution" xml:id="struct-300009" status="VALID">
<orgName>Institut National de Recherche en Informatique et en Automatique</orgName>
<orgName type="acronym">Inria</orgName>
<desc>
<address>
<addrLine>Domaine de VoluceauRocquencourt - BP 10578153 Le Chesnay Cedex</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/en/</ref>
</desc>
</org>
</tutelle>
</tutelles>
</hal:affiliation>
<country>France</country>
<placeName>
<settlement type="city">Grenoble</settlement>
<region type="region" nuts="2">Rhône-Alpes</region>
</placeName>
<orgName type="university">Université Pierre-Mendès-France</orgName>
<orgName type="institution" wicri:auto="newGroup">Université de Grenoble</orgName>
<placeName>
<settlement type="city">Grenoble</settlement>
<region type="region" nuts="2">Rhône-Alpes</region>
</placeName>
<orgName type="university">Université Joseph Fourier</orgName>
<orgName type="institution" wicri:auto="newGroup">Université de Grenoble</orgName>
</affiliation>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">HAL</idno>
<idno type="RBID">Hal:tel-01119207</idno>
<idno type="halId">tel-01119207</idno>
<idno type="halUri">https://tel.archives-ouvertes.fr/tel-01119207</idno>
<idno type="url">https://tel.archives-ouvertes.fr/tel-01119207</idno>
<date when="2014-12-18">2014-12-18</date>
<idno type="wicri:Area/Hal/Corpus">000032</idno>
<idno type="wicri:Area/Hal/Curation">000032</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en">Automatic Rush Generation with Application to Theatre Performances</title>
<title xml:lang="fr">Généation Automatique de Prises de Vues Cinématographiques avec Applications aux Captations de Théâtre</title>
<author>
<name sortKey="Gandhi, Vineet" sort="Gandhi, Vineet" uniqKey="Gandhi V" first="Vineet" last="Gandhi">Vineet Gandhi</name>
<affiliation wicri:level="1">
<hal:affiliation type="researchteam" xml:id="struct-174814" status="VALID">
<orgName>Intuitive Modeling and Animation for Interactive Graphics & Narrative Environments</orgName>
<orgName type="acronym">IMAGINE</orgName>
<desc>
<address>
<addrLine>655 avenue de l'Europe 38 334 Saint Ismier cedex</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/equipes/imagine</ref>
</desc>
<listRelation>
<relation active="#struct-24474" type="direct"></relation>
<relation active="#struct-3886" type="indirect"></relation>
<relation active="#struct-51016" type="indirect"></relation>
<relation active="#struct-300339" type="indirect"></relation>
<relation name="UMR5224" active="#struct-441569" type="indirect"></relation>
<relation active="#struct-300275" type="direct"></relation>
<relation active="#struct-2497" type="direct"></relation>
<relation active="#struct-300009" type="indirect"></relation>
</listRelation>
<tutelles>
<tutelle active="#struct-24474" type="direct">
<org type="laboratory" xml:id="struct-24474" status="VALID">
<orgName>Laboratoire Jean Kuntzmann</orgName>
<orgName type="acronym">LJK</orgName>
<desc>
<address>
<addrLine>Tour IRMA 51 rue des Mathématiques - 53 38041 GRENOBLE CEDEX 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://ljk.imag.fr</ref>
</desc>
<listRelation>
<relation active="#struct-3886" type="direct"></relation>
<relation active="#struct-51016" type="direct"></relation>
<relation active="#struct-300339" type="direct"></relation>
<relation name="UMR5224" active="#struct-441569" type="direct"></relation>
</listRelation>
</org>
</tutelle>
<tutelle active="#struct-3886" type="indirect">
<org type="institution" xml:id="struct-3886" status="OLD">
<orgName>Université Pierre Mendès France</orgName>
<orgName type="acronym">Grenoble 2 UPMF</orgName>
<date type="end">2015-12-31</date>
<desc>
<address>
<addrLine>BP 47 - 38040 Grenoble Cedex 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.upmf-grenoble.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-51016" type="indirect">
<org type="institution" xml:id="struct-51016" status="OLD">
<orgName>Université Joseph Fourier</orgName>
<orgName type="acronym">UJF</orgName>
<date type="end">2015-12-31</date>
<desc>
<address>
<addrLine>BP 53 - 38041 Grenoble Cedex 9</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.ujf-grenoble.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-300339" type="indirect">
<org type="institution" xml:id="struct-300339" status="VALID">
<orgName>Institut Polytechnique de Grenoble - Grenoble Institute of Technology</orgName>
<desc>
<address>
<country key="FR"></country>
</address>
</desc>
</org>
</tutelle>
<tutelle name="UMR5224" active="#struct-441569" type="indirect">
<org type="institution" xml:id="struct-441569" status="VALID">
<idno type="IdRef">02636817X</idno>
<idno type="ISNI">0000000122597504</idno>
<orgName>Centre National de la Recherche Scientifique</orgName>
<orgName type="acronym">CNRS</orgName>
<date type="start">1939-10-19</date>
<desc>
<address>
<country key="FR"></country>
</address>
<ref type="url">http://www.cnrs.fr/</ref>
</desc>
</org>
</tutelle>
<tutelle active="#struct-300275" type="direct">
<org type="institution" xml:id="struct-300275" status="VALID">
<orgName>Institut National Polytechnique de Grenoble (INPG)</orgName>
<desc>
<address>
<country key="FR"></country>
</address>
</desc>
</org>
</tutelle>
<tutelle active="#struct-2497" type="direct">
<org type="laboratory" xml:id="struct-2497" status="VALID">
<orgName>Inria Grenoble - Rhône-Alpes</orgName>
<desc>
<address>
<addrLine>Inovallée655 avenue de l'Europe38330 Montbonnot</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/centre/grenoble</ref>
</desc>
<listRelation>
<relation active="#struct-300009" type="direct"></relation>
</listRelation>
</org>
</tutelle>
<tutelle active="#struct-300009" type="indirect">
<org type="institution" xml:id="struct-300009" status="VALID">
<orgName>Institut National de Recherche en Informatique et en Automatique</orgName>
<orgName type="acronym">Inria</orgName>
<desc>
<address>
<addrLine>Domaine de VoluceauRocquencourt - BP 10578153 Le Chesnay Cedex</addrLine>
<country key="FR"></country>
</address>
<ref type="url">http://www.inria.fr/en/</ref>
</desc>
</org>
</tutelle>
</tutelles>
</hal:affiliation>
<country>France</country>
<placeName>
<settlement type="city">Grenoble</settlement>
<region type="region" nuts="2">Rhône-Alpes</region>
</placeName>
<orgName type="university">Université Pierre-Mendès-France</orgName>
<orgName type="institution" wicri:auto="newGroup">Université de Grenoble</orgName>
<placeName>
<settlement type="city">Grenoble</settlement>
<region type="region" nuts="2">Rhône-Alpes</region>
</placeName>
<orgName type="university">Université Joseph Fourier</orgName>
<orgName type="institution" wicri:auto="newGroup">Université de Grenoble</orgName>
</affiliation>
</author>
</analytic>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass>
<keywords scheme="mix" xml:lang="en">
<term>Actor detection</term>
<term>Actor recognition</term>
<term>And/Or graph</term>
<term>Film Editing</term>
<term>Virtual cinematography</term>
</keywords>
<keywords scheme="mix" xml:lang="fr">
<term>Cinématographie virtuelle</term>
<term>Détection d'acteurs</term>
<term>Montage vidéo</term>
<term>Reconnaissance d'acteurs</term>
</keywords>
</textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">Professional quality videos of live staged performances are created by recording them fromdifferent appropriate viewpoints. These are then edited together to portray an eloquent storyreplete with the ability to draw out the intended emotion from the viewers. Creating such competentvideos typically requires a team of skilled camera operators to capture the scene frommultiple viewpoints. In this thesis, we explore an alternative approach where we automaticallycompute camera movements in post-production using specially designed computer visionmethods.A high resolution static camera replaces the plural camera crew and their efficient cameramovements are then simulated by virtually panning - tilting - zooming within the originalrecordings. We show that multiple virtual cameras can be simulated by choosing different trajectoriesof cropping windows inside the original recording. One of the key novelties of thiswork is an optimization framework for computing the virtual camera trajectories using the informationextracted from the original video based on computer vision techniques.The actors present on stage are considered as the most important elements of the scene.For the task of localizing and naming actors, we introduce generative models for learning viewindependent person and costume specific detectors from a set of labeled examples. We explainhow to learn the models from a small number of labeled keyframes or video tracks, and how todetect novel appearances of the actors in a maximum likelihood framework. We demonstratethat such actor specific models can accurately localize actors despite changes in view point andocclusions, and significantly improve the detection recall rates over generic detectors.The thesis then proposes an offline algorithm for tracking objects and actors in long videosequences using these actor specific models. Detections are first performed to independentlyselect candidate locations of the actor/object in each frame of the video. The candidate detectionsare then combined into smooth trajectories by minimizing a cost function accounting forfalse detections and occlusions.Using the actor tracks, we then describe a method for automatically generating multipleclips suitable for video editing by simulating pan-tilt-zoom camera movements within theframe of a single static camera. Our method requires only minimal user input to define thesubject matter of each sub-clip. The composition of each sub-clip is automatically computedin a novel convex optimization framework. Our approach encodes several common cinematographicpractices into a single convex cost function minimization problem, resulting inaesthetically-pleasing sub-clips which can easily be edited together using off-the-shelf multiclipvideo editing software.The proposed methods have been tested and validated on a challenging corpus of theatrerecordings. They open the way to novel applications of computer vision methods for costeffectivevideo production of live performances including, but not restricted to, theatre, musicand opera.</div>
</front>
</TEI>
<hal api="V3">
<titleStmt>
<title xml:lang="en">Automatic Rush Generation with Application to Theatre Performances</title>
<title xml:lang="fr">Généation Automatique de Prises de Vues Cinématographiques avec Applications aux Captations de Théâtre</title>
<author role="aut">
<persName>
<forename type="first">Vineet</forename>
<surname>Gandhi</surname>
</persName>
<email>Vineet.Gandhi@inria.fr</email>
<idno type="halauthor">1131170</idno>
<affiliation ref="#struct-174814"></affiliation>
</author>
<editor role="depositor">
<persName>
<forename>Remi</forename>
<surname>Ronfard</surname>
</persName>
<email>remi.ronfard@inria.fr</email>
</editor>
<funder ref="#projanr-35582"></funder>
<funder>Région Rhone-Alpes, projet CIBLE SCENOPTIQUE</funder>
</titleStmt>
<editionStmt>
<edition n="v1" type="current">
<date type="whenSubmitted">2015-02-22 09:12:21</date>
<date type="whenModified">2015-03-01 22:13:56</date>
<date type="whenReleased">2015-02-23 10:51:34</date>
<date type="whenProduced">2014-12-18</date>
<date type="whenEndEmbargoed">2015-02-22</date>
<ref type="file" target="https://tel.archives-ouvertes.fr/tel-01119207/document">
<date notBefore="2015-02-22"></date>
</ref>
<ref type="file" n="1" target="https://tel.archives-ouvertes.fr/tel-01119207/file/Phd_Thesis_Vineet_Gandhi.pdf">
<date notBefore="2015-02-22"></date>
</ref>
</edition>
<respStmt>
<resp>contributor</resp>
<name key="147607">
<persName>
<forename>Remi</forename>
<surname>Ronfard</surname>
</persName>
<email>remi.ronfard@inria.fr</email>
</name>
</respStmt>
</editionStmt>
<publicationStmt>
<distributor>CCSD</distributor>
<idno type="halId">tel-01119207</idno>
<idno type="halUri">https://tel.archives-ouvertes.fr/tel-01119207</idno>
<idno type="halBibtex">gandhi:tel-01119207</idno>
<idno type="halRefHtml">Computer Vision and Pattern Recognition [cs.CV]. Grenoble University, 2014. English</idno>
<idno type="halRef">Computer Vision and Pattern Recognition [cs.CV]. Grenoble University, 2014. English</idno>
</publicationStmt>
<seriesStmt>
<idno type="stamp" n="CNRS">CNRS - Centre national de la recherche scientifique</idno>
<idno type="stamp" n="INRIA">INRIA - Institut National de Recherche en Informatique et en Automatique</idno>
<idno type="stamp" n="UNIV-GRENOBLE1" p="UGA">Université Joseph Fourier - Grenoble I</idno>
<idno type="stamp" n="INPG" p="UGA">Institut polytechnique de Grenoble</idno>
<idno type="stamp" n="INRIA-RHA">INRIA Grenoble - Rhône-Alpes</idno>
<idno type="stamp" n="LJK" p="UGA">Laboratoire Jean Kuntzmann</idno>
<idno type="stamp" n="UGA">HAL Grenoble Alpes</idno>
<idno type="stamp" n="LJK_GI" p="LJK">Département Géométrie et Images</idno>
<idno type="stamp" n="LJK_GI_IMAGINE" p="LJK_GI">IMAGINE</idno>
</seriesStmt>
<notesStmt></notesStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en">Automatic Rush Generation with Application to Theatre Performances</title>
<title xml:lang="fr">Généation Automatique de Prises de Vues Cinématographiques avec Applications aux Captations de Théâtre</title>
<author role="aut">
<persName>
<forename type="first">Vineet</forename>
<surname>Gandhi</surname>
</persName>
<email>Vineet.Gandhi@inria.fr</email>
<idno type="halAuthorId">1131170</idno>
<affiliation ref="#struct-174814"></affiliation>
</author>
</analytic>
<monogr>
<imprint>
<date type="dateDefended">2014-12-18</date>
</imprint>
<authority type="institution">Grenoble University</authority>
<authority type="school">EDMSTII</authority>
<authority type="supervisor">Rémi Ronfard</authority>
<authority type="jury">James Crowley</authority>
<authority type="jury">Patrick Perez</authority>
<authority type="jury">Frédéric Jurie</authority>
<authority type="jury">Alexander Sorkine-Hornung</authority>
<authority type="jury">Michael Gleicher</authority>
<authority type="jury">Remi Ronfard</authority>
</monogr>
</biblStruct>
</sourceDesc>
<profileDesc>
<langUsage>
<language ident="en">English</language>
</langUsage>
<textClass>
<keywords scheme="author">
<term xml:lang="en">And/Or graph</term>
<term xml:lang="en">Film Editing</term>
<term xml:lang="en">Actor recognition</term>
<term xml:lang="en">Actor detection</term>
<term xml:lang="en">Virtual cinematography</term>
<term xml:lang="fr">Reconnaissance d'acteurs</term>
<term xml:lang="fr">Montage vidéo</term>
<term xml:lang="fr">Détection d'acteurs</term>
<term xml:lang="fr">Cinématographie virtuelle</term>
</keywords>
<classCode scheme="halDomain" n="info.info-cv">Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]</classCode>
<classCode scheme="halTypology" n="THESE">Theses</classCode>
</textClass>
<abstract xml:lang="en">Professional quality videos of live staged performances are created by recording them fromdifferent appropriate viewpoints. These are then edited together to portray an eloquent storyreplete with the ability to draw out the intended emotion from the viewers. Creating such competentvideos typically requires a team of skilled camera operators to capture the scene frommultiple viewpoints. In this thesis, we explore an alternative approach where we automaticallycompute camera movements in post-production using specially designed computer visionmethods.A high resolution static camera replaces the plural camera crew and their efficient cameramovements are then simulated by virtually panning - tilting - zooming within the originalrecordings. We show that multiple virtual cameras can be simulated by choosing different trajectoriesof cropping windows inside the original recording. One of the key novelties of thiswork is an optimization framework for computing the virtual camera trajectories using the informationextracted from the original video based on computer vision techniques.The actors present on stage are considered as the most important elements of the scene.For the task of localizing and naming actors, we introduce generative models for learning viewindependent person and costume specific detectors from a set of labeled examples. We explainhow to learn the models from a small number of labeled keyframes or video tracks, and how todetect novel appearances of the actors in a maximum likelihood framework. We demonstratethat such actor specific models can accurately localize actors despite changes in view point andocclusions, and significantly improve the detection recall rates over generic detectors.The thesis then proposes an offline algorithm for tracking objects and actors in long videosequences using these actor specific models. Detections are first performed to independentlyselect candidate locations of the actor/object in each frame of the video. The candidate detectionsare then combined into smooth trajectories by minimizing a cost function accounting forfalse detections and occlusions.Using the actor tracks, we then describe a method for automatically generating multipleclips suitable for video editing by simulating pan-tilt-zoom camera movements within theframe of a single static camera. Our method requires only minimal user input to define thesubject matter of each sub-clip. The composition of each sub-clip is automatically computedin a novel convex optimization framework. Our approach encodes several common cinematographicpractices into a single convex cost function minimization problem, resulting inaesthetically-pleasing sub-clips which can easily be edited together using off-the-shelf multiclipvideo editing software.The proposed methods have been tested and validated on a challenging corpus of theatrerecordings. They open the way to novel applications of computer vision methods for costeffectivevideo production of live performances including, but not restricted to, theatre, musicand opera.</abstract>
<abstract xml:lang="fr">Les captations professionnelles de pièces de théâtre utilisent un grand nombre de caméras afinde montrer l’ensemble du spectacle sous tous ses angles. C’est un processus complexe et coûteux,qui fait appel aux compétences d’un grand nombre de techniciens qualifiés pour assurer lecadrage puis le montage de toutes les prises de vues. Dans cette thèse, nous explorons une approchedifférente, consistant à calculer automatiquement en post-production des cadrages dynamiquesà partir d’un petit nombre de prises de vues obtenues en caméra fixe, sans opérateurs.Pour atteindre cet objectif, nous proposons de nouveaux algorithmes de vision par ordinateurqui nous permettent de formaliser et reproduire les régles du cadrage cinématographique.Dans cette thèse, nous proposons de nouvelles méthodes d’analyse vidéo pour calculer automatiquementle cadrage le plus approprié aux mouvements des acteurs qui évoluent sur scène.Nous simulons pour cela les mouvements d’une caméra "pan-tilt-zoom" extraite du cadre d’uneprise de vue en caméra fixe. Une contribution importante de la thèse consiste à formaliser leproblème du cadrage cinématographique comme un problème d’optimisation convexe.Dans une première partie de la thèse, nous proposons des méthodes nouvelles pour détecteret reconnaitre les acteurs à l’aide d’une modélisation explicite de leurs apparences, quiinclue leurs caractères physiques ainsi que leurs costumes et chevelures. Nous présentons uneapproche pour apprendre ces modèles d’apparence à partir d’un petit nombre d’exemples, enmaximisant leur vraisemblance. Nous montrons l’efficacité de ces méthodes sur des exemplesde films de théâtre et de cinéma.Dans une seconde partie de la thèse, nous montrons comment ces méthodes peuvent êtreutilisées pour effectuer le suivi des acteurs d’une pièce de théâtre, y compris sur de longuesséquences de plusieurs minutes, par l’utilisation de méthodes efficaces de programmation dynamique,qui permettent de prendre en compte les entrées et sorties de scène des acteurs, ainsique leurs occultations mutuelles.Dans une troisième partie de la thèse, nous décrivons une méthode générale pour calculerdynamiquement le cadrage d’une caméra virtuelle incluant un nombre quelconque d’acteurs,tout en excluant les autre acteurs dans la mesure du possible. Notre méthode prend en compteun grand nombre de considérations esthétiques que nous avons extraites des ouvrages techniquesconsacrés à la cinématographie et au montage. Notre approche présente l’avantagede formaliser les règles de la cinématographie et du montage sous la forme d’un problèmed’optimisation convexe, pour lequel nous pouvons proposer une solution efficace.Tout au long de la thèse, nous illustrons et validons les approches proposées sur des exemplesréels et complexes, que nous avons filmés au Théâtre de Lyon - Célestins. Les méthodesque nous proposons s’appliquent généralement au spectacle vivant (théâtre, musique, opéra) etpermettent d’envisager de nouvelles applications de la vision par ordinateur dans le domainede la production audio-visuelle.</abstract>
</profileDesc>
</hal>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Wicri/Musique/explor/OperaV1/Data/Hal/Curation
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000032 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Hal/Curation/biblio.hfd -nk 000032 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Wicri/Musique
   |area=    OperaV1
   |flux=    Hal
   |étape=   Curation
   |type=    RBID
   |clé=     Hal:tel-01119207
   |texte=   Automatic Rush Generation with Application to Theatre Performances
}}

Wicri

This area was generated with Dilib version V0.6.21.
Data generation: Thu Apr 14 14:59:05 2016. Site generation: Thu Jan 4 23:09:23 2024