Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits
Identifieur interne : 000C80 ( PascalFrancis/Corpus ); précédent : 000C79; suivant : 000C81Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits
Auteurs : Paul G. Griffiths ; R. Brent GillespieSource :
- Human factors [ 0018-7208 ] ; 2005.
Descripteurs français
- Pascal (Inist)
English descriptors
- KwdEn :
Abstract
This paper describes a paradigm for human/automation control sharing in which the automation acts through a motor coupled to a machine's manual control interface. The manual interface becomes a haptic display, continually informing the human about automation actions. While monitoring by feel, users may choose either to conform to the automation or override it and express their own control intentions. This paper's objective is to demonstrate that adding automation through haptic display can be used not only to improve performance on a primary task but also to reduce perceptual demands or free attention for a secondary task. Results are presented from three experiments in which 11 participants completed a lane-following task using a motorized steering wheel on a fixed-base driving simulator. The automation behaved like a copilot, assisting with lane following by applying torques to the steering wheel. Results indicate that haptic assist improves lane following by least 30%, p <.0001, while reducing visual demand by 29%, p <.0001, or improving reaction time in a secondary tone localization task by 18 ms, p = .0009. Potential applications of this research include the design of automation interfaces based on haptics that support human/automation control sharing better than traditional push-button automation interfaces.
Notice en format standard (ISO 2709)
Pour connaître la documentation sur le format Inist Standard.
pA |
|
---|
Format Inist (serveur)
NO : | PASCAL 06-0366216 INIST |
---|---|
ET : | Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits |
AU : | GRIFFITHS (Paul G.); GILLESPIE (R. Brent) |
AF : | University of Michigan/Ann Arbor, Michigan/Etats-Unis (1 aut., 2 aut.) |
DT : | Publication en série; Niveau analytique |
SO : | Human factors; ISSN 0018-7208; Coden HUFAA6; Etats-Unis; Da. 2005; Vol. 47; No. 3; Pp. 574-590; Bibl. 30 ref. |
LA : | Anglais |
EA : | This paper describes a paradigm for human/automation control sharing in which the automation acts through a motor coupled to a machine's manual control interface. The manual interface becomes a haptic display, continually informing the human about automation actions. While monitoring by feel, users may choose either to conform to the automation or override it and express their own control intentions. This paper's objective is to demonstrate that adding automation through haptic display can be used not only to improve performance on a primary task but also to reduce perceptual demands or free attention for a secondary task. Results are presented from three experiments in which 11 participants completed a lane-following task using a motorized steering wheel on a fixed-base driving simulator. The automation behaved like a copilot, assisting with lane following by applying torques to the steering wheel. Results indicate that haptic assist improves lane following by least 30%, p <.0001, while reducing visual demand by 29%, p <.0001, or improving reaction time in a secondary tone localization task by 18 ms, p = .0009. Potential applications of this research include the design of automation interfaces based on haptics that support human/automation control sharing better than traditional push-button automation interfaces. |
CC : | 002A26L07 |
FD : | Automatisation; Sensibilité tactile; Perception; Performance; Interface utilisateur; Ergonomie; Activité manuelle; Homme |
FG : | Motricité |
ED : | Automation; Tactile sensitivity; Perception; Performance; User interface; Ergonomics; Manual activity; Human |
EG : | Motricity |
SD : | Automatización; Sensibilidad tactil; Percepción; Rendimiento; Interfase usuario; Ergonomía; Actividad manual; Hombre |
LO : | INIST-2452.354000135077450100 |
ID : | 06-0366216 |
Links to Exploration step
Pascal:06-0366216Le document en format XML
<record><TEI><teiHeader><fileDesc><titleStmt><title xml:lang="en" level="a">Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits</title>
<author><name sortKey="Griffiths, Paul G" sort="Griffiths, Paul G" uniqKey="Griffiths P" first="Paul G." last="Griffiths">Paul G. Griffiths</name>
<affiliation><inist:fA14 i1="01"><s1>University of Michigan</s1>
<s2>Ann Arbor, Michigan</s2>
<s3>USA</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Gillespie, R Brent" sort="Gillespie, R Brent" uniqKey="Gillespie R" first="R. Brent" last="Gillespie">R. Brent Gillespie</name>
<affiliation><inist:fA14 i1="01"><s1>University of Michigan</s1>
<s2>Ann Arbor, Michigan</s2>
<s3>USA</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
</titleStmt>
<publicationStmt><idno type="wicri:source">INIST</idno>
<idno type="inist">06-0366216</idno>
<date when="2005">2005</date>
<idno type="stanalyst">PASCAL 06-0366216 INIST</idno>
<idno type="RBID">Pascal:06-0366216</idno>
<idno type="wicri:Area/PascalFrancis/Corpus">000C80</idno>
</publicationStmt>
<sourceDesc><biblStruct><analytic><title xml:lang="en" level="a">Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits</title>
<author><name sortKey="Griffiths, Paul G" sort="Griffiths, Paul G" uniqKey="Griffiths P" first="Paul G." last="Griffiths">Paul G. Griffiths</name>
<affiliation><inist:fA14 i1="01"><s1>University of Michigan</s1>
<s2>Ann Arbor, Michigan</s2>
<s3>USA</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Gillespie, R Brent" sort="Gillespie, R Brent" uniqKey="Gillespie R" first="R. Brent" last="Gillespie">R. Brent Gillespie</name>
<affiliation><inist:fA14 i1="01"><s1>University of Michigan</s1>
<s2>Ann Arbor, Michigan</s2>
<s3>USA</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
</analytic>
<series><title level="j" type="main">Human factors</title>
<title level="j" type="abbreviated">Hum. factors</title>
<idno type="ISSN">0018-7208</idno>
<imprint><date when="2005">2005</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
<seriesStmt><title level="j" type="main">Human factors</title>
<title level="j" type="abbreviated">Hum. factors</title>
<idno type="ISSN">0018-7208</idno>
</seriesStmt>
</fileDesc>
<profileDesc><textClass><keywords scheme="KwdEn" xml:lang="en"><term>Automation</term>
<term>Ergonomics</term>
<term>Human</term>
<term>Manual activity</term>
<term>Perception</term>
<term>Performance</term>
<term>Tactile sensitivity</term>
<term>User interface</term>
</keywords>
<keywords scheme="Pascal" xml:lang="fr"><term>Automatisation</term>
<term>Sensibilité tactile</term>
<term>Perception</term>
<term>Performance</term>
<term>Interface utilisateur</term>
<term>Ergonomie</term>
<term>Activité manuelle</term>
<term>Homme</term>
</keywords>
</textClass>
</profileDesc>
</teiHeader>
<front><div type="abstract" xml:lang="en">This paper describes a paradigm for human/automation control sharing in which the automation acts through a motor coupled to a machine's manual control interface. The manual interface becomes a haptic display, continually informing the human about automation actions. While monitoring by feel, users may choose either to conform to the automation or override it and express their own control intentions. This paper's objective is to demonstrate that adding automation through haptic display can be used not only to improve performance on a primary task but also to reduce perceptual demands or free attention for a secondary task. Results are presented from three experiments in which 11 participants completed a lane-following task using a motorized steering wheel on a fixed-base driving simulator. The automation behaved like a copilot, assisting with lane following by applying torques to the steering wheel. Results indicate that haptic assist improves lane following by least 30%, p <.0001, while reducing visual demand by 29%, p <.0001, or improving reaction time in a secondary tone localization task by 18 ms, p = .0009. Potential applications of this research include the design of automation interfaces based on haptics that support human/automation control sharing better than traditional push-button automation interfaces.</div>
</front>
</TEI>
<inist><standard h6="B"><pA><fA01 i1="01" i2="1"><s0>0018-7208</s0>
</fA01>
<fA02 i1="01"><s0>HUFAA6</s0>
</fA02>
<fA03 i2="1"><s0>Hum. factors</s0>
</fA03>
<fA05><s2>47</s2>
</fA05>
<fA06><s2>3</s2>
</fA06>
<fA08 i1="01" i2="1" l="ENG"><s1>Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits</s1>
</fA08>
<fA11 i1="01" i2="1"><s1>GRIFFITHS (Paul G.)</s1>
</fA11>
<fA11 i1="02" i2="1"><s1>GILLESPIE (R. Brent)</s1>
</fA11>
<fA14 i1="01"><s1>University of Michigan</s1>
<s2>Ann Arbor, Michigan</s2>
<s3>USA</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
</fA14>
<fA20><s1>574-590</s1>
</fA20>
<fA21><s1>2005</s1>
</fA21>
<fA23 i1="01"><s0>ENG</s0>
</fA23>
<fA43 i1="01"><s1>INIST</s1>
<s2>2452</s2>
<s5>354000135077450100</s5>
</fA43>
<fA44><s0>0000</s0>
<s1>© 2006 INIST-CNRS. All rights reserved.</s1>
</fA44>
<fA45><s0>30 ref.</s0>
</fA45>
<fA47 i1="01" i2="1"><s0>06-0366216</s0>
</fA47>
<fA60><s1>P</s1>
</fA60>
<fA61><s0>A</s0>
</fA61>
<fA64 i1="01" i2="1"><s0>Human factors</s0>
</fA64>
<fA66 i1="01"><s0>USA</s0>
</fA66>
<fC01 i1="01" l="ENG"><s0>This paper describes a paradigm for human/automation control sharing in which the automation acts through a motor coupled to a machine's manual control interface. The manual interface becomes a haptic display, continually informing the human about automation actions. While monitoring by feel, users may choose either to conform to the automation or override it and express their own control intentions. This paper's objective is to demonstrate that adding automation through haptic display can be used not only to improve performance on a primary task but also to reduce perceptual demands or free attention for a secondary task. Results are presented from three experiments in which 11 participants completed a lane-following task using a motorized steering wheel on a fixed-base driving simulator. The automation behaved like a copilot, assisting with lane following by applying torques to the steering wheel. Results indicate that haptic assist improves lane following by least 30%, p <.0001, while reducing visual demand by 29%, p <.0001, or improving reaction time in a secondary tone localization task by 18 ms, p = .0009. Potential applications of this research include the design of automation interfaces based on haptics that support human/automation control sharing better than traditional push-button automation interfaces.</s0>
</fC01>
<fC02 i1="01" i2="X"><s0>002A26L07</s0>
</fC02>
<fC03 i1="01" i2="X" l="FRE"><s0>Automatisation</s0>
<s5>01</s5>
</fC03>
<fC03 i1="01" i2="X" l="ENG"><s0>Automation</s0>
<s5>01</s5>
</fC03>
<fC03 i1="01" i2="X" l="SPA"><s0>Automatización</s0>
<s5>01</s5>
</fC03>
<fC03 i1="02" i2="X" l="FRE"><s0>Sensibilité tactile</s0>
<s5>02</s5>
</fC03>
<fC03 i1="02" i2="X" l="ENG"><s0>Tactile sensitivity</s0>
<s5>02</s5>
</fC03>
<fC03 i1="02" i2="X" l="SPA"><s0>Sensibilidad tactil</s0>
<s5>02</s5>
</fC03>
<fC03 i1="03" i2="X" l="FRE"><s0>Perception</s0>
<s5>03</s5>
</fC03>
<fC03 i1="03" i2="X" l="ENG"><s0>Perception</s0>
<s5>03</s5>
</fC03>
<fC03 i1="03" i2="X" l="SPA"><s0>Percepción</s0>
<s5>03</s5>
</fC03>
<fC03 i1="04" i2="X" l="FRE"><s0>Performance</s0>
<s5>04</s5>
</fC03>
<fC03 i1="04" i2="X" l="ENG"><s0>Performance</s0>
<s5>04</s5>
</fC03>
<fC03 i1="04" i2="X" l="SPA"><s0>Rendimiento</s0>
<s5>04</s5>
</fC03>
<fC03 i1="05" i2="X" l="FRE"><s0>Interface utilisateur</s0>
<s5>05</s5>
</fC03>
<fC03 i1="05" i2="X" l="ENG"><s0>User interface</s0>
<s5>05</s5>
</fC03>
<fC03 i1="05" i2="X" l="SPA"><s0>Interfase usuario</s0>
<s5>05</s5>
</fC03>
<fC03 i1="06" i2="X" l="FRE"><s0>Ergonomie</s0>
<s5>06</s5>
</fC03>
<fC03 i1="06" i2="X" l="ENG"><s0>Ergonomics</s0>
<s5>06</s5>
</fC03>
<fC03 i1="06" i2="X" l="SPA"><s0>Ergonomía</s0>
<s5>06</s5>
</fC03>
<fC03 i1="07" i2="X" l="FRE"><s0>Activité manuelle</s0>
<s5>07</s5>
</fC03>
<fC03 i1="07" i2="X" l="ENG"><s0>Manual activity</s0>
<s5>07</s5>
</fC03>
<fC03 i1="07" i2="X" l="SPA"><s0>Actividad manual</s0>
<s5>07</s5>
</fC03>
<fC03 i1="08" i2="X" l="FRE"><s0>Homme</s0>
<s5>18</s5>
</fC03>
<fC03 i1="08" i2="X" l="ENG"><s0>Human</s0>
<s5>18</s5>
</fC03>
<fC03 i1="08" i2="X" l="SPA"><s0>Hombre</s0>
<s5>18</s5>
</fC03>
<fC07 i1="01" i2="X" l="FRE"><s0>Motricité</s0>
<s5>37</s5>
</fC07>
<fC07 i1="01" i2="X" l="ENG"><s0>Motricity</s0>
<s5>37</s5>
</fC07>
<fC07 i1="01" i2="X" l="SPA"><s0>Motricidad</s0>
<s5>37</s5>
</fC07>
<fN21><s1>240</s1>
</fN21>
</pA>
</standard>
<server><NO>PASCAL 06-0366216 INIST</NO>
<ET>Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits</ET>
<AU>GRIFFITHS (Paul G.); GILLESPIE (R. Brent)</AU>
<AF>University of Michigan/Ann Arbor, Michigan/Etats-Unis (1 aut., 2 aut.)</AF>
<DT>Publication en série; Niveau analytique</DT>
<SO>Human factors; ISSN 0018-7208; Coden HUFAA6; Etats-Unis; Da. 2005; Vol. 47; No. 3; Pp. 574-590; Bibl. 30 ref.</SO>
<LA>Anglais</LA>
<EA>This paper describes a paradigm for human/automation control sharing in which the automation acts through a motor coupled to a machine's manual control interface. The manual interface becomes a haptic display, continually informing the human about automation actions. While monitoring by feel, users may choose either to conform to the automation or override it and express their own control intentions. This paper's objective is to demonstrate that adding automation through haptic display can be used not only to improve performance on a primary task but also to reduce perceptual demands or free attention for a secondary task. Results are presented from three experiments in which 11 participants completed a lane-following task using a motorized steering wheel on a fixed-base driving simulator. The automation behaved like a copilot, assisting with lane following by applying torques to the steering wheel. Results indicate that haptic assist improves lane following by least 30%, p <.0001, while reducing visual demand by 29%, p <.0001, or improving reaction time in a secondary tone localization task by 18 ms, p = .0009. Potential applications of this research include the design of automation interfaces based on haptics that support human/automation control sharing better than traditional push-button automation interfaces.</EA>
<CC>002A26L07</CC>
<FD>Automatisation; Sensibilité tactile; Perception; Performance; Interface utilisateur; Ergonomie; Activité manuelle; Homme</FD>
<FG>Motricité</FG>
<ED>Automation; Tactile sensitivity; Perception; Performance; User interface; Ergonomics; Manual activity; Human</ED>
<EG>Motricity</EG>
<SD>Automatización; Sensibilidad tactil; Percepción; Rendimiento; Interfase usuario; Ergonomía; Actividad manual; Hombre</SD>
<LO>INIST-2452.354000135077450100</LO>
<ID>06-0366216</ID>
</server>
</inist>
</record>
Pour manipuler ce document sous Unix (Dilib)
EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/HapticV1/Data/PascalFrancis/Corpus
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000C80 | SxmlIndent | more
Ou
HfdSelect -h $EXPLOR_AREA/Data/PascalFrancis/Corpus/biblio.hfd -nk 000C80 | SxmlIndent | more
Pour mettre un lien sur cette page dans le réseau Wicri
{{Explor lien |wiki= Ticri/CIDE |area= HapticV1 |flux= PascalFrancis |étape= Corpus |type= RBID |clé= Pascal:06-0366216 |texte= Sharing control between humans and automation using haptic interface : Primary and secondary task performance benefits }}
![]() | This area was generated with Dilib version V0.6.23. | ![]() |