Serveur d'exploration sur les dispositifs haptiques

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Motor imagery facilitates force field learning.

Identifieur interne : 002B17 ( Main/Merge ); précédent : 002B16; suivant : 002B18

Motor imagery facilitates force field learning.

Auteurs : Muhammad Nabeel Anwar [Japon] ; Naoki Tomi ; Koji Ito

Source :

RBID : pubmed:21555118

English descriptors

Abstract

Humans have the ability to produce an internal reproduction of a specific motor action without any overt motor output. Recent findings show that the processes underlying motor imagery are similar to those active during motor execution and both share common neural substrates. This suggests that the imagery of motor movements might play an important role in acquiring new motor skills. In this study we used haptic robot in conjunction with motor imagery technique to improve learning in a robot-based adaptation task. Two groups of subjects performed reaching movements with or without motor imagery in a velocity-dependent and position-dependent mixed force field. The groups performed movements with motor imagery produced higher after effects and decreased muscle co-contraction with respect to no-motor imagery group. These results showed a positive influence of motor imagery on acquiring new motor skill and suggest that motor learning can be facilitated by mental practice and could be used to increase the rate of adaptation.

DOI: 10.1016/j.brainres.2011.04.030
PubMed: 21555118

Links toward previous steps (curation, corpus...)


Links to Exploration step

pubmed:21555118

Le document en format XML

<record>
<TEI>
<teiHeader>
<fileDesc>
<titleStmt>
<title xml:lang="en">Motor imagery facilitates force field learning.</title>
<author>
<name sortKey="Anwar, Muhammad Nabeel" sort="Anwar, Muhammad Nabeel" uniqKey="Anwar M" first="Muhammad Nabeel" last="Anwar">Muhammad Nabeel Anwar</name>
<affiliation wicri:level="1">
<nlm:affiliation>Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, 4259 Nagatsuta, Midori-ku, Yokohama 226-8502, Japan. nabeel@smme.nust.edu.pk</nlm:affiliation>
<country xml:lang="fr">Japon</country>
<wicri:regionArea>Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, 4259 Nagatsuta, Midori-ku, Yokohama 226-8502</wicri:regionArea>
<wicri:noRegion>Yokohama 226-8502</wicri:noRegion>
</affiliation>
</author>
<author>
<name sortKey="Tomi, Naoki" sort="Tomi, Naoki" uniqKey="Tomi N" first="Naoki" last="Tomi">Naoki Tomi</name>
</author>
<author>
<name sortKey="Ito, Koji" sort="Ito, Koji" uniqKey="Ito K" first="Koji" last="Ito">Koji Ito</name>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">PubMed</idno>
<date when="2011">2011</date>
<idno type="doi">10.1016/j.brainres.2011.04.030</idno>
<idno type="RBID">pubmed:21555118</idno>
<idno type="pmid">21555118</idno>
<idno type="wicri:Area/PubMed/Corpus">000E71</idno>
<idno type="wicri:Area/PubMed/Curation">000E71</idno>
<idno type="wicri:Area/PubMed/Checkpoint">000D03</idno>
<idno type="wicri:Area/Ncbi/Merge">001A30</idno>
<idno type="wicri:Area/Ncbi/Curation">001A30</idno>
<idno type="wicri:Area/Ncbi/Checkpoint">001A30</idno>
<idno type="wicri:Area/Main/Merge">002B17</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title xml:lang="en">Motor imagery facilitates force field learning.</title>
<author>
<name sortKey="Anwar, Muhammad Nabeel" sort="Anwar, Muhammad Nabeel" uniqKey="Anwar M" first="Muhammad Nabeel" last="Anwar">Muhammad Nabeel Anwar</name>
<affiliation wicri:level="1">
<nlm:affiliation>Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, 4259 Nagatsuta, Midori-ku, Yokohama 226-8502, Japan. nabeel@smme.nust.edu.pk</nlm:affiliation>
<country xml:lang="fr">Japon</country>
<wicri:regionArea>Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, 4259 Nagatsuta, Midori-ku, Yokohama 226-8502</wicri:regionArea>
<wicri:noRegion>Yokohama 226-8502</wicri:noRegion>
</affiliation>
</author>
<author>
<name sortKey="Tomi, Naoki" sort="Tomi, Naoki" uniqKey="Tomi N" first="Naoki" last="Tomi">Naoki Tomi</name>
</author>
<author>
<name sortKey="Ito, Koji" sort="Ito, Koji" uniqKey="Ito K" first="Koji" last="Ito">Koji Ito</name>
</author>
</analytic>
<series>
<title level="j">Brain research</title>
<idno type="eISSN">1872-6240</idno>
<imprint>
<date when="2011" type="published">2011</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<textClass>
<keywords scheme="KwdEn" xml:lang="en">
<term>Adaptation, Physiological (physiology)</term>
<term>Adult</term>
<term>Biomechanical Phenomena (physiology)</term>
<term>Humans</term>
<term>Imagination (physiology)</term>
<term>Learning (physiology)</term>
<term>Male</term>
<term>Motor Skills (physiology)</term>
<term>Movement (physiology)</term>
<term>Muscle Contraction (physiology)</term>
<term>Muscle, Skeletal (physiology)</term>
<term>Physical Therapy Modalities (instrumentation)</term>
<term>Physical Therapy Modalities (trends)</term>
<term>Robotics (instrumentation)</term>
<term>Robotics (methods)</term>
<term>Young Adult</term>
</keywords>
<keywords scheme="MESH" qualifier="instrumentation" xml:lang="en">
<term>Physical Therapy Modalities</term>
<term>Robotics</term>
</keywords>
<keywords scheme="MESH" qualifier="methods" xml:lang="en">
<term>Robotics</term>
</keywords>
<keywords scheme="MESH" qualifier="physiology" xml:lang="en">
<term>Adaptation, Physiological</term>
<term>Biomechanical Phenomena</term>
<term>Imagination</term>
<term>Learning</term>
<term>Motor Skills</term>
<term>Movement</term>
<term>Muscle Contraction</term>
<term>Muscle, Skeletal</term>
</keywords>
<keywords scheme="MESH" qualifier="trends" xml:lang="en">
<term>Physical Therapy Modalities</term>
</keywords>
<keywords scheme="MESH" xml:lang="en">
<term>Adult</term>
<term>Humans</term>
<term>Male</term>
<term>Young Adult</term>
</keywords>
</textClass>
</profileDesc>
</teiHeader>
<front>
<div type="abstract" xml:lang="en">Humans have the ability to produce an internal reproduction of a specific motor action without any overt motor output. Recent findings show that the processes underlying motor imagery are similar to those active during motor execution and both share common neural substrates. This suggests that the imagery of motor movements might play an important role in acquiring new motor skills. In this study we used haptic robot in conjunction with motor imagery technique to improve learning in a robot-based adaptation task. Two groups of subjects performed reaching movements with or without motor imagery in a velocity-dependent and position-dependent mixed force field. The groups performed movements with motor imagery produced higher after effects and decreased muscle co-contraction with respect to no-motor imagery group. These results showed a positive influence of motor imagery on acquiring new motor skill and suggest that motor learning can be facilitated by mental practice and could be used to increase the rate of adaptation.</div>
</front>
</TEI>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/HapticV1/Data/Main/Merge
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 002B17 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Main/Merge/biblio.hfd -nk 002B17 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Ticri/CIDE
   |area=    HapticV1
   |flux=    Main
   |étape=   Merge
   |type=    RBID
   |clé=     pubmed:21555118
   |texte=   Motor imagery facilitates force field learning.
}}

Pour générer des pages wiki

HfdIndexSelect -h $EXPLOR_AREA/Data/Main/Merge/RBID.i   -Sk "pubmed:21555118" \
       | HfdSelect -Kh $EXPLOR_AREA/Data/Main/Merge/biblio.hfd   \
       | NlmPubMed2Wicri -a HapticV1 

Wicri

This area was generated with Dilib version V0.6.23.
Data generation: Mon Jun 13 01:09:46 2016. Site generation: Wed Mar 6 09:54:07 2024