Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments
Identifieur interne : 000085 ( Pmc/Checkpoint ); précédent : 000084; suivant : 000086Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments
Auteurs : Yingli Tian [États-Unis] ; Xiaodong Yang [États-Unis] ; Chucai Yi [États-Unis] ; Aries Arditi [États-Unis]Source :
- Machine vision and applications [ 0932-8092 ] ; 2012.
Abstract
Independent travel is a well known challenge for blind and visually impaired persons. In this paper, we propose a proof-of-concept computer vision-based wayfinding aid for blind people to independently access unfamiliar indoor environments. In order to find different rooms (e.g. an office, a lab, or a bathroom) and other building amenities (e.g. an exit or an elevator), we incorporate object detection with text recognition. First we develop a robust and efficient algorithm to detect doors, elevators, and cabinets based on their general geometric shape, by combining edges and corners. The algorithm is general enough to handle large intra-class variations of objects with different appearances among different indoor environments, as well as small inter-class differences between different objects such as doors and door-like cabinets. Next, in order to distinguish intra-class objects (e.g. an office door from a bathroom door), we extract and recognize text information associated with the detected objects. For text recognition, we first extract text regions from signs with multiple colors and possibly complex backgrounds, and then apply character localization and topological analysis to filter out background interference. The extracted text is recognized using off-the-shelf optical character recognition (OCR) software products. The object type, orientation, location, and text information are presented to the blind traveler as speech.
Url:
DOI: 10.1007/s00138-012-0431-7
PubMed: 23630409
PubMed Central: 3636776
Affiliations:
Links toward previous steps (curation, corpus...)
Links to Exploration step
PMC:3636776Le document en format XML
<record><TEI><teiHeader><fileDesc><titleStmt><title xml:lang="en">Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments</title>
<author><name sortKey="Tian, Yingli" sort="Tian, Yingli" uniqKey="Tian Y" first="Yingli" last="Tian">Yingli Tian</name>
<affiliation wicri:level="2"><nlm:aff id="A1">Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York, NY 10031</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Yang, Xiaodong" sort="Yang, Xiaodong" uniqKey="Yang X" first="Xiaodong" last="Yang">Xiaodong Yang</name>
<affiliation wicri:level="2"><nlm:aff id="A1">Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York, NY 10031</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Yi, Chucai" sort="Yi, Chucai" uniqKey="Yi C" first="Chucai" last="Yi">Chucai Yi</name>
<affiliation wicri:level="2"><nlm:aff id="A2">The Graduate Center, City University of New York, New York, NY 10036</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>The Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Arditi, Aries" sort="Arditi, Aries" uniqKey="Arditi A" first="Aries" last="Arditi">Aries Arditi</name>
<affiliation wicri:level="2"><nlm:aff id="A3">Visibility Metrics LLC, Chappaqua, NY 10514 USA</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Visibility Metrics LLC, Chappaqua</wicri:cityArea>
</affiliation>
</author>
</titleStmt>
<publicationStmt><idno type="wicri:source">PMC</idno>
<idno type="pmid">23630409</idno>
<idno type="pmc">3636776</idno>
<idno type="url">http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3636776</idno>
<idno type="RBID">PMC:3636776</idno>
<idno type="doi">10.1007/s00138-012-0431-7</idno>
<date when="2012">2012</date>
<idno type="wicri:Area/Pmc/Corpus">000139</idno>
<idno type="wicri:Area/Pmc/Curation">000139</idno>
<idno type="wicri:Area/Pmc/Checkpoint">000085</idno>
</publicationStmt>
<sourceDesc><biblStruct><analytic><title xml:lang="en" level="a" type="main">Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments</title>
<author><name sortKey="Tian, Yingli" sort="Tian, Yingli" uniqKey="Tian Y" first="Yingli" last="Tian">Yingli Tian</name>
<affiliation wicri:level="2"><nlm:aff id="A1">Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York, NY 10031</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Yang, Xiaodong" sort="Yang, Xiaodong" uniqKey="Yang X" first="Xiaodong" last="Yang">Xiaodong Yang</name>
<affiliation wicri:level="2"><nlm:aff id="A1">Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York, NY 10031</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Yi, Chucai" sort="Yi, Chucai" uniqKey="Yi C" first="Chucai" last="Yi">Chucai Yi</name>
<affiliation wicri:level="2"><nlm:aff id="A2">The Graduate Center, City University of New York, New York, NY 10036</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>The Graduate Center, City University of New York, New York</wicri:cityArea>
</affiliation>
</author>
<author><name sortKey="Arditi, Aries" sort="Arditi, Aries" uniqKey="Arditi A" first="Aries" last="Arditi">Aries Arditi</name>
<affiliation wicri:level="2"><nlm:aff id="A3">Visibility Metrics LLC, Chappaqua, NY 10514 USA</nlm:aff>
<country xml:lang="fr">États-Unis</country>
<placeName><region type="state">État de New York</region>
</placeName>
<wicri:cityArea>Visibility Metrics LLC, Chappaqua</wicri:cityArea>
</affiliation>
</author>
</analytic>
<series><title level="j">Machine vision and applications</title>
<idno type="ISSN">0932-8092</idno>
<idno type="eISSN">1432-1769</idno>
<imprint><date when="2012">2012</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc><textClass></textClass>
</profileDesc>
</teiHeader>
<front><div type="abstract" xml:lang="en"><p id="P1">Independent travel is a well known challenge for blind and visually impaired persons. In this paper, we propose a proof-of-concept computer vision-based wayfinding aid for blind people to independently access unfamiliar indoor environments. In order to find different rooms (e.g. an office, a lab, or a bathroom) and other building amenities (e.g. an exit or an elevator), we incorporate object detection with text recognition. First we develop a robust and efficient algorithm to detect doors, elevators, and cabinets based on their general geometric shape, by combining edges and corners. The algorithm is general enough to handle large intra-class variations of objects with different appearances among different indoor environments, as well as small inter-class differences between different objects such as doors and door-like cabinets. Next, in order to distinguish intra-class objects (e.g. an office door from a bathroom door), we extract and recognize text information associated with the detected objects. For text recognition, we first extract text regions from signs with multiple colors and possibly complex backgrounds, and then apply character localization and topological analysis to filter out background interference. The extracted text is recognized using off-the-shelf optical character recognition (OCR) software products. The object type, orientation, location, and text information are presented to the blind traveler as speech.</p>
</div>
</front>
</TEI>
<pmc article-type="research-article"><pmc-comment>The publisher of this article does not allow downloading of the full text in XML form.</pmc-comment>
<pmc-dir>properties manuscript</pmc-dir>
<front><journal-meta><journal-id journal-id-type="nlm-journal-id">101562623</journal-id>
<journal-id journal-id-type="pubmed-jr-id">39268</journal-id>
<journal-id journal-id-type="nlm-ta">Mach Vis Appl</journal-id>
<journal-id journal-id-type="iso-abbrev">Mach Vis Appl</journal-id>
<journal-title-group><journal-title>Machine vision and applications</journal-title>
</journal-title-group>
<issn pub-type="ppub">0932-8092</issn>
<issn pub-type="epub">1432-1769</issn>
</journal-meta>
<article-meta><article-id pub-id-type="pmid">23630409</article-id>
<article-id pub-id-type="pmc">3636776</article-id>
<article-id pub-id-type="doi">10.1007/s00138-012-0431-7</article-id>
<article-id pub-id-type="manuscript">NIHMS371367</article-id>
<article-categories><subj-group subj-group-type="heading"><subject>Article</subject>
</subj-group>
</article-categories>
<title-group><article-title>Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments</article-title>
</title-group>
<contrib-group><contrib contrib-type="author"><name><surname>Tian</surname>
<given-names>YingLi</given-names>
</name>
<xref ref-type="aff" rid="A1">1</xref>
<email>ytian@ccny.cuny.edu</email>
</contrib>
<contrib contrib-type="author"><name><surname>Yang</surname>
<given-names>Xiaodong</given-names>
</name>
<xref ref-type="aff" rid="A1">1</xref>
<email>xyang02@ccny.cuny.edu</email>
</contrib>
<contrib contrib-type="author"><name><surname>Yi</surname>
<given-names>Chucai</given-names>
</name>
<xref ref-type="aff" rid="A2">2</xref>
<email>cyi@gc.cuny.edu</email>
</contrib>
<contrib contrib-type="author"><name><surname>Arditi</surname>
<given-names>Aries</given-names>
</name>
<xref ref-type="aff" rid="A3">3</xref>
<email>arditi@visibilitymetrics.com</email>
</contrib>
</contrib-group>
<aff id="A1"><label>1</label>
Electrical Engineering Department, The City College, and Graduate Center, City University of New York, New York, NY 10031</aff>
<aff id="A2"><label>2</label>
The Graduate Center, City University of New York, New York, NY 10036</aff>
<aff id="A3"><label>3</label>
Visibility Metrics LLC, Chappaqua, NY 10514 USA</aff>
<pub-date pub-type="nihms-submitted"><day>19</day>
<month>4</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="epub"><day>24</day>
<month>5</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="ppub"><day>1</day>
<month>4</month>
<year>2013</year>
</pub-date>
<pub-date pub-type="pmc-release"><day>01</day>
<month>4</month>
<year>2014</year>
</pub-date>
<volume>24</volume>
<issue>3</issue>
<fpage>521</fpage>
<lpage>535</lpage>
<abstract><p id="P1">Independent travel is a well known challenge for blind and visually impaired persons. In this paper, we propose a proof-of-concept computer vision-based wayfinding aid for blind people to independently access unfamiliar indoor environments. In order to find different rooms (e.g. an office, a lab, or a bathroom) and other building amenities (e.g. an exit or an elevator), we incorporate object detection with text recognition. First we develop a robust and efficient algorithm to detect doors, elevators, and cabinets based on their general geometric shape, by combining edges and corners. The algorithm is general enough to handle large intra-class variations of objects with different appearances among different indoor environments, as well as small inter-class differences between different objects such as doors and door-like cabinets. Next, in order to distinguish intra-class objects (e.g. an office door from a bathroom door), we extract and recognize text information associated with the detected objects. For text recognition, we first extract text regions from signs with multiple colors and possibly complex backgrounds, and then apply character localization and topological analysis to filter out background interference. The extracted text is recognized using off-the-shelf optical character recognition (OCR) software products. The object type, orientation, location, and text information are presented to the blind traveler as speech.</p>
</abstract>
<kwd-group><kwd>Indoor wayfinding</kwd>
<kwd>computer vision</kwd>
<kwd>object detection</kwd>
<kwd>text extraction</kwd>
<kwd>optical character recognition (OCR)</kwd>
<kwd>blind/visually impaired persons</kwd>
</kwd-group>
<funding-group><award-group><funding-source country="United States">National Eye Institute : NEI</funding-source>
<award-id>R21 EY020990-01 || EY</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
</pmc>
<affiliations><list><country><li>États-Unis</li>
</country>
<region><li>État de New York</li>
</region>
</list>
<tree><country name="États-Unis"><region name="État de New York"><name sortKey="Tian, Yingli" sort="Tian, Yingli" uniqKey="Tian Y" first="Yingli" last="Tian">Yingli Tian</name>
</region>
<name sortKey="Arditi, Aries" sort="Arditi, Aries" uniqKey="Arditi A" first="Aries" last="Arditi">Aries Arditi</name>
<name sortKey="Yang, Xiaodong" sort="Yang, Xiaodong" uniqKey="Yang X" first="Xiaodong" last="Yang">Xiaodong Yang</name>
<name sortKey="Yi, Chucai" sort="Yi, Chucai" uniqKey="Yi C" first="Chucai" last="Yi">Chucai Yi</name>
</country>
</tree>
</affiliations>
</record>
Pour manipuler ce document sous Unix (Dilib)
EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/OcrV1/Data/Pmc/Checkpoint
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000085 | SxmlIndent | more
Ou
HfdSelect -h $EXPLOR_AREA/Data/Pmc/Checkpoint/biblio.hfd -nk 000085 | SxmlIndent | more
Pour mettre un lien sur cette page dans le réseau Wicri
{{Explor lien |wiki= Ticri/CIDE |area= OcrV1 |flux= Pmc |étape= Checkpoint |type= RBID |clé= PMC:3636776 |texte= Toward a Computer Vision-based Wayfinding Aid for Blind Persons to Access Unfamiliar Indoor Environments }}
Pour générer des pages wiki
HfdIndexSelect -h $EXPLOR_AREA/Data/Pmc/Checkpoint/RBID.i -Sk "pubmed:23630409" \ | HfdSelect -Kh $EXPLOR_AREA/Data/Pmc/Checkpoint/biblio.hfd \ | NlmPubMed2Wicri -a OcrV1
This area was generated with Dilib version V0.6.32. |