Text extraction from natural scene image: A survey
Identifieur interne : 000043 ( PascalFrancis/Corpus ); précédent : 000042; suivant : 000044Text extraction from natural scene image: A survey
Auteurs : HONGGANG ZHANG ; KAILI ZHAO ; Yi-Zhe Song ; JUN GUOSource :
- Neurocomputing : (Amsterdam) [ 0925-2312 ] ; 2013.
Descripteurs français
- Pascal (Inist)
- Extraction information, Texte, Analyse scène, Calculateur embarqué, Traitement image, Réalité virtuelle, Interface utilisateur, Recherche information, Vision ordinateur, Localisation, Reconnaissance optique caractère, Reconnaissance caractère, Appareil photographique, Réalité augmentée, Alignement, Luminance, Procédé extraction, Extraction forme, Algorithmique, Segmentation, Eclairement, Scène naturelle, Accentuation image, Traitement image document.
English descriptors
- KwdEn :
- Algorithmics, Alignment, Augmented reality, Boarded computer, Camera, Character recognition, Computer vision, Document image processing, Extraction process, Illumination, Image enhancement, Image processing, Information extraction, Information retrieval, Localization, Luminance, Natural scenes, Optical character recognition, Pattern extraction, Scene analysis, Segmentation, Text, User interface, Virtual reality.
Abstract
With the increasing popularity of portable camera devices and embedded visual processing, text extraction from natural scene images has become a key problem that is deemed to change our everyday lives via novel applications such as augmented reality. Text extraction from natural scene images algorithms is generally composed of the following three stages: (i) detection and localization, (ii) text enhancement and segmentation and (iii) optical character recognition (OCR). The problem is challenging in nature due to variations in the font size and color, text alignment, illumination change and reflections. This paper aims to classify and assess the latest algorithms. More specifically, we draw attention to studies on the first two steps in the extraction process, since OCR is a well-studied area where powerful algorithms already exist. This paper offers to the researchers a link to public image database for the algorithm assessment of text extraction from natural scene images.
Notice en format standard (ISO 2709)
Pour connaître la documentation sur le format Inist Standard.
pA |
|
---|
Format Inist (serveur)
NO : | PASCAL 13-0324769 INIST |
---|---|
ET : | Text extraction from natural scene image: A survey |
AU : | HONGGANG ZHANG; KAILI ZHAO; SONG (Yi-Zhe); JUN GUO |
AF : | School of Communication and Information Engineering, Beijing University of Posts and Telecommunications/Beijing/Chine (1 aut., 2 aut., 4 aut.); School of Electronic Engineering and Computer Science, Queen Mary, University of London/London E1 4NS/Royaume-Uni (3 aut.) |
DT : | Publication en série; Niveau analytique |
SO : | Neurocomputing : (Amsterdam); ISSN 0925-2312; Pays-Bas; Da. 2013; Vol. 122; Pp. 310-323; Bibl. 91 ref. |
LA : | Anglais |
EA : | With the increasing popularity of portable camera devices and embedded visual processing, text extraction from natural scene images has become a key problem that is deemed to change our everyday lives via novel applications such as augmented reality. Text extraction from natural scene images algorithms is generally composed of the following three stages: (i) detection and localization, (ii) text enhancement and segmentation and (iii) optical character recognition (OCR). The problem is challenging in nature due to variations in the font size and color, text alignment, illumination change and reflections. This paper aims to classify and assess the latest algorithms. More specifically, we draw attention to studies on the first two steps in the extraction process, since OCR is a well-studied area where powerful algorithms already exist. This paper offers to the researchers a link to public image database for the algorithm assessment of text extraction from natural scene images. |
CC : | 001D02C03; 001D02C04; 001D02B04; 001D02A05 |
FD : | Extraction information; Texte; Analyse scène; Calculateur embarqué; Traitement image; Réalité virtuelle; Interface utilisateur; Recherche information; Vision ordinateur; Localisation; Reconnaissance optique caractère; Reconnaissance caractère; Appareil photographique; Réalité augmentée; Alignement; Luminance; Procédé extraction; Extraction forme; Algorithmique; Segmentation; Eclairement; Scène naturelle; Accentuation image; Traitement image document |
ED : | Information extraction; Text; Scene analysis; Boarded computer; Image processing; Virtual reality; User interface; Information retrieval; Computer vision; Localization; Optical character recognition; Character recognition; Camera; Augmented reality; Alignment; Luminance; Extraction process; Pattern extraction; Algorithmics; Segmentation; Illumination; Natural scenes; Image enhancement; Document image processing |
SD : | Extracción información; Texto; Análisis escena; Calculador embarque; Procesamiento imagen; Realidad virtual; Interfase usuario; Búsqueda información; Visión ordenador; Localización; Reconocimento óptico de caracteres; Reconocimiento carácter; Máquina fotográfica; Realidad aumentada; Alineamiento; Luminancia; Procedimiento extracción; Extracción forma; Algorítmica; Segmentación; Alumbrado; Escena natural; Mejoramiento de imágenes; Tratamiento de imágenes de documentos |
LO : | INIST-28121.354000501570810320 |
ID : | 13-0324769 |
Links to Exploration step
Pascal:13-0324769Le document en format XML
<record><TEI><teiHeader><fileDesc><titleStmt><title xml:lang="en" level="a">Text extraction from natural scene image: A survey</title>
<author><name sortKey="Honggang Zhang" sort="Honggang Zhang" uniqKey="Honggang Zhang" last="Honggang Zhang">HONGGANG ZHANG</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Kaili Zhao" sort="Kaili Zhao" uniqKey="Kaili Zhao" last="Kaili Zhao">KAILI ZHAO</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Song, Yi Zhe" sort="Song, Yi Zhe" uniqKey="Song Y" first="Yi-Zhe" last="Song">Yi-Zhe Song</name>
<affiliation><inist:fA14 i1="02"><s1>School of Electronic Engineering and Computer Science, Queen Mary, University of London</s1>
<s2>London E1 4NS</s2>
<s3>GBR</s3>
<sZ>3 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Jun Guo" sort="Jun Guo" uniqKey="Jun Guo" last="Jun Guo">JUN GUO</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
</titleStmt>
<publicationStmt><idno type="wicri:source">INIST</idno>
<idno type="inist">13-0324769</idno>
<date when="2013">2013</date>
<idno type="stanalyst">PASCAL 13-0324769 INIST</idno>
<idno type="RBID">Pascal:13-0324769</idno>
<idno type="wicri:Area/PascalFrancis/Corpus">000043</idno>
</publicationStmt>
<sourceDesc><biblStruct><analytic><title xml:lang="en" level="a">Text extraction from natural scene image: A survey</title>
<author><name sortKey="Honggang Zhang" sort="Honggang Zhang" uniqKey="Honggang Zhang" last="Honggang Zhang">HONGGANG ZHANG</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Kaili Zhao" sort="Kaili Zhao" uniqKey="Kaili Zhao" last="Kaili Zhao">KAILI ZHAO</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Song, Yi Zhe" sort="Song, Yi Zhe" uniqKey="Song Y" first="Yi-Zhe" last="Song">Yi-Zhe Song</name>
<affiliation><inist:fA14 i1="02"><s1>School of Electronic Engineering and Computer Science, Queen Mary, University of London</s1>
<s2>London E1 4NS</s2>
<s3>GBR</s3>
<sZ>3 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
<author><name sortKey="Jun Guo" sort="Jun Guo" uniqKey="Jun Guo" last="Jun Guo">JUN GUO</name>
<affiliation><inist:fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</inist:fA14>
</affiliation>
</author>
</analytic>
<series><title level="j" type="main">Neurocomputing : (Amsterdam)</title>
<title level="j" type="abbreviated">Neurocomputing : (Amst.)</title>
<idno type="ISSN">0925-2312</idno>
<imprint><date when="2013">2013</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
<seriesStmt><title level="j" type="main">Neurocomputing : (Amsterdam)</title>
<title level="j" type="abbreviated">Neurocomputing : (Amst.)</title>
<idno type="ISSN">0925-2312</idno>
</seriesStmt>
</fileDesc>
<profileDesc><textClass><keywords scheme="KwdEn" xml:lang="en"><term>Algorithmics</term>
<term>Alignment</term>
<term>Augmented reality</term>
<term>Boarded computer</term>
<term>Camera</term>
<term>Character recognition</term>
<term>Computer vision</term>
<term>Document image processing</term>
<term>Extraction process</term>
<term>Illumination</term>
<term>Image enhancement</term>
<term>Image processing</term>
<term>Information extraction</term>
<term>Information retrieval</term>
<term>Localization</term>
<term>Luminance</term>
<term>Natural scenes</term>
<term>Optical character recognition</term>
<term>Pattern extraction</term>
<term>Scene analysis</term>
<term>Segmentation</term>
<term>Text</term>
<term>User interface</term>
<term>Virtual reality</term>
</keywords>
<keywords scheme="Pascal" xml:lang="fr"><term>Extraction information</term>
<term>Texte</term>
<term>Analyse scène</term>
<term>Calculateur embarqué</term>
<term>Traitement image</term>
<term>Réalité virtuelle</term>
<term>Interface utilisateur</term>
<term>Recherche information</term>
<term>Vision ordinateur</term>
<term>Localisation</term>
<term>Reconnaissance optique caractère</term>
<term>Reconnaissance caractère</term>
<term>Appareil photographique</term>
<term>Réalité augmentée</term>
<term>Alignement</term>
<term>Luminance</term>
<term>Procédé extraction</term>
<term>Extraction forme</term>
<term>Algorithmique</term>
<term>Segmentation</term>
<term>Eclairement</term>
<term>Scène naturelle</term>
<term>Accentuation image</term>
<term>Traitement image document</term>
</keywords>
</textClass>
</profileDesc>
</teiHeader>
<front><div type="abstract" xml:lang="en">With the increasing popularity of portable camera devices and embedded visual processing, text extraction from natural scene images has become a key problem that is deemed to change our everyday lives via novel applications such as augmented reality. Text extraction from natural scene images algorithms is generally composed of the following three stages: (i) detection and localization, (ii) text enhancement and segmentation and (iii) optical character recognition (OCR). The problem is challenging in nature due to variations in the font size and color, text alignment, illumination change and reflections. This paper aims to classify and assess the latest algorithms. More specifically, we draw attention to studies on the first two steps in the extraction process, since OCR is a well-studied area where powerful algorithms already exist. This paper offers to the researchers a link to public image database for the algorithm assessment of text extraction from natural scene images.</div>
</front>
</TEI>
<inist><standard h6="B"><pA><fA01 i1="01" i2="1"><s0>0925-2312</s0>
</fA01>
<fA03 i2="1"><s0>Neurocomputing : (Amst.)</s0>
</fA03>
<fA05><s2>122</s2>
</fA05>
<fA08 i1="01" i2="1" l="ENG"><s1>Text extraction from natural scene image: A survey</s1>
</fA08>
<fA11 i1="01" i2="1"><s1>HONGGANG ZHANG</s1>
</fA11>
<fA11 i1="02" i2="1"><s1>KAILI ZHAO</s1>
</fA11>
<fA11 i1="03" i2="1"><s1>SONG (Yi-Zhe)</s1>
</fA11>
<fA11 i1="04" i2="1"><s1>JUN GUO</s1>
</fA11>
<fA14 i1="01"><s1>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications</s1>
<s2>Beijing</s2>
<s3>CHN</s3>
<sZ>1 aut.</sZ>
<sZ>2 aut.</sZ>
<sZ>4 aut.</sZ>
</fA14>
<fA14 i1="02"><s1>School of Electronic Engineering and Computer Science, Queen Mary, University of London</s1>
<s2>London E1 4NS</s2>
<s3>GBR</s3>
<sZ>3 aut.</sZ>
</fA14>
<fA20><s1>310-323</s1>
</fA20>
<fA21><s1>2013</s1>
</fA21>
<fA23 i1="01"><s0>ENG</s0>
</fA23>
<fA43 i1="01"><s1>INIST</s1>
<s2>28121</s2>
<s5>354000501570810320</s5>
</fA43>
<fA44><s0>0000</s0>
<s1>© 2013 INIST-CNRS. All rights reserved.</s1>
</fA44>
<fA45><s0>91 ref.</s0>
</fA45>
<fA47 i1="01" i2="1"><s0>13-0324769</s0>
</fA47>
<fA60><s1>P</s1>
</fA60>
<fA61><s0>A</s0>
</fA61>
<fA64 i1="01" i2="1"><s0>Neurocomputing : (Amsterdam)</s0>
</fA64>
<fA66 i1="01"><s0>NLD</s0>
</fA66>
<fC01 i1="01" l="ENG"><s0>With the increasing popularity of portable camera devices and embedded visual processing, text extraction from natural scene images has become a key problem that is deemed to change our everyday lives via novel applications such as augmented reality. Text extraction from natural scene images algorithms is generally composed of the following three stages: (i) detection and localization, (ii) text enhancement and segmentation and (iii) optical character recognition (OCR). The problem is challenging in nature due to variations in the font size and color, text alignment, illumination change and reflections. This paper aims to classify and assess the latest algorithms. More specifically, we draw attention to studies on the first two steps in the extraction process, since OCR is a well-studied area where powerful algorithms already exist. This paper offers to the researchers a link to public image database for the algorithm assessment of text extraction from natural scene images.</s0>
</fC01>
<fC02 i1="01" i2="X"><s0>001D02C03</s0>
</fC02>
<fC02 i1="02" i2="X"><s0>001D02C04</s0>
</fC02>
<fC02 i1="03" i2="X"><s0>001D02B04</s0>
</fC02>
<fC02 i1="04" i2="X"><s0>001D02A05</s0>
</fC02>
<fC03 i1="01" i2="X" l="FRE"><s0>Extraction information</s0>
<s5>06</s5>
</fC03>
<fC03 i1="01" i2="X" l="ENG"><s0>Information extraction</s0>
<s5>06</s5>
</fC03>
<fC03 i1="01" i2="X" l="SPA"><s0>Extracción información</s0>
<s5>06</s5>
</fC03>
<fC03 i1="02" i2="X" l="FRE"><s0>Texte</s0>
<s5>07</s5>
</fC03>
<fC03 i1="02" i2="X" l="ENG"><s0>Text</s0>
<s5>07</s5>
</fC03>
<fC03 i1="02" i2="X" l="SPA"><s0>Texto</s0>
<s5>07</s5>
</fC03>
<fC03 i1="03" i2="X" l="FRE"><s0>Analyse scène</s0>
<s5>08</s5>
</fC03>
<fC03 i1="03" i2="X" l="ENG"><s0>Scene analysis</s0>
<s5>08</s5>
</fC03>
<fC03 i1="03" i2="X" l="SPA"><s0>Análisis escena</s0>
<s5>08</s5>
</fC03>
<fC03 i1="04" i2="X" l="FRE"><s0>Calculateur embarqué</s0>
<s5>09</s5>
</fC03>
<fC03 i1="04" i2="X" l="ENG"><s0>Boarded computer</s0>
<s5>09</s5>
</fC03>
<fC03 i1="04" i2="X" l="SPA"><s0>Calculador embarque</s0>
<s5>09</s5>
</fC03>
<fC03 i1="05" i2="X" l="FRE"><s0>Traitement image</s0>
<s5>10</s5>
</fC03>
<fC03 i1="05" i2="X" l="ENG"><s0>Image processing</s0>
<s5>10</s5>
</fC03>
<fC03 i1="05" i2="X" l="SPA"><s0>Procesamiento imagen</s0>
<s5>10</s5>
</fC03>
<fC03 i1="06" i2="X" l="FRE"><s0>Réalité virtuelle</s0>
<s5>11</s5>
</fC03>
<fC03 i1="06" i2="X" l="ENG"><s0>Virtual reality</s0>
<s5>11</s5>
</fC03>
<fC03 i1="06" i2="X" l="SPA"><s0>Realidad virtual</s0>
<s5>11</s5>
</fC03>
<fC03 i1="07" i2="X" l="FRE"><s0>Interface utilisateur</s0>
<s5>12</s5>
</fC03>
<fC03 i1="07" i2="X" l="ENG"><s0>User interface</s0>
<s5>12</s5>
</fC03>
<fC03 i1="07" i2="X" l="SPA"><s0>Interfase usuario</s0>
<s5>12</s5>
</fC03>
<fC03 i1="08" i2="X" l="FRE"><s0>Recherche information</s0>
<s5>13</s5>
</fC03>
<fC03 i1="08" i2="X" l="ENG"><s0>Information retrieval</s0>
<s5>13</s5>
</fC03>
<fC03 i1="08" i2="X" l="SPA"><s0>Búsqueda información</s0>
<s5>13</s5>
</fC03>
<fC03 i1="09" i2="X" l="FRE"><s0>Vision ordinateur</s0>
<s5>14</s5>
</fC03>
<fC03 i1="09" i2="X" l="ENG"><s0>Computer vision</s0>
<s5>14</s5>
</fC03>
<fC03 i1="09" i2="X" l="SPA"><s0>Visión ordenador</s0>
<s5>14</s5>
</fC03>
<fC03 i1="10" i2="X" l="FRE"><s0>Localisation</s0>
<s5>15</s5>
</fC03>
<fC03 i1="10" i2="X" l="ENG"><s0>Localization</s0>
<s5>15</s5>
</fC03>
<fC03 i1="10" i2="X" l="SPA"><s0>Localización</s0>
<s5>15</s5>
</fC03>
<fC03 i1="11" i2="X" l="FRE"><s0>Reconnaissance optique caractère</s0>
<s5>16</s5>
</fC03>
<fC03 i1="11" i2="X" l="ENG"><s0>Optical character recognition</s0>
<s5>16</s5>
</fC03>
<fC03 i1="11" i2="X" l="SPA"><s0>Reconocimento óptico de caracteres</s0>
<s5>16</s5>
</fC03>
<fC03 i1="12" i2="X" l="FRE"><s0>Reconnaissance caractère</s0>
<s5>17</s5>
</fC03>
<fC03 i1="12" i2="X" l="ENG"><s0>Character recognition</s0>
<s5>17</s5>
</fC03>
<fC03 i1="12" i2="X" l="SPA"><s0>Reconocimiento carácter</s0>
<s5>17</s5>
</fC03>
<fC03 i1="13" i2="X" l="FRE"><s0>Appareil photographique</s0>
<s5>18</s5>
</fC03>
<fC03 i1="13" i2="X" l="ENG"><s0>Camera</s0>
<s5>18</s5>
</fC03>
<fC03 i1="13" i2="X" l="SPA"><s0>Máquina fotográfica</s0>
<s5>18</s5>
</fC03>
<fC03 i1="14" i2="X" l="FRE"><s0>Réalité augmentée</s0>
<s5>19</s5>
</fC03>
<fC03 i1="14" i2="X" l="ENG"><s0>Augmented reality</s0>
<s5>19</s5>
</fC03>
<fC03 i1="14" i2="X" l="SPA"><s0>Realidad aumentada</s0>
<s5>19</s5>
</fC03>
<fC03 i1="15" i2="X" l="FRE"><s0>Alignement</s0>
<s5>20</s5>
</fC03>
<fC03 i1="15" i2="X" l="ENG"><s0>Alignment</s0>
<s5>20</s5>
</fC03>
<fC03 i1="15" i2="X" l="SPA"><s0>Alineamiento</s0>
<s5>20</s5>
</fC03>
<fC03 i1="16" i2="X" l="FRE"><s0>Luminance</s0>
<s5>21</s5>
</fC03>
<fC03 i1="16" i2="X" l="ENG"><s0>Luminance</s0>
<s5>21</s5>
</fC03>
<fC03 i1="16" i2="X" l="SPA"><s0>Luminancia</s0>
<s5>21</s5>
</fC03>
<fC03 i1="17" i2="X" l="FRE"><s0>Procédé extraction</s0>
<s5>22</s5>
</fC03>
<fC03 i1="17" i2="X" l="ENG"><s0>Extraction process</s0>
<s5>22</s5>
</fC03>
<fC03 i1="17" i2="X" l="SPA"><s0>Procedimiento extracción</s0>
<s5>22</s5>
</fC03>
<fC03 i1="18" i2="X" l="FRE"><s0>Extraction forme</s0>
<s5>23</s5>
</fC03>
<fC03 i1="18" i2="X" l="ENG"><s0>Pattern extraction</s0>
<s5>23</s5>
</fC03>
<fC03 i1="18" i2="X" l="SPA"><s0>Extracción forma</s0>
<s5>23</s5>
</fC03>
<fC03 i1="19" i2="X" l="FRE"><s0>Algorithmique</s0>
<s5>24</s5>
</fC03>
<fC03 i1="19" i2="X" l="ENG"><s0>Algorithmics</s0>
<s5>24</s5>
</fC03>
<fC03 i1="19" i2="X" l="SPA"><s0>Algorítmica</s0>
<s5>24</s5>
</fC03>
<fC03 i1="20" i2="X" l="FRE"><s0>Segmentation</s0>
<s5>25</s5>
</fC03>
<fC03 i1="20" i2="X" l="ENG"><s0>Segmentation</s0>
<s5>25</s5>
</fC03>
<fC03 i1="20" i2="X" l="SPA"><s0>Segmentación</s0>
<s5>25</s5>
</fC03>
<fC03 i1="21" i2="X" l="FRE"><s0>Eclairement</s0>
<s5>41</s5>
</fC03>
<fC03 i1="21" i2="X" l="ENG"><s0>Illumination</s0>
<s5>41</s5>
</fC03>
<fC03 i1="21" i2="X" l="SPA"><s0>Alumbrado</s0>
<s5>41</s5>
</fC03>
<fC03 i1="22" i2="X" l="FRE"><s0>Scène naturelle</s0>
<s4>CD</s4>
<s5>96</s5>
</fC03>
<fC03 i1="22" i2="X" l="ENG"><s0>Natural scenes</s0>
<s4>CD</s4>
<s5>96</s5>
</fC03>
<fC03 i1="22" i2="X" l="SPA"><s0>Escena natural</s0>
<s4>CD</s4>
<s5>96</s5>
</fC03>
<fC03 i1="23" i2="X" l="FRE"><s0>Accentuation image</s0>
<s4>CD</s4>
<s5>97</s5>
</fC03>
<fC03 i1="23" i2="X" l="ENG"><s0>Image enhancement</s0>
<s4>CD</s4>
<s5>97</s5>
</fC03>
<fC03 i1="23" i2="X" l="SPA"><s0>Mejoramiento de imágenes</s0>
<s4>CD</s4>
<s5>97</s5>
</fC03>
<fC03 i1="24" i2="X" l="FRE"><s0>Traitement image document</s0>
<s4>CD</s4>
<s5>98</s5>
</fC03>
<fC03 i1="24" i2="X" l="ENG"><s0>Document image processing</s0>
<s4>CD</s4>
<s5>98</s5>
</fC03>
<fC03 i1="24" i2="X" l="SPA"><s0>Tratamiento de imágenes de documentos</s0>
<s4>CD</s4>
<s5>98</s5>
</fC03>
<fN21><s1>308</s1>
</fN21>
<fN44 i1="01"><s1>OTO</s1>
</fN44>
<fN82><s1>OTO</s1>
</fN82>
</pA>
</standard>
<server><NO>PASCAL 13-0324769 INIST</NO>
<ET>Text extraction from natural scene image: A survey</ET>
<AU>HONGGANG ZHANG; KAILI ZHAO; SONG (Yi-Zhe); JUN GUO</AU>
<AF>School of Communication and Information Engineering, Beijing University of Posts and Telecommunications/Beijing/Chine (1 aut., 2 aut., 4 aut.); School of Electronic Engineering and Computer Science, Queen Mary, University of London/London E1 4NS/Royaume-Uni (3 aut.)</AF>
<DT>Publication en série; Niveau analytique</DT>
<SO>Neurocomputing : (Amsterdam); ISSN 0925-2312; Pays-Bas; Da. 2013; Vol. 122; Pp. 310-323; Bibl. 91 ref.</SO>
<LA>Anglais</LA>
<EA>With the increasing popularity of portable camera devices and embedded visual processing, text extraction from natural scene images has become a key problem that is deemed to change our everyday lives via novel applications such as augmented reality. Text extraction from natural scene images algorithms is generally composed of the following three stages: (i) detection and localization, (ii) text enhancement and segmentation and (iii) optical character recognition (OCR). The problem is challenging in nature due to variations in the font size and color, text alignment, illumination change and reflections. This paper aims to classify and assess the latest algorithms. More specifically, we draw attention to studies on the first two steps in the extraction process, since OCR is a well-studied area where powerful algorithms already exist. This paper offers to the researchers a link to public image database for the algorithm assessment of text extraction from natural scene images.</EA>
<CC>001D02C03; 001D02C04; 001D02B04; 001D02A05</CC>
<FD>Extraction information; Texte; Analyse scène; Calculateur embarqué; Traitement image; Réalité virtuelle; Interface utilisateur; Recherche information; Vision ordinateur; Localisation; Reconnaissance optique caractère; Reconnaissance caractère; Appareil photographique; Réalité augmentée; Alignement; Luminance; Procédé extraction; Extraction forme; Algorithmique; Segmentation; Eclairement; Scène naturelle; Accentuation image; Traitement image document</FD>
<ED>Information extraction; Text; Scene analysis; Boarded computer; Image processing; Virtual reality; User interface; Information retrieval; Computer vision; Localization; Optical character recognition; Character recognition; Camera; Augmented reality; Alignment; Luminance; Extraction process; Pattern extraction; Algorithmics; Segmentation; Illumination; Natural scenes; Image enhancement; Document image processing</ED>
<SD>Extracción información; Texto; Análisis escena; Calculador embarque; Procesamiento imagen; Realidad virtual; Interfase usuario; Búsqueda información; Visión ordenador; Localización; Reconocimento óptico de caracteres; Reconocimiento carácter; Máquina fotográfica; Realidad aumentada; Alineamiento; Luminancia; Procedimiento extracción; Extracción forma; Algorítmica; Segmentación; Alumbrado; Escena natural; Mejoramiento de imágenes; Tratamiento de imágenes de documentos</SD>
<LO>INIST-28121.354000501570810320</LO>
<ID>13-0324769</ID>
</server>
</inist>
</record>
Pour manipuler ce document sous Unix (Dilib)
EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/OcrV1/Data/PascalFrancis/Corpus
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000043 | SxmlIndent | more
Ou
HfdSelect -h $EXPLOR_AREA/Data/PascalFrancis/Corpus/biblio.hfd -nk 000043 | SxmlIndent | more
Pour mettre un lien sur cette page dans le réseau Wicri
{{Explor lien |wiki= Ticri/CIDE |area= OcrV1 |flux= PascalFrancis |étape= Corpus |type= RBID |clé= Pascal:13-0324769 |texte= Text extraction from natural scene image: A survey }}
This area was generated with Dilib version V0.6.32. |