Serveur d'exploration sur les dispositifs haptiques

Attention, ce site est en cours de développement !
Attention, site généré par des moyens informatiques à partir de corpus bruts.
Les informations ne sont donc pas validées.

Visuohaptic Discrimination of 3D Gross Shape

Identifieur interne : 002077 ( Istex/Corpus ); précédent : 002076; suivant : 002078

Visuohaptic Discrimination of 3D Gross Shape

Auteurs : Hong Z. Tan ; Kwangtaek Kim ; Mauro Barni ; Domenico Prattichizzo

Source :

RBID : ISTEX:AFAF5DD994945AA8FECF4AE82311F74BB899DEAC
Url:
DOI: 10.1163/187847612X629937

Links to Exploration step

ISTEX:AFAF5DD994945AA8FECF4AE82311F74BB899DEAC

Le document en format XML

<record>
<TEI wicri:istexFullTextTei="biblStruct">
<teiHeader>
<fileDesc>
<titleStmt>
<title>Visuohaptic Discrimination of 3D Gross Shape</title>
<author>
<name sortKey="Tan, Hong Z" sort="Tan, Hong Z" uniqKey="Tan H" first="Hong Z." last="Tan">Hong Z. Tan</name>
</author>
<author>
<name sortKey="Kim, Kwangtaek" sort="Kim, Kwangtaek" uniqKey="Kim K" first="Kwangtaek" last="Kim">Kwangtaek Kim</name>
</author>
<author>
<name sortKey="Barni, Mauro" sort="Barni, Mauro" uniqKey="Barni M" first="Mauro" last="Barni">Mauro Barni</name>
</author>
<author>
<name sortKey="Prattichizzo, Domenico" sort="Prattichizzo, Domenico" uniqKey="Prattichizzo D" first="Domenico" last="Prattichizzo">Domenico Prattichizzo</name>
</author>
</titleStmt>
<publicationStmt>
<idno type="wicri:source">ISTEX</idno>
<idno type="RBID">ISTEX:AFAF5DD994945AA8FECF4AE82311F74BB899DEAC</idno>
<date when="2012" year="2012">2012</date>
<idno type="doi">10.1163/187847612X629937</idno>
<idno type="url">https://api.istex.fr/document/AFAF5DD994945AA8FECF4AE82311F74BB899DEAC/fulltext/pdf</idno>
<idno type="wicri:Area/Istex/Corpus">002077</idno>
</publicationStmt>
<sourceDesc>
<biblStruct>
<analytic>
<title level="a">Visuohaptic Discrimination of 3D Gross Shape</title>
<author>
<name sortKey="Tan, Hong Z" sort="Tan, Hong Z" uniqKey="Tan H" first="Hong Z." last="Tan">Hong Z. Tan</name>
</author>
<author>
<name sortKey="Kim, Kwangtaek" sort="Kim, Kwangtaek" uniqKey="Kim K" first="Kwangtaek" last="Kim">Kwangtaek Kim</name>
</author>
<author>
<name sortKey="Barni, Mauro" sort="Barni, Mauro" uniqKey="Barni M" first="Mauro" last="Barni">Mauro Barni</name>
</author>
<author>
<name sortKey="Prattichizzo, Domenico" sort="Prattichizzo, Domenico" uniqKey="Prattichizzo D" first="Domenico" last="Prattichizzo">Domenico Prattichizzo</name>
</author>
</analytic>
<monogr></monogr>
<series>
<title level="j">Seeing and Perceiving</title>
<title level="j" type="sub">A Journal of Multisensory Science</title>
<title level="j" type="abbrev">SP</title>
<idno type="ISSN">1878-4755</idno>
<idno type="eISSN">1878-4763</idno>
<imprint>
<publisher>BRILL</publisher>
<pubPlace>The Netherlands</pubPlace>
<date type="published" when="2012">2012</date>
<biblScope unit="volume">25</biblScope>
<biblScope unit="issue">3-4</biblScope>
<biblScope unit="page" from="351">351</biblScope>
<biblScope unit="page" to="364">364</biblScope>
</imprint>
<idno type="ISSN">1878-4755</idno>
</series>
<idno type="istex">AFAF5DD994945AA8FECF4AE82311F74BB899DEAC</idno>
<idno type="DOI">10.1163/187847612X629937</idno>
<idno type="href">18784763_025_03-04_s007_text.pdf</idno>
</biblStruct>
</sourceDesc>
<seriesStmt>
<idno type="ISSN">1878-4755</idno>
</seriesStmt>
</fileDesc>
<profileDesc>
<textClass></textClass>
<langUsage>
<language ident="en">en</language>
</langUsage>
</profileDesc>
</teiHeader>
</TEI>
<istex>
<corpusName>brill-journals</corpusName>
<author>
<json:item>
<name>Hong Z. Tan</name>
</json:item>
<json:item>
<name>Kwangtaek Kim</name>
</json:item>
<json:item>
<name>Mauro Barni</name>
</json:item>
<json:item>
<name>Domenico Prattichizzo</name>
</json:item>
</author>
<subject>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>3D visuohaptic watermarking</value>
</json:item>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>Visuohaptic perception</value>
</json:item>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>shape perception</value>
</json:item>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>discrimination</value>
</json:item>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>gross shape</value>
</json:item>
<json:item>
<lang>
<json:string>eng</json:string>
</lang>
<value>psychophysics</value>
</json:item>
</subject>
<language>
<json:string>eng</json:string>
</language>
<qualityIndicators>
<score>5.344</score>
<pdfVersion>1.4</pdfVersion>
<pdfPageSize>453.543 x 680.315 pts</pdfPageSize>
<refBibsNative>false</refBibsNative>
<keywordCount>6</keywordCount>
<abstractCharCount>0</abstractCharCount>
<pdfWordCount>4832</pdfWordCount>
<pdfCharCount>27527</pdfCharCount>
<pdfPageCount>14</pdfPageCount>
<abstractWordCount>1</abstractWordCount>
</qualityIndicators>
<title>Visuohaptic Discrimination of 3D Gross Shape</title>
<genre.original>
<json:string>research-article</json:string>
</genre.original>
<genre>
<json:string>research-article</json:string>
</genre>
<host>
<volume>25</volume>
<pages>
<last>364</last>
<first>351</first>
</pages>
<issn>
<json:string>1878-4755</json:string>
</issn>
<issue>3-4</issue>
<genre>
<json:string>Journal</json:string>
</genre>
<language>
<json:string>unknown</json:string>
</language>
<eissn>
<json:string>1878-4763</json:string>
</eissn>
<title>Seeing and Perceiving</title>
</host>
<publicationDate>2012</publicationDate>
<doi>
<json:string>10.1163/187847612X629937</json:string>
</doi>
<id>AFAF5DD994945AA8FECF4AE82311F74BB899DEAC</id>
<score>1</score>
<fulltext>
<json:item>
<original>true</original>
<mimetype>application/pdf</mimetype>
<extension>pdf</extension>
<uri>https://api.istex.fr/document/AFAF5DD994945AA8FECF4AE82311F74BB899DEAC/fulltext/pdf</uri>
</json:item>
<json:item>
<original>false</original>
<mimetype>application/zip</mimetype>
<extension>zip</extension>
<uri>https://api.istex.fr/document/AFAF5DD994945AA8FECF4AE82311F74BB899DEAC/fulltext/zip</uri>
</json:item>
<json:item>
<original>false</original>
<mimetype>text/plain</mimetype>
<extension>txt</extension>
<uri>https://api.istex.fr/document/AFAF5DD994945AA8FECF4AE82311F74BB899DEAC/fulltext/txt</uri>
</json:item>
<istex:fulltextTEI uri="https://api.istex.fr/document/AFAF5DD994945AA8FECF4AE82311F74BB899DEAC/fulltext/tei">
<teiHeader>
<fileDesc>
<titleStmt>
<title level="a">Visuohaptic Discrimination of 3D Gross Shape</title>
<respStmt xml:id="ISTEX-API" resp="Références bibliographiques récupérées via GROBID" name="ISTEX-API (INIST-CNRS)"></respStmt>
</titleStmt>
<publicationStmt>
<authority>ISTEX</authority>
<publisher>BRILL</publisher>
<pubPlace>The Netherlands</pubPlace>
<availability>
<p>BRILL Journals</p>
</availability>
<date>2012</date>
</publicationStmt>
<sourceDesc>
<biblStruct type="inbook">
<analytic>
<title level="a">Visuohaptic Discrimination of 3D Gross Shape</title>
<author>
<persName>
<forename type="first">Hong Z.</forename>
<surname>Tan</surname>
</persName>
</author>
<author>
<persName>
<forename type="first">Kwangtaek</forename>
<surname>Kim</surname>
</persName>
</author>
<author>
<persName>
<forename type="first">Mauro</forename>
<surname>Barni</surname>
</persName>
</author>
<author>
<persName>
<forename type="first">Domenico</forename>
<surname>Prattichizzo</surname>
</persName>
</author>
</analytic>
<monogr>
<title level="j">Seeing and Perceiving</title>
<title level="j" type="sub">A Journal of Multisensory Science</title>
<title level="j" type="abbrev">SP</title>
<idno type="pISSN">1878-4755</idno>
<idno type="eISSN">1878-4763</idno>
<imprint>
<publisher>BRILL</publisher>
<pubPlace>The Netherlands</pubPlace>
<date type="published" when="2012"></date>
<biblScope unit="volume">25</biblScope>
<biblScope unit="issue">3-4</biblScope>
<biblScope unit="page" from="351">351</biblScope>
<biblScope unit="page" to="364">364</biblScope>
</imprint>
</monogr>
<idno type="istex">AFAF5DD994945AA8FECF4AE82311F74BB899DEAC</idno>
<idno type="DOI">10.1163/187847612X629937</idno>
<idno type="href">18784763_025_03-04_s007_text.pdf</idno>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc>
<creation>
<date>2012</date>
</creation>
<langUsage>
<language ident="en">en</language>
</langUsage>
<textClass>
<keywords scheme="keyword">
<list>
<head>Keywords</head>
<item>
<term>3D visuohaptic watermarking</term>
</item>
<item>
<term>Visuohaptic perception</term>
</item>
<item>
<term>shape perception</term>
</item>
<item>
<term>discrimination</term>
</item>
<item>
<term>gross shape</term>
</item>
<item>
<term>psychophysics</term>
</item>
</list>
</keywords>
</textClass>
</profileDesc>
<revisionDesc>
<change when="2012">Created</change>
<change when="2012">Published</change>
<change xml:id="refBibs-istex" who="#ISTEX-API" when="2016-3-13">References added</change>
</revisionDesc>
</teiHeader>
</istex:fulltextTEI>
</fulltext>
<metadata>
<istex:metadataXml wicri:clean="corpus brill-journals not found" wicri:toSee="no header">
<istex:xmlDeclaration>version="1.0" encoding="UTF-8"</istex:xmlDeclaration>
<istex:docType PUBLIC="-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" URI="http://dtd.nlm.nih.gov/publishing/2.3/journalpublishing.dtd" name="istex:docType"></istex:docType>
<istex:document>
<article article-type="research-article" dtd-version="2.3">
<front>
<journal-meta>
<journal-id journal-id-type="e-issn">18784763</journal-id>
<journal-title>Seeing and Perceiving</journal-title>
<journal-subtitle>A Journal of Multisensory Science</journal-subtitle>
<abbrev-journal-title>SP</abbrev-journal-title>
<issn pub-type="ppub">1878-4755</issn>
<issn pub-type="epub">1878-4763</issn>
<publisher>
<publisher-name>BRILL</publisher-name>
<publisher-loc>The Netherlands</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.1163/187847612X629937</article-id>
<title-group>
<article-title>Visuohaptic Discrimination of 3D Gross Shape</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Tan</surname>
<given-names>Hong Z.</given-names>
</name>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Kim</surname>
<given-names>Kwangtaek</given-names>
</name>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Barni</surname>
<given-names>Mauro</given-names>
</name>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Prattichizzo</surname>
<given-names>Domenico</given-names>
</name>
</contrib>
</contrib-group>
<pub-date pub-type="epub">
<year>2012</year>
</pub-date>
<volume>25</volume>
<issue>3-4</issue>
<fpage>351</fpage>
<lpage>364</lpage>
<permissions>
<copyright-statement>© Koninklijke Brill NV, Leiden, The Netherlands</copyright-statement>
<copyright-holder>Koninklijke Brill NV, Leiden, The Netherlands</copyright-holder>
</permissions>
<self-uri content-type="pdf" xlink:href="18784763_025_03-04_s007_text.pdf"></self-uri>
<kwd-group>
<kwd>3D visuohaptic watermarking</kwd>
<kwd>Visuohaptic perception</kwd>
<kwd>shape perception</kwd>
<kwd>discrimination</kwd>
<kwd>gross shape</kwd>
<kwd>psychophysics</kwd>
</kwd-group>
<custom-meta-wrap>
<custom-meta>
<meta-name>version</meta-name>
<meta-value>header</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<p>Seeing and Perceiving 25 (2012) 351–364 brill.nl/sp Visuohaptic Discrimination of 3D Gross Shape Kwangtaek Kim 1 , ∗ , Mauro Barni 2 , Domenico Prattichizzo 2 , 3 and Hong Z. Tan 1 1 Haptic Interface Research Lab, School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN, USA 2 Department of Information Engineering, University of Siena, Siena, Italy 3 Department of Advanced Robotics, Istituto Italiano di Technologia, Genova, Italy Received 30 August 2010; accepted 22 December 2011 Abstract Human sensitivity to 3D gross shape changes was measured for the visual and haptic sensory channels. Three volume-invariant affine transformations were defined: compressing, shearing and stretching. Partici- pants discriminated a reference 3D object (cube or sphere) from its deformed shape under three experimental conditions: visual only (on a computer monitor), haptic only (through a point-contact force-feedback device) and visuohaptic simulations. The results indicate that vision is more sensitive to gross shape changes than point-based touch, and that vision dominated in the visuohaptic condition. In the haptic alone condition, thresholds were higher for shearing and stretching than for compressing. Thresholds were otherwise sim- ilar for the three transformations in the vision only or visuohaptic conditions. These trends were similar for the two shapes tested. A second experiment, conducted under similar conditions but preventing partic- ipants from manipulating object orientations, verified that the main conclusion of our research still holds when visual inspection can rely only on a single perspective view of the object. Our earlier studies on 3D visuohaptic watermarking showed that the haptic channel is more sensitive to surface texture and rough- ness changes than vision. The thresholds from the present and our earlier studies can potentially be used as the upper limits for selecting watermark strengths in order to ensure watermark imperceptibility in a 3D visuohaptic watermarking system. © Koninklijke Brill NV, Leiden, 2012 Keywords Visuohaptic perception, shape perception, gross shape, discrimination, psychophysics, 3D visuohaptic wa- termarking 1. Introduction Perception of 3D shapes through 2D retinal projections is a remarkable ability of human vision. From a computational point of view, the recognition of 3D scenes * To whom correspondence should be addressed. E-mail: samuelkim91@gmail.com © Koninklijke Brill NV, Leiden, 2012 DOI:10.1163/187847612X629937</p>
<p>352 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 from 2D images is ill posed in the sense that the solution is not unique and prone to noise perturbation (see Pizlo, 2001, for a review). Scientists have studied the nature of shape perception for about a thousand years, yet it remains an active research area in many fields including psychology, neuroscience, physics, computer science and engineering (see Pizlo, 2008). As an initial study of human visual perception, Alhazen (1083/1989) developed a theory explaining the process of vision by rays of light reaching the eye from points on an object. He also defined shape constancy to refer to the fact that an object keeps its perceived shape despite changes in its retinal projections (see Pizlo and Stevenson, 1999, for a review on shape constancy). A subsequent research area is the sources of information needed to resolve ambiguity associate with 3D shapes. Examples include variations of image intensity or shading, surface contours, color, size, and line drawings that represent edges and vertices of 3D objects (Todd, 2004). Empirically, psychophysical experiments were conducted on shape perception with various cues such as depth, curvature, length from simple lines, sphere and cube and complex objects (Hecht et al. , 1999; Kleffner and Ramachandran, 1992; Koen- derink et al. , 1996, 1997, 2000; Norman et al. , 1996). Compared to visual shape perception, haptic shape perception involves the com- bination of information from the cutaneous stimulation of skin surface and kines- thetic sense of joint positions. Although the haptic channel does not suffer from many of the visual ambiguities caused by projective transformations, there also ex- ist many illusions in haptic shape perception (Hayward, 2008). Nonetheless, studies by Klatzky et al. (1985) reported that haptic object recognition can be both rapid and accurate. Many scientists have investigated how the brain integrates inputs from individual receptors to form the percept of global shapes through the haptic chan- nel (Reed et al. , 2004; Zangaladze et al. , 1999). Others have studied multimodal shape perception and found that vision generally dominates perception when infor- mation from two modalities (e.g., vision and touch) are in conflict (Rock and Victor, 1964), but progressively more weight is given to haptics when visual information becomes less reliable (Ernst et al. , 2002; Helbig and Ernst, 2007). Moreover, the integrated visuohaptic estimates are more reliable than those from either modal- ity alone. These findings should be carefully considered when shape perception through both vision and touch is investigated. Our interest in visuohaptic perception of 3D object shapes originated from an ongoing research program on watermarking of 3D objects. As haptic technology matures, it is becoming easier to see and touch 3D virtual objects at the same time. With the development of 3D digital media suitable for both visual and haptic ren- dering, the need will soon arise to protect 3D visuohaptic contents from misuse. An important requirement of any 3D visuohaptic watermarking scheme is the imper- ceptibility of embedded watermarks. Previous studies of visuohaptic watermarking have investigated the perceptibility of additive noise embedded on object surfaces (Formaglio et al. 2006; Prattichizzo et al. , 2005, 2007). This kind of modification to 3D meshes is akin to modifying the surface details of 3D objects. The results of the</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 353 earlier studies suggest that touch is more sensitive than vision to changes in surface roughness due to watermarking. It is arguable that touch may not remain more sensitive than vision if watermarks are embedded in such a way that only the gross shape of the object is changed, since vision is better at global perception than touch. The goal of the present study is therefore to investigate the perceptibility of gross shape deformation of 3D ob- jects through vision and touch. The results will inform the design of visuohaptic watermarking systems that embed watermarks in both the object shape and surface details of 3D objects. In the present study, three types of deformations (compressing, shearing, and stretching) of 3D gross shapes were designed and implemented in a PC environ- ment. Two psychophysical experiments were conducted to estimate the discrimina- tion thresholds for 3D gross shape in Euclidean space using two simulated basic shapes (cube and sphere) over three conditions (vision alone, haptics alone and vi- suohaptic). In the first experiment, participants were able to manipulate an object’s orientation with the computer mouse and to reset the object’s orientation by press- ing the ‘R’ key on the keyboard. It could be argued that some participants may orient objects in such a way as to reduce a 3D shape discrimination task to a 2D length discrimination task. Therefore, in the second experiment, the object’s orientation was fixed throughout the entire trial. In both experiments, the initial orientation of each object presentation was randomized. In Section 2, we describe 3D gross shape deformation and its simulation in a virtual environment, followed by the design of the psychophysical experiment. Thresholds of the two experiments are reported in Section 3. Concluding remarks are drawn in Section 4. 2. General Methods 2.1. Gross Shape Deformation We use the term gross shape to refer to the global shape of a 3D object in Euclidean space, regardless of the surface details such as color or texture. In this section, we describe three linear deformations we had developed to manipulate the 3D gross shapes of cubes and spheres used in the present study. Among possible deformations of 3D object shapes, we focused on the more com- mon types of compressing, shearing and stretching, which can be simulated by affine transformations. They are defined as follows: • compressing makes the gross shape of an object more compact by pressing in one direction; • shearing deforms the gross shape in which parallel planes remain parallel but are shifted in a direction parallel to themselves;</p>
<p>354 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 (a) (b) (c) Figure 1. Three geometrical deformations: (a) compressing, (b) shearing and (c) stretching. F denotes an external force and α, β and γ denote side lengths. • stretching makes the gross shape of an object longer along the stretched di- rection: the three deformations are illustrated in Fig. 1 using the cube as an example. Compression and stretching are similar in that the two external forces are applied along the same axis, but the direction of the force causes either flat- tening or elongation of the cube along the axis, respectively (see Fig. 1(a) and (c)). Shearing, however, causes an angular change such that the resulting 3D object is no longer a rectangular prism (see Fig. 1(b)). Mathematically, the 3D linear deformations can be formulated as S new = T (S old ) , where S old and S new are polygonal surfaces before and after deformation, and T denotes a linear function. T can be set up as T 4 × 4 , a four-by-four matrix for homogeneous coordinates, x , y , z , 1 of a 3D point that is an affine transforma- tion matrix without translation. An analysis of the matrix T 4 × 4 , parameterized as in equation (1), concluded that parameters a , e and i play a role in compressing and stretching, while the other parameters ( b , c , d , f , g and h ) cause shearing of 3D shapes. Additionally, all entries in the matrix can be minimized since they are symmetric with respect to an axis ( x , y or z ). Therefore, T 4 × 4 can be further simpli- fied to form T 1 for compressing and stretching, and T 2 for shearing as shown below where k 1 and k 2 denote any constant (real number). T 4 × 4 = ⎡ ⎢ ⎣ a b c 0 d e f 0 g h i 0 0 0 0 1 ⎤ ⎥ ⎦ , T 1 = ⎡ ⎢ ⎣ a 0 0 0 0 k 1 0 0 0 0 k 2 0 0 0 0 1 ⎤ ⎥ ⎦ and (1)</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 355 T 2 = ⎡ ⎢ ⎣ k 1 b 0 0 0 k 2 0 0 0 0 k 2 0 0 0 0 1 ⎤ ⎥ ⎦ . Of the matrices derived above, transformations based on T 1 does not guarantee volume-invariance; i.e., the 3D object’s volume may change after the deformation. For the present study, we are interested in 3D shape transformations that preserve the object’s 3D volume. This is similar to kneeling a play-dough to change its shape but not the volume. This is accomplished by deriving two more matrices, T compress and T stretch , from the matrix T 1 as shown below. The matrix T 2 , which preserves 3D volumes, is rewritten below as T shear with γ as the parameter, indicating that the original values on the x axis are changed with respect to the y plane. The three matrices were used to generate the 3D objects used in the present study by com- pressing, shearing, or stretching cubes and spheres. T compress = ⎡ ⎢ ⎣ γ 0 0 0 0 1 /γ 2 0 0 0 0 1 /γ 0 0 0 0 1 ⎤ ⎥ ⎦ , T stretch = ⎡ ⎢ ⎣ γ 0 0 0 0 1 / √ γ 0 0 0 0 1 / √ γ 0 0 0 0 1 ⎤ ⎥ ⎦ , (2) T shear = ⎡ ⎢ ⎣ 1 γ 0 0 0 1 0 0 0 0 1 0 0 0 0 1 ⎤ ⎥ ⎦ . 2.2. Development of Simulator For virtual simulations of gross shape changes, a simulator of the three defor- mations employing the matrices in equation (2) was developed with Visual C++, Chai3D (www.chai3d.org) and OpenGL libraries on a PC. Gouraud shading tech- nique (Gouraud, 1971) was used for visual rendering. For haptic rendering, Rus- pini’s finger proxy rendering method (Ruspini et al. , 1997) built into Chai3D li- braries was used. Figure 2 shows the effects of the three deformations on the gross shape of a Bunny model by transforming the bounding rectangular prism. Any 3D shapes can be deformed by the three linear transformations introduced above by applying the transformation matrices to the bounding object. In the present study, cubes and spheres were used as the base reference objects for 3D shape dis- crimination as they can serve as the simplest bounding shapes for any 3D objects. 2.3. Participants A total of twenty participants, divided into two groups, took part in the present study (Experiment I: 5 males and 5 females, age range 22–37 years old, average</p>
<p>356 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 (a) (b) (c) Figure 2. Gross shape manipulations with a Bunny model bounded by a rectangular prism. The value of γ for the original shape is 1.0 for both compressing and stretching, and 0 for shearing. (a) Com- pressing ( γ = 1 . 2), (b) Shearing ( γ = 1), (c) Stretching ( γ = 1 . 5). (This object model was not used in the present study.) Figure 3. The ministick. This figure is published in colour in the online version. age 28.2 years old; Experiment II: 5 males and 5 females, age range 20–36 years old, average age 29.7). All but one participant were right-handed by self-report. Four participants in Experiment I and five participants in Experiment II had pre- vious experience with haptic interfaces and perception experiments. None of the participants reported any deficiencies in vision or touch. 2.4. Apparatus For haptic rendering of virtual 3D objects, a custom-designed 3-DOF (degrees of freedom) force-feedback device, the ‘ministick’ (see Fig. 3), was used for both Ex- periments I and II. The participant interacts with virtual 3D objects by moving a stylus that is magnetically connected to the end effector of the ministick. Whenever the stylus collided with the virtual object, a restoring force resisting the penetration was sent to the ministick. The participant perceived the gross shape of the virtual object in a manner that is similar to poking around a real object with the tip of a pen. The ministick was initially designed based on a parallel multi-loop mecha- nism invented by Adelstein (1998) and implemented by Traylor (2005). Detailed documentation of the ministick can be found in Traylor’s Master’s thesis (Traylor, 2005). The ministick has a usable, interference-free, and bowl-shaped hemispheri-</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 357 Figure 4. Examples of undeformed cube and sphere used as the base shapes in the present study. cal workspace measuring approximately 9 × 9 × 6 cm. It is capable of an update of rate of 3.8 kHz, has a position resolution of ≈ 1.5 µm and a velocity resolution of ∼ 3 mm/s at the center of its workspace. The ministick produces a stable 8 N/mm stiffness at its typical update rate of 2 kHz. A standard TFT LCD 19 ′′ PC monitor (1280 by 1024 pixels) was used for visual rendering of virtual 3D objects. A keyboard was used by the participants to enter the responses. 2.5. Stimuli Two reference 3D shapes, a cube and a sphere, were created with 3DS Max software (see Fig. 4) for both Experiments I and II. The undeformed shapes were presented as the reference stimuli to the participants through the visual and/or haptic interfaces. The test stimuli consisted of compressed, sheared, or stretched cubes or spheres with the amount of deformation controlled by γ (see equation (2)). Specifically, γ test = γ ref + δ, (3) where γ test denotes the value of γ in T compress , T stretch or T shear after deformation, γ ref the value of γ before deformation, and δ the change in γ for deformation. During the experiment, δ was increased or decreased from trial to trial based on the participants’ responses. The values for the reference stimulus ( γ ref ) were 1.1 for the T compress and T stretch matrices and 0.1 for the T shear matrix. 2.6. Procedures A three-interval forced choice (3IFC) one-up three-down adaptive procedure (Levitt, 1971) was used for both Experiments I and II to estimate the discrimination thresholds of gross shape changes. Participants were tested with three conditions: vision alone (V), haptics alone (H) or both (VH). On each trial, for the V (or H) con- dition, the participants looked at (or touched) the sphere or cube presented through the monitor (or the ministick). For the VH condition, the participants looked at and touched the sphere or cube by using both the monitor and the ministick as shown in Fig. 5. Two of the three stimuli contained the non-deformed reference cube or sphere. The remaining randomly-selected stimulus was a deformed object with the amount of deformation specified by δ . The participant’s task was to indicate which of the three objects looked and/or felt different from the other two. The initial ori- entation of each object was randomly chosen on each trial for both Experiments I</p>
<p>358 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 Figure 5. Experimental setup. and II. There was only one difference between Experiments I and II concerning the manipulation of the object orientations. In Experiment I, the participant was able to rotate the stimulus at any time with a computer mouse and the object’s orienta- tion could be reset to its default position as seen in Fig. 1 by pressing the ‘R’ key on the computer keyboard. By allowing the participants to manipulate and reset object orientations for both visually and haptically rendered objects, we removed the difficulty in object shape perception due to a particular viewing and/or feeling angle. However, this could potentially allow the participants to align the axis of shape change with the computer monitor so that the 3D object shape discrimination task was accomplished by 2D length discrimination. In Experiment II, no option was given for the participant to manipulate the object orientation. The participant had to perform 3D shape discrimination based on the object orientation that was randomized at the beginning of each interval. According to the one-up three-down adaptive rule, the value of δ was increased after a single incorrect response and decreased after three successive correct re- sponses; otherwise, the value of δ remained the same. The initial δ value was chosen to be large enough so that the gross shape change was clearly perceptible to the par- ticipant. The value of δ then decreased or increased by 6 dB, depending on the participant’s responses. After the initial three reversals (a reversal occurred when the value of δ decreased after an increase, or vice versa), the value of δ changed by 2 dB. The initial larger change in δ was necessary for a fast convergence of the values, whereas the later smaller change improved the resolution of threshold es- timates. Each adaptive series was terminated after 8 reversals at the smaller step size. The participants were comfortably seated in front of a computer monitor, the ministick haptic device, and a keyboard as shown in Fig. 5. They wore headphones to block any sound from the equipment. Initial training was provided where a se- ries of stimuli were presented to familiarize the participants with the three types of deformations (compress, shear, stretch) and the three experimental conditions (V, H, VH). The training time varied from participant to participant and averaged 20 min for both Experiments I and II. Each participant was tested once per defor-</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 359 mation type, reference object shape and experimental condition, resulting in a total of eighteen adaptive series per participant. Each participant was tested over two ses- sions. It took between 3 to 4.5 h for each participant to complete either Experiment I or II. 2.7. Data Analysis For each adaptive series, thresholds were calculated from the values over the last eight reversals at the 2 dB step size. Specifically, four threshold values were esti- mated by averaging the four pairs of peak/valley values recorded during the last 8 reversals. The mean and the standard deviation of the four discrimination thresholds were then calculated. According to Levitt (1971), the resulting thresholds corre- sponded to the 79.4 perceptible point on the psychometric function. 3. Results 3.1. Experiment I Figure 6 shows the results of Experiment I. The bars represent the average thresh- olds for all participants, and the error bars show standard deviations. The general trend of thresholds over the three types of deformations and the three experimental conditions were similar for the cube (top panel) and sphere (bottom panel) shapes. For both cubes and spheres, the thresholds for the H condition were significantly larger than those for the V condition over all three types of deformations. It was also apparent that the thresholds for the V and VH conditions were almost identi- cal regardless of the shape or deformation type, indicating that the discrimination thresholds for the VH condition were very likely determined by those for the V con- dition. For the H condition, the thresholds for stretching and shearing were similar, which in turn were significantly larger than those for compressing. This was true for both cubes and spheres. A three-way ANOVA with the factors Condition (V, VH, H), Deformation (compress, shear, stretch) and Shape (cube, sphere) for Experiment I showed that each factor was significant (Condition: F ( 2 , 166 ) = 104 . 88, p < 0 . 0001; Deformation: F ( 2 , 166 ) = 18 . 89, p < 0 . 0001; Shape: F ( 1 , 166 ) = 8 . 03, p = 0 . 0052). The only significant interaction was between Condition and Deforma- tion ( F ( 4 , 166 ) = 11 . 28, p < 0 . 0001), indicating similar trends of thresholds for the two shapes. A posthoc Tukey test revealed two groups for Deformation (com- press vs. stretch/shear) for the H condition ( μ comp = 0 . 07947, μ shear = 0 . 27947, μ str = 0 . 23315; p < 0 . 0001) but a single group for Deformation for the V and VH conditions ( p ∼ 0 . 9836 for all pair comparisons). A Tukey test also confirmed that there was a significant difference between the V and H conditions ( μ V = 0 . 03051, μ H = 0 . 19737; p < 0 . 0001), but there was no significant difference between the V and VH conditions ( μ V = 0 . 03051, μ VH = 0 . 02826; p = 0 . 9845).</p>
<p>360 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 (a) (b) Figure 6. Average thresholds from Experiment I for (a) cubes and (b) spheres. Error bars indicate standard deviations. 3.2. Experiment II Figure 7 shows the results of the psychophysical Experiment II. The bars repre- sent the average thresholds for all participants, and the error bars show standard deviations. It is apparent that the general trends of thresholds are similar in Figs 6 and 7. The thresholds of Experiment II were generally higher than those of Exper- iment I over all three types of deformations and the three experimental conditions, especially in the H condition. A three-way ANOVA with the factors Condition (V, VH, H), Deformation (compress, shear, stretch) and Shape (cube, sphere) for Experiment II showed that each factor was significant (Condition: F ( 2 , 166 ) = 259 . 25, p < 0 . 0001 ; Deformation: F ( 2 , 166 ) = 36 . 90, p < 0 . 0001; Shape: F ( 1 , 166 ) = 39 . 08, p < 0 . 0001). The only significant interaction was between Condition and Deforma-</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 361 (a) (b) Figure 7. Average thresholds from Experiment II for (a) cubes and (b) spheres. Error bars indicate standard deviations. tion ( F ( 4 , 166 ) = 21 . 18, p < 0 . 0001), indicating similar trends of thresholds for the two shapes. A posthoc Tukey test revealed two groups for Deformation (com- press vs. stretch/shear) for the H condition ( μ comp = 0 . 14169, μ shear = 0 . 35207, μ str = 0 . 32203126; p < 0 . 0001) but a single group for Deformation for the V and VH conditions ( p ≈ 0 . 0777 for all pair comparisons). A Tukey test also con- firmed that there was a significant difference between the V and H conditions ( μ V = 0 . 05296, μ H = 0 . 27193; p < 0 . 0001), but there was no significant differ- ence between the V and VH conditions ( μ V = 0 . 053, μ VH = 0 . 0607; p = 0 . 7583). Finally, a posthoc Tukey test confirmed a statistically significant difference in the H thresholds between Experiments I and II ( μ Expt . I = 0 . 1973663, μ Expt . II = 0 . 2719286; p < 0 . 0001) but found no significant differences for the V or VH con- dition (V: μ Expt . I = 0 . 031, μ Expt . II = 0 . 05296, p = 0 . 4348; VH: μ Expt . I = 0 . 0283, μ Expt . II = 0 . 060701, p = 0 . 0830).</p>
<p>362 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 4. Discussion The present study investigated discrimination thresholds of 3D gross shapes with three deformations (compress, shear and stretch) over three conditions (vision only, haptics only and visuohaptic) using cubes and spheres. The results of Experiments I and II clearly demonstrate vision dominance in 3D gross shape discrimination in that the visuohaptic thresholds were determined by the visual thresholds. This re- sult was to be expected since vision is better at capturing global information like 3D gross shapes than haptic perception of 3D shapes via a single contact point. Interacting with a stylus that touches only one point of a 3D object at a time is not conductive to the perception of the global shape (see Lederman and Klatzky, 2004). The significant difference between the visual and haptic thresholds may be reduced by using a multiple point-of-contact haptic device (see Frisoli et al. , 2005; Lederman and Klatzky, 2004). It is apparent from Figs 6 and 7 that for the H condition, discriminating 3D gross shapes with compressing was easier than with stretching or shearing. One reason for this result may be due to the way most participants explored the 3D shapes. At the beginning of a trial, the participant would lower the stylus until it touched the top of the 3D virtual object. The participant would then start a lateral stroking motion (e.g., along the γ edge as seen in Fig. 1(a)). Since compression affected the length of the top panel most significantly along the direction of the deformation, the participant could more easily discern a change in size on top of the object. Further investigation is needed in order to fully understand why the haptic sensory modality was more sensitive to compressing than to shearing or stretching. Comparing the thresholds from Experiments I and II, there was a general trend of increased thresholds in Experiment II although only the increase of the H thresholds was statistically significant. This is likely due to the fact that it was difficult for participants to identify an object’s orientation and shape with a point-contact force- feedback device. Regardless of the significant increase in H thresholds, however, the effects of experimental conditions on gross-shape discrimination thresholds for the three deformations examined in the present study are similar for Experiments I and II, thereby eliminating the possibility that the results of Experiment I were due simply to 2D object length discrimination at specific orientations. For our intended application of 3D visuohaptic watermarking, the results of the present study show that vision is more sensitive to gross shape (low frequency contents) deformation than touch with a single point of contact, and therefore the strengths of watermarks hidden in gross shape deformations should be selected with the visual thresholds as the upper limits. In addition, since users of watermark- ing applications can freely manipulate the object orientation on their watermarking system, the thresholds of Experiment I can be used as the maximum watermark strengths that can be embedded in order to ensure imperceptibility. The results of the present study complement those from earlier studies on haptic watermarking presented in Formaglio et al. (2006) and Prattichizzo et al. (2005, 2007) showing that touch is more sensitive to distortions of surface details (high frequency con-</p>
<p>K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 363 tents) of 3D objects with additive random noise. A comprehensive 3D visuohaptic watermarking scheme can achieve watermark imperceptibility by taking advantage of the different sensitivity of vision and touch to distortions in gross shapes and surface details. Acknowledgements The first and last authors (KK and HZT) were partially supported by the US Na- tional Science Foundation under Grant no. 0836664. The second author (MB) was partially supported by the Italian Ministry of Research and Education under FIRB project no. RBIN04AC9W. The authors thank anonymous reviewers for their in- sightful comments on an earlier version of the manuscript. References Adelstein, B. (1998). Three degrees of freedom parallel mechanical linkage, US Patent No. 5816105. Alhazen (1083/1989). The Optics Books , pp. 1–3. The Warhurg Institute, London, UK. (Translation by A. I. Sahra.) Ernst, M. O., Banks, M. S., Grafton, S. T. and Sathian, K. (2002). Humans integrate visual and haptic information in a statistically optimal fashion, Nature 15 , 429–433. Formaglio, A., Belloni, S., Menegaz, G., Tan, H. Z., Prattichizzo, D. and Barni, M. (2006). Percepti- bility of digital watermarking in haptically enabled 3D meshes, in: Proc. Eurohaptics Conf. 2006 , Paris, France, pp. 407–412. Frisoli, A., Bergamasco, M., Wu, S. and Ruffaldi, E. (2005). Evaluation of multipoint contact inter- faces in haptic perception of shapes. Multi-point interaction with real and virtual objects, Springer Tracts in Advanced Robotics 18 , 177–188. Gouraud, H. (1971). Continuous shading of curved surfaces, IEEE Trans. Comput. 20 , 623–629. Hayward, V. (2008). A brief taxonomy of tactile illusions and demonstrations that can be done in a hardware store, Brain Res. Bull. 75 , 742–752. (Special Issue: Robotics and Neuroscience.) Hecht, H., Doorn, A. V. and Koenderink, J. J. (1999). Compression of visual space in natural scenes and in their photographic counterparts, Percept. Psychophys. 61 , 1269–1286. Helbig, H. and Ernst, M. (2007). Optimal integration of shape information from vision and touch, Exper. Brain Res. 179 , 595–606. Klatzky, R. L., Lederman, S. and Metzger, V. A. (1985). Identifying objects by touch: an ‘expert system’, Percept. Psychophys. 37 , 299–302. Kleffner, D. A. and Ramachandran, V. S. (1992). On the perception of shape from shading, Percept. Psychophys. 52 , 18–36. Koenderink, J. J., Kappers, A. M. L., Todd, J. T., Norman, F. and Philips, F. (1996). Surface range and attitude probing in stereoscopically presented dynamic scenes, J. Exper. Psychol.: Human Percept. Perform. 22 , 869–878. Koenderink, J., Doorn, A. V., Kappers, A. and Todd, J. (1997). The visual contour in depth, Percept. Psychophys. 59 , 828–838. Koenderink, J. J., Doorn, A. J. and Lappin, J. S. (2000). Direct measurement of the curvature of visual space, Perception 9 , 69–79. Lederman, S. J. and Klatzky, R. L. (2004). Haptic identification of common objects: effects of con- straining the manual exploration process, Percept. Psychophys. 66 , 618–628.</p>
<p>364 K. Kim et al. / Seeing and Perceiving 25 (2012) 351–364 Levitt, H. (1971). Transformed up–down methods in psychoacoustics, J. Acoust. Soc. Amer. 9 , 467– 477. Norman, J. F., Todd, J. T. and Perotti, V. J. (1996). The visual perception of 3D length, J. Exper. Psychol.: Human Percept. Perform. 22 , 173–186. Pizlo, Z. (2001). Perception viewed as an inverse problem, Vision Res. 1 , 3145–3161. Pizlo, Z. (2008). 3D Shape: Its Unique Place in Visual Perception . MIT Press, Cambridge, MA, USA. Pizlo, Z. and Stevenson, A. K. (1999). Shape constancy from novel views, Percept. Psychophys. 67 , 1299–1307. Prattichizzo, D., Barni, M., Tan, H. Z. and Choi, S. (2005). Perceptibility of haptic digital watermark- ing of virtual textures, in: Proc. 2005 World Haptics Conf., The First Joint Eurohaptics Conference and Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems , Pisa, Italy, pp. 50–55. Prattichizzo, D., Barni, M., Menegaz, G., Formaglio, A., Tan, H. Z. and Choi, S. (2007). Perceptual issues in haptic digital watermarking, IEEE Trans. Multimedia 14 , 84–91. Reed, C. L., Shoham, S. and Halgren, E. (2004). Neural substrates of tactile object recognition: an fMRI study, Human Brain Mapping 1 , 236–246. Rock, I. and Victor, J. (1964). Vision and touch: an experimentally created conflict between the two senses, Science 143 , 594–596. Ruspini, D. C., Kolarov, K. and Khatib, O. (1997). The haptic display of complex graphical environ- ments, in: Proc. 24th Ann. Conf. Computer Graphics and Interactive Techniques , pp. 345–352. ACM Press/Addison-Wesley Publishing, New York, NY, USA. Todd, J. T. (2004). The visual perception of 3D shape, Trends Cognit. Sci. 8 , 115–121. Traylor, R. (2005). Design of an ethernet enabled embedded controller for stand-alone haptic inter- faces, MS Thesis , Purdue University, School of Electrical and Computer Engineering. Zangaladze, A., Epstein, C. M., Grafton, S. T. and Sathian, K. (1999). Involvement of visual cortex in tactile discrimination of orientation, Nature 401 , 587–590.</p>
</body>
</article>
</istex:document>
</istex:metadataXml>
<mods version="3.6">
<titleInfo>
<title>Visuohaptic Discrimination of 3D Gross Shape</title>
</titleInfo>
<titleInfo type="alternative" contentType="CDATA">
<title>Visuohaptic Discrimination of 3D Gross Shape</title>
</titleInfo>
<name type="personal">
<namePart type="given">Hong Z.</namePart>
<namePart type="family">Tan</namePart>
<role>
<roleTerm type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Kwangtaek</namePart>
<namePart type="family">Kim</namePart>
<role>
<roleTerm type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mauro</namePart>
<namePart type="family">Barni</namePart>
<role>
<roleTerm type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Domenico</namePart>
<namePart type="family">Prattichizzo</namePart>
<role>
<roleTerm type="text">author</roleTerm>
</role>
</name>
<typeOfResource>text</typeOfResource>
<genre type="research-article" displayLabel="research-article"></genre>
<originInfo>
<publisher>BRILL</publisher>
<place>
<placeTerm type="text">The Netherlands</placeTerm>
</place>
<dateIssued encoding="w3cdtf">2012</dateIssued>
<dateCreated encoding="w3cdtf">2012</dateCreated>
</originInfo>
<language>
<languageTerm type="code" authority="iso639-2b">eng</languageTerm>
<languageTerm type="code" authority="rfc3066">en</languageTerm>
</language>
<physicalDescription>
<internetMediaType>text/html</internetMediaType>
</physicalDescription>
<subject>
<genre>Keywords</genre>
<topic>3D visuohaptic watermarking</topic>
<topic>Visuohaptic perception</topic>
<topic>shape perception</topic>
<topic>discrimination</topic>
<topic>gross shape</topic>
<topic>psychophysics</topic>
</subject>
<relatedItem type="host">
<titleInfo>
<title>Seeing and Perceiving</title>
<subTitle>A Journal of Multisensory Science</subTitle>
</titleInfo>
<titleInfo type="abbreviated">
<title>SP</title>
</titleInfo>
<genre type="Journal">journal</genre>
<identifier type="ISSN">1878-4755</identifier>
<identifier type="eISSN">1878-4763</identifier>
<part>
<date>2012</date>
<detail type="volume">
<caption>vol.</caption>
<number>25</number>
</detail>
<detail type="issue">
<caption>no.</caption>
<number>3-4</number>
</detail>
<extent unit="pages">
<start>351</start>
<end>364</end>
</extent>
</part>
</relatedItem>
<identifier type="istex">AFAF5DD994945AA8FECF4AE82311F74BB899DEAC</identifier>
<identifier type="DOI">10.1163/187847612X629937</identifier>
<identifier type="href">18784763_025_03-04_s007_text.pdf</identifier>
<accessCondition type="use and reproduction" contentType="copyright">© Koninklijke Brill NV, Leiden, The Netherlands</accessCondition>
<recordInfo>
<recordContentSource>BRILL Journals</recordContentSource>
<recordOrigin>Koninklijke Brill NV, Leiden, The Netherlands</recordOrigin>
</recordInfo>
</mods>
</metadata>
<serie></serie>
</istex>
</record>

Pour manipuler ce document sous Unix (Dilib)

EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/HapticV1/Data/Istex/Corpus
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 002077 | SxmlIndent | more

Ou

HfdSelect -h $EXPLOR_AREA/Data/Istex/Corpus/biblio.hfd -nk 002077 | SxmlIndent | more

Pour mettre un lien sur cette page dans le réseau Wicri

{{Explor lien
   |wiki=    Ticri/CIDE
   |area=    HapticV1
   |flux=    Istex
   |étape=   Corpus
   |type=    RBID
   |clé=     ISTEX:AFAF5DD994945AA8FECF4AE82311F74BB899DEAC
   |texte=   Visuohaptic Discrimination of 3D Gross Shape
}}

Wicri

This area was generated with Dilib version V0.6.23.
Data generation: Mon Jun 13 01:09:46 2016. Site generation: Wed Mar 6 09:54:07 2024