You are on page 1of 21

404 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO.

2, APRIL-JUNE 2020

Haptic Material Analysis and Classification Inspired


by Human Exploratory Procedures
Matti Strese , Member, IEEE, Lara Brudermueller , Jonas Kirsch , and Eckehard Steinbach , Fellow, IEEE

Abstract—We present a framework for the acquisition and and macroscopic roughness of a surface. These two tactile
parametrization of object material properties. The introduced dimensions received immense attention during almost a
acquisition device, denoted as Texplorer2, is able to extract
century, beginning with the duplex theory of roughness by
surface material properties while a human operator is performing
exploratory procedures. Using the Texplorer2, we scanned Katz [4], the research studies and survey of Lederman et al. [5],
184 material classes which we labeled according to biological, Klatzky et al. [6], Bensma€ıa et al. [7], [8], and Bergmann Tiest
chemical, and geological naming conventions. Based on these real et al. [9], [10]. In parallel, auditory perception during finger-
material recordings, we introduce a novel set of mathematical surface interaction is linked to the perception of roughness and
features which align with corresponding material properties is cognitively fused into an overall understanding of the touched
defined in perceptual studies from related work and classify the
materials using common machine learning techniques. Validation material according to Schroeder et al. [11] and Yau et al. [12].
results of the proposed multi-modal features lead to an overall Altogether, these EPs lead to a set of haptic material proper-
classification accuracy of 90.2%  1.2% and an F1 score of 0.90  ties which we intend to capture by manually recording object
0.01 using the random forest classifier. For the sake of surface interaction data using a novel human-operated object
comparison, a deep neural network is trained and tested on scanner, denoted as Texplorer2. Data traces resulting from
images of the material surfaces; it outperforms (90.7%  1.0%)
these scans provide the input for supervised material classifica-
the hand-crafted feature-based approach yet leads to more critical
misclassifications in terms of the proposed taxonomy. tion based on multi-modal features, e.g., content-based tactile
features. Additionally, such features also enable the potential
Index Terms—Surface Haptics, material scanning, content-
based features. reproduction, i.e. haptic display, of the material properties.

A. Tactile Dimensions
I. INTRODUCTION
There is still an active discussion about the feature space

T HE human haptic perception system relies on both tactile


and kinesthetic sensory information when interacting with
objects. Lederman et al. [1] summarize that humans typically
describing all relevant tactile experiences [13]. By summarizing
previous studies, Okamoto et al. [14] identified the five major
tactile dimensions: friction, hardness, warmth (thermal proper-
perform six exploratory procedures (EPs) to identify unknown ties), macroscopic, and microscopic roughness. Further sub-
objects: 1) enclosure reveals global shape information about the dimensions have already been inferred by Fishel et al. (Syn-
object which is also linked to its visual inspection. The relation Touch, USA) [15], who propose a 15-dimensional feature
between objects’ visual and haptic representations is an impor- space [15], [16]. Their definitions allow for a further subtle dis-
tant issue in multi-sensory processing according to the insights tinction between these five major dimensions by separating, for
of Lacey et al. [2] and has also been confirmed by Vardar example, hardness into compliance, yielding, relaxation, damp-
et al. [3] for fingertip interaction data and visual textures. ing, and local deformation around the sensing device. To avoid
Humans then 2) lift objects to identify their weights. In parallel, potential confusion with the features defined by Fishel
3) static touch is used to identify thermal properties; 4) pressing et al. [16], we adapt their naming convention with specific
on the materials further enables assessment of the hardness. changes to credit the original thoughts about tactile features. We
5) Contour following and 6) lateral sliding motions allow users capitalize the first letter (which indicates the five major tactile
to identify the contour, and also to let them feel the fine dimension from [14]: Friction F, Microscopic Roughness Mi,
roughness and the object surface’s friction properties. During Macroscopic Roughness Ma, Compliance C, Thermal Conduc-
this procedure, humans can distinguish between microscopic tivity T), and add TUM as superscript. For example, our feature
addressing the SynTouch-inspired dimension of compliance
Manuscript received February 5, 2019; revised October 1, 2019; accepted cCP is denoted as CCPTUM .
October 29, 2019. Date of publication November 8, 2019; date of current ver-
sion June 8, 2020. This paper was recommended for publication by Associate
Editor A. M. L. Kappers and Editor L. Jones upon evaluation of the reviewers’ B. Material Scanners
comments. (Corresponding author: Matti Strese.)
The authors are with the Chair of Media Technology, Technical University Kuchenbecker et al. [17] defined the term haptography,
of Munich9184, Munich 80333, Germany (e-mail: matti.strese@tum.de; lara. which describes the process of recording and displaying all hap-
brudermueller@tum.de; jonas.kirsch@tum.de; eckehard.steinbach@tum.de). tic properties of an object similar to capturing and showing an
This paper has supplementary downloadable material available at http://
ieeexplore.ieee.org, provided by the authors. image in photography. Unlike to the conditions faced in photog-
Digital Object Identifier 10.1109/TOH.2019.2952118 raphy, haptography data acquisition is complicated by the
1939-1412 ß 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 405

necessary physical interaction with an object, and hence, is extracts relevant haptic properties to perform the task of object
influenced by active scan-time parameters. For example, recognition. The robot applied a series of four of the six EPs
Romano et al. [18] have shown that acceleration signals result- (except lifting and contour following) on each of the 60 objects
ing from tool–surface interaction heavily depend on the scan- used in the experiments. They computed the mean and the maxi-
ning speed and force. Current technical systems that attempt mum value of the low-frequency signals measured with the
to solve this challenging task mainly concentrate on the BioTac and converted the recorded high-frequency signals into
acquisition and display of vibro-tactile signals for roughness a non-normalized energy spectral density (ESD). The total
acquisition and display using acceleration sensors during tool- energy of the ESD curve, the spectral centroid, the variance, the
mediated material interactions as investigated by Kuchenbecker skewness, and the kurtosis transform the ESD into single-valued
et al. [17] and Culbertson et al. [19], [20]. These vibro-tactile features. Kaboli et al. [25] extended this idea using a Shadow
signals are also used to recognize material surfaces using Hand equipped with five BioTac sensors to perform classifica-
robots [21]–[25] and during human freehand movements by tion of 120 objects with an accuracy of 100%. However, it is not
Burka et al. [26] and in our previous work [27], [28]. Beyond clear which of the presented material samples belonged to one
using tool-mediated accelerometer scanning setups, other meth- and the same material class. The ambiguous class names, such
odologies measure the vibro-tactile signal propagation through as carpets (textures 58–78), do not allow for a generalization of
the human skin as shown by Sano et al. [29], Vardar et al. [30] the material identification. It is also not clear whether such basic
and Visell et al. [31], or, are based on tribometry scans as for features are inherently descriptive enough to sufficiently repre-
instance conducted by Colgate et al. [32]. The vibro-tactile sent the material samples in a haptic context.
sensing capabilities can even exceed human performance of We argue that both approaches have their advantages and pro-
material identification, as shown by Fishel et al. [22] using the vide solutions for different applications. For example, the Bio-
BioTac sensor. Tac Toccare might be a good choice to scan planar material
Acquiring and modeling other relevant tactile dimensions samples for industrial or commercial applications, whereas
has also been studied extensively in the past. For example, ste- hand-held operator systems perform well as mobile and versatile
reoscopy-based approaches which determine the intensity of haptic property scanners. Generally, robotic systems can reliably
surface structures have been presented recently by Choi reproduce desired recording procedures and generate more sta-
et al. [33]. Comparably, Dulik et al. [34] use infrared light tionary material interaction signals on planar material samples.
during non-contact scans to infer the surface height profile. By contrast, a human operator can easily adapt the scan proce-
Thermal conductivities can be measured using thermistors or, dure to complex object geometries and freely scan any material
as shown by Aujeszky et al. [35] or Choi et al. [36], using a sample without prior preparation of the scanned object. Humans
thermal camera based on nondestructive infrared recordings. can evaluate much faster where and how a scan should be per-
Low-frequency hardness is usually modeled as spring stiffness formed, especially for very complex object geometries. Sec-
according to Hooke’s law as done by Basdogan et al. [37] and ondly, robotic solutions require great effort to reduce the motor
can be measured using normal force measurements that are vibrations which affect the accelerometer recordings during the
divided by the indentation depth. High-frequency hardness is robot-surface interaction. The cost and effort to set up and run
related to tapping responses measured with accelerometers such a system are significantly higher than a human-operated
and has been extensively examined by Okamura et al. [38] scanner. This point might be leveraged by cost scalability in the
and Kuchenbecker et al. [39]. future. To the best of our knowledge, the BioTac Toccare system
Two different approaches exist to capture all the aforemen- currently costs about $3  105 and the hand-held operator-
tioned tactile dimensions. A material scanner can be wielded wielded device that we propose in this article about $3  103 .
either by a human operator, or by a robotic scanning system
which precisely controls its interaction parameters.
C. Contributions
A human operator-based approach reveals a specific vari-
ability in these scan-time parameters. For example, a sensor- We envision a mobile human-operated material scanner to
ized tool generally is moved within a speed range of about capture multi-modal haptic data that is associated to meaningful
20 mm/s up to 240 mm/s, as shown by Culbertson et al. [19]. labels of real world material samples. We present mathematical
The application of different sensing domains and the design of formulations of tactile features which were conceptualized and
robust features as shown in our previous work [27], [28] can partly summarized by other experts in the field to allow for suc-
mitigate the dependencies of these parameters. cessful material classification in combination with audio-visual
Robotic scanning setups, as presented by Sinapov et al. [40], features and to provide parameters possibly used for tactile dis-
Jamali et al. [21], Fishel et al. [22] or Hoelscher et al. [24] have play. The following list summarizes the main contributions of
the advantage of precise and controllable movement procedures this paper.
on predefined material sample slices. The current state-of-the-  We present a taxonomy based on geological, biological,
art scanning platform is the BioTac Toccare [15], which allows and chemical naming conventions and provide a novel
for a comprehensive scan of a predefined material sample to database (LMT Haptic Material Database) consisting
assess its tactile properties and can be considered as the gold of 184 material classes. The collected images and data
standard for automated planar tactile sensing. Beyond two- traces have a total file size of about 80 GB.
dimensional surface scans, Chu et al. [41] have shown that a  We present a mobile material scanning setup
PR2 robot equipped with two BioTac sensors successfully (Texplorer2, two units are shown in Fig. 1), which

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
406 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

Fig. 2. Granite is an igneous (S1 ) stone (C6 ) which is either polished (left:
P2 ) or otherwise physically processed (right: pointed P3 ). The difference in
the surface structure heavily influences its haptic feel.

processing shapes P . For example, wood can be veneered or


naturally cut from the corresponding tree; aluminum can con-
tain differently shaped holes or gratings, and sandpaper pos-
Fig. 1. Two components of the proposed Texplorer2. sesses different grit sizes. The applied processing procedures
can lead to completely different visual and tangible perception,
extracts relevant material properties. A video describing although the actual material is the same. Another identifier X
the system is included in our website [42] and uploaded denotes class-specific attributes. For example, we need to take
as supplementary material. the moisture of woods into account, which changes their density
 We mathematically define perceptual tactile features, or metals may contain more or less percentages of their compo-
denoted as TacTUM features, which are inspired by the nents, especially steel alloys. Table I shows an overview of the
biomimetic process features from Fishel et al. [15], [16]. taxonomy with respect to the classes and sub-classes. The more
 We compare the hand-crafted feature-based results with extensive Table 6 in the appendix further shows the different
the classification results obtained with a commonly processing shapes and extra class properties. We have also
used deep neural network based on a large number of included a complete material name list on our website [42].
The taxonomy is mainly inferred from Weber et al. [45]
64,400 [500 x 500] RGB texture images from our new
and [46]. The database currently consists of 88 wooden, 24
database using the 184 defined material labels.
organic, 18 plastic, 4 ceramic, 4 glass, 28 stone, 12 metallic,
This paper is organized as follows: In Section II, we
and 6 composite material classes, in total 184. The relevance
describe our substantially novel haptic database taxonomy,
of this taxonomy is given in the following based on the exam-
and Section III provides an overview of our hardware setup.
ple C6 S1 M1 (granite) and illustrated in Fig. 2.
Section IV presents the multi-modal feature calculation, and
Considering a pure granite tile we either may face a pol-
Section V their performance in the desired task of material
ished P2 (left image in Fig. 2) or a pointed P3 (right image in
classification and further compares their performance to sur-
Fig. 2) sample, which leads to a completely different percep-
face image-based deep learning-based (DL) classification. We
tion of the material surface as well as different scan results.
conclude the paper in Section VI.

II. DATABASE AND CATEGORIZATION B. Discussion

Publicly available material databases for haptics research Due to the large amount of woods available, we currently face
(e.g., from Culbertson et al. [43], Hassan et al. [44], or our prior a class imbalance towards C1 which needs to be considered
database from our website [42]), provide a great variety of dif- whenever the classification system has to distinguish on a coarse
ferent haptic data scans. However, these databases are based on scale between, e.g., woods and other classes. However, fine level
colloquial material names (e.g., stone tile) which can lead to material classification is not affected since we have the same
ambiguities across different industrial or commercial branches. number of observations (instances) for each material class.
In addition, these databases typically contain between 100 and We actively record new material samples to extend the already
200 differently shaped surface textures which often belong to extensive database. Several materials, e.g., concrete or glass,
the same material classes, though. Consequently, we performed exist in dozens of different processing shapes and mixtures. We
an interdisciplinary survey on the chemical, biological, and introduced the identifiers P and X to take these material-specific
geological material names and defined the material class labels properties into account and will significantly increase the number
according to these observations. of differently processed materials in the database in future work.
Considering the number of different sensing domains and
required data to represent these different material appearances,
A. Taxonomy
we still face an ongoing task. That is also why we suggest to fol-
Our naming convention works as follows. C denotes the low taxonomies like the one proposed in this article to unambigu-
major material class, e.g., metals or woods. The identifier S con- ously connect haptic data traces to the exact material. Note that
siders the material sub-class, e.g., hardwood or softwood. M the number of woods goes well beyond the 10,000, for example.
denotes a specific material, e.g., lead, tin, or aluminum on a fine The number of various minerals, stones, or metal alloys and their
classification level. Many materials come along with different processing shapes is similarly excessive.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 407

TABLE I
TAXONOMY OF THE CLASSES C AND SUB-CLASSES S. ALL MATERIALS M ARE INTEGRATED INTO THIS LABELING STRUCTURE

Fig. 3. Overview of the Texplorer2 units. The operator bag and laptop tray
are not depicted.

III. TEXPLORER2 MATERIAL SCANNING SETUP


We seek to capture and parametrize material samples of the
different material classes. Consequently, a material scanning
system is required to record relevant device-surface interac-
tion data. Varying scanning parameters, such as speed and
force for contact-based approaches, or distance, rotation and
illumination conditions for non-contact approaches complicate
robust data acquisition. Fig. 4. First Texplorer2 unit for the acquisition of microscopic and macro-
scopic roughness information. Sensing components are shown in red. The
images showing the exploratory procedures are adapted from [1].

A. Texplorer2 Overview
Based on the idea of a haptic camera (haptography) from describe hardness, and thermal cooling data to describe
Kuchenbecker et al. [17], we see the necessity for an intuitive warmth. Beyond, it acts as metal and magnetic detector.
and flexible low-cost operator approach for the measurement This unit is first placed on the material sample to collect
of any material sample. We completely redesigned the static touch information, and thereafter, used to press on
Texplorer device from our previous work [28] and added new and squeeze the material sample.
components. Inspired by the design of the Proton Pack by  TU3 captures surface images, magnified surface
Burka et al. [26], we embed the scanning system into a bag to images, and images with additional illumination using a
allow for mobile scanning using a laptop, a laptop supporter/ smartphone (Apple iPhone 8) camera.
tray (Rocket Packs) and a power bank (Anker). A USB data  TU4 measures the mass and the volume to estimate the
acquisition card (NI USB-6002), attached to a Windows 7 lap- density of an object sample by placing the sample on
top Lenovo 80VR (i7-7700HQ CPU at 2,800 MHz, 16 GB top of the scale and inside the measuring jug.
RAM) collects the multi-modal sensor data. In addition to the
lower cost and simple scanning procedure, we further identify B. Texplorer2 Units and Data Processing
the advantage of possibly scanning materials of arbitrary
shape, geometry, and size. The operator performs multiple 1) Texplorer2 Unit 1: Micro- and Macroscopic Rough-
scanning steps with four mobile devices, which we denote as ness: The first Texplorer2 unit is rolled by the operator over
Texplorer2 units (TU1–4). the surface of an object as shown in Fig. 4. Three bearings
Fig. 3 shows the four main units of the Texplorer2 and the ensure that the influence of the applied force is mitigated and
following list summarizes the general operating principle and the distance between the surface and the IR sensor is constant.
the target haptic dimension of each unit.
 TU1 records acceleration and audio signals to calculate a) Vibro-Tactile Signals: We use a three-axis accelera-
microscopic roughness features as well as reflectance tion sensor ADXL335 (Adafruit) with a range of 3g to
data for macroscopic roughness definition. It is slid acquire vibro-tactile signals v, i.e., to capture microscopic
over the material sample. roughness information with a sampling rate of 3,000 Hz. We
 TU2 records normal and tangential forces to describe the use two different kinds of tool tips during separate scans: the
surface friction, pressure and folding FSR information to bare finger which propagates finger-surface interaction signals

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
408 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

to a held accelerometer, and a stainless steel tool tip mounted


directly at the accelerometer.
We capture bare finger recordings vf;t during tapping on
and vf;s during sliding over the material surface. The operator
holds the acceleration unit between the thumb and index finger
as shown in Fig. 4 (upper middle illustration). The index finger
is placed on the accelerometer and simultaneously touches the
material surface directly. Additionally, tool tip tapping vt;t
and sliding vt;s recordings were collected after the bare-finger
recordings. The operator mounts a 3D-printed stainless steel
tool tip (hemispherical tip with a diameter of 6 mm) on the
ADXL335, clips it (using Velcro tape) on the TU1, and
repeats the same scanning procedure. The majority of the
operator force is compensated by the bearings during a scan.
We collect the two versions (bare finger-based and steel tool
tip-based) of the vibro-tactile signal recording procedures for
two reasons. We need to determine which works more robustly
for material classification, and, which signals are more appro-
priate for haptic display. During our testing we observed that
stainless steel tool tip recordings work better for classification,
presumably because they amplify the tool-surface interaction.
The bare finger recordings intuitively work better for haptic
display, which still needs to be proven in future work address-
ing the display of the 15 TacTUM dimensions.

b) Audio Signals: We place an omni-directional capaci-


tive microphone (Blusmart Lavalier Clip) on the TU1 to capture
audio signals of a frequency range between 10 Hz – 7,500 Hz
using a sampling rate of 44,100 Hz both for bare finger af;s as
well as for the stainless steel tool tip at;s audio signals, which
are recorded in parallel to the vibro-tactile signals.

c) Reflectance Signals: Dulik et al. [34] worked


with infrared reflective (IR) scans to extract height-related Fig. 5. Second Texplorer2 unit for the acquisition of thermal, friction, and
material properties. We employ a comparable IR sensor hardness material information. Sensing components are shown in red. The
(Adafruit AF2349) consisting of an emitter, a detector, and a images showing the exploratory procedures are adapted from Lederman
et al. [1].
10kV resistor in a voltage divider circuit. We measure one-
dimensional reflectance signals r to characterize macro-
scopic roughness information. Note that the IR sensor has a acceleration signal components into one using the DFT321
fixed distance of 7 mm to the surface which is ensured due algorithm from Landin et al. [47] as used by Culbertson et al.
to the stable rolling of the three bearings during a slide. in [48] and all following works.
For completely flat surfaces, this distance is below the sens-
ing range, and hence, the values for flat surfaces are constant. 2) Texplorer2 Unit 2: Warmth, Friction, and Hardness:
However, if a material surface contains coarse structures or a) Thermal Signals: Thermal conductivity properties
holes, the measured distance varies accordingly. help to distinguish between different object surfaces [1].
Although surfaces conduct heat from the human skin differ-
d) Data Processing: A raw vibro-tactile scan with TU1 ently during bare-finger touch, they obtain ambient tempera-
is 5 seconds long. When the recording begins, only accelerom- ture after an indefinite time. Aujeszky et al. [35] use a laser to
eter noise is present until the operator starts moving the TU1. heat a material sample and locally measure thermal cooling
We shift a 0.1 ms frame over the raw vibro-tactile signal to using infrared sensing. We follow a comparable approach to
identify the start of the sliding motion. Compared to the signal infer thermal properties of the material samples. Our setup is
noise energy, the window frame energy rises notably during shown in the upper part of Fig. 5. We actively heat the mate-
this procedure which leads to the start index. We then extract rial surface and measure the subsequent local cooling of the
one-second-long acceleration signals as vibro-tactile data. The material surface. More specifically, we heat a Peltier element
audio data indices are the same due to the parallel recording (TES1-127021) for three seconds while pressing it on the
procedure. The vibro-tactile signals are band-limited between material surface. Subsequently, we move the unit to measure
10 Hz and 1,000 Hz to remove the effects of human hand the thermal cooling rate over the heated spot using an infrared
motion and accelerometer sensor noise. We combine the three temperature sensor (Melexis MLX90614) for ten seconds and

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 409

Fig. 7. Recorded magnetic data during static touch. Aluminum (left) does
not change the magnetic field, steel wool (middle) is paramagnetic, and cast
Fig. 6. Close-up view of the static touch part of TU2. The electric conductive iron (right) reveals significant magnetic properties (ferromagnetic). Diamag-
tissues reliably detect metals, the Peltier element and MLX90614 measure netic materials, which are currently not part of the database, would generate
thermal cooling, and an additional magnetometer unit (right) instantly detects positive values. All other materials lead to a mean value of DB ¼ 0.
the change of the external magnetic field caused by the neodymium stack
through a material.
material to identify metals. This procedure leads to an inverted
denote the recorded temperature data array as t. Note that the binary array m ~ 2 f5 V, 0 Vg, which we map to the values
required change in position can be automated in a future ver- m 2 f1; 1g. Only electrically conductive materials, e.g.,
sion of the Texplorer2. metals, pull down the DAQ input to ground (0 V) during scan
time, leading to an m ¼ 1 array. We set a binary value
b) Force Sensitive Resistor (FSR) Signals: Surface metal ¼ 1 if m > 0.
friction information is recorded using the index finger during As a minor contribution, we add a magnetometer to measure
a sliding motion of the object surface. Our second Texplorer2 the change of the magnetic field through a material sample.
unit contains two FSR (FSR400, Interlink) for normal and tan- More specifically, we apply a stationary external magnetic field
gential sliding force measurements, see Fig. 5, middle. These in a constant distance (thickness of casing) and measure the
sensors are used to determine the exerted normal force and the influence of the underlying material sample. We 3D-printed a
tangentially exerted friction force during the sliding move- casing which contains a 3-axis digital magnetometer
ment. In the following, we consider the intensity values of MAG3110 (Sparkfun) PCB and a stack of eight neodymium
these sensor measurements only; the orientation of the forces permanent magnets, shown in Fig. 6 (right). We zero-calibrate
is captured by the sensor placement (normal and tangential). the MAG3110 using these permanent magnets which expose a
We denote the force arrays f ¼ ½f1 ; f2 ; . . .fn  from a scan as strong and constant magnetic field to rule out the influence of
f0fr;n and f0fr;t , respectively. The raw FSR sensor values f0 (in the Earth’s magnetic field. The resulting data traces are denoted
volts) grow logarithmically with increasing force. We convert as DB and shown for selected materials in Fig. 7.
each voltage value f 0 in f0 to a force applying the expression Magnetism is not a tactile property, but the change of the
magnetic field characterizes several metals and their alloys.
0 1 We can reliably distinguish between ferro-, non-, and diamag-
f ¼ 9.81  m=s2  0.01 kg  ef 1:33 V (1)
netic materials within a second during static touch and thereby
according to the inverse force-voltage relation diagram (Inter- even exceed human material classification capabilities. The
link FSR integration guide), leading to the normal and friction thickness of the material sample ideally should be included in
force arrays ffr;n and ffr;t . Note that the normal FSR, which the future, still. Features based on metal or magnetic material
contacts the surface, is equipped with a 6 mm diameter PLA detection as well as thermal cooling are not influenced by
sphere. other sensing domains (such as vibro-tactile signals) and lead
The same FSR used for the normal force is also used to infer to further independent, and hence, characteristic features.
hardness properties. Additionally, a Flex Sensor (FS-L-0095-
103-ST, Spectra Symbol) collects information about the object d) Data Processing: Since the operator index finger
flexibility. For example, if a piece of paper lies flat on a table may create an offset force while touching the TU2 friction
and a robotic material scanner presses on it, the hardness of the FSR, we need to subtract the mean value of the first force sam-
underlying table is measured. However, humans can easily lift ples from the friction force signal.
the paper to fold it. In this case, the Flex Sensor will change its
angular value indicating flexibility. Note that the FSR for the
3) Texplorer2 Unit 3: Visual Properties: Visual informa-
normal force from the friction estimate is used alongside with
tion is linked to haptic perception as summarized by Lacey
the Flex Sensor during the pressing and folding procedure. This
et al. [2], and hence, contains information about relevant
step helps to distinguish between papers, leather, and fabrics,
material properties.
for example. The operator first presses on the material surface
and releases it, using the FSR for normal force acquisition. Sec-
ondly, the operator grabs the whole material, if possible, and a) Camera: The images of the materials, denoted as
tries to fold it between the thumb and index finger. The Flex Idisp , are captured using an iPhone 8 (Apple) camera with a
Sensor records the folding angle a f while the normal force FSR resolution of 3024 x 4032 pixels. To infer macroscopic details,
records the applied folding force ff in parallel. we further attach a magnifying lens (four times zoom) and
capture macro images Imacro . Note that the distance between
c) Metal and Magnet Detector Signals: As in our previ- each surface and the camera was held constant using a Styro-
ous work [28], we use two electric conductive tissues as elec- foam bumper to ensure the same recording distance and per-
trodes (see Fig. 6, left) during the static contact with a spective. We further use the camera flashlight to capture

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
410 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

are available on our website [42]. This collection is intended to


contain a broad set of different recording modalities. The data-
base shall allow other researchers to test various material classi-
fication approaches and define further relevant tactile features.
Available material archives, e.g., the one presented in [50],
indicate which materials still need to be added in the future.

D. Discussion
We do not explicitly measure the scanning parameters during
the device – surface interaction. Burka et al. [51] have shown
that variable scan-time parameters do not necessarily decrease
the classification performance whenever options exist to miti-
gate their influence by intelligent acquisition device design or
Fig. 8. Volumetric approach based on Archimedes’ principle. We use the application of robust features. Several aspects of the
different measuring jugs to infer the volume of arbitrarily-shaped objects by
displacing either water or micropearls (diameter 1 mm). The images showing Texplorer2 account for these dependencies, and we also identify
the exploratory procedures originate from Lederman et al. [1]. more options for future optimization. The bearings of TU1 com-
pensate for heavy pressure forces during the acquisition of
illuminated images Iillu which provide further information vibro-tactile signals. The fixed distance of the IR reflectance sen-
about the object’s surface reflectivity. sor to the surface also makes sliding scans robust in terms of sur-
face structure detection. The major remaining dependency is the
b) Data Processing: We remove 100 pixels from each scanning speed. Different operators may scan the surfaces differ-
border of the image to reduce the number of artifacts. ently fast, yet within a specific range of this dependency. Cul-
bertson et al.’s [19] work on tool-mediated setups shows an
4) Texplorer2 Unit 4: Mass and Volume: The material average speed of 130 mm/s which gives a good indication of
samples are placed on an electronic scale (Smart Weight, how fast humans wield such acquisition systems. We need to
range of 0.1 g – 2 kg) to measure their weight m which is accept a remaining variance in the vibro-tactile and audio signal
noted manually. The density r ¼ m V has been determined and scans, but are confident that the inclusion of other modalities,
published online for the majority of all available materials good device design, and robust features can still lead to success-
(e.g., in [49]). However, for a new query to the database, very ful classification.
complex object geometries, and undocumented densities, we Another important point to discuss is the number of modalities
still need an approach to estimate the volume V . This task is used in this article. Fig. 9 shows all currently recorded modalities
complex and an automatic robotic setup may not be able to for one material sample. Haptic material interaction is a multi-
extract the volume of complex object geometries without faceted procedure; we already employ nine different sensors
additional effort, time, and cost, whereas a human operator leading to twenty different kinds of signals. The application of
can intuitively carry out displacement techniques to estimate different sensors reduces potential correlations from subse-
the volume. For example, a set of measuring jugs can be filled quently extracted features, but poses challenges for the required
with water and the object is placed inside. The displacement is amount of data to be collected. We believe that content-based
proportional to the volume of the object. If the object could be haptic data acquisition, potentially used for display as well,
damaged by water, we use polystyrene micro-pearls instead. requires multi-modal data acquisition. Versatile material scan-
To read the measuring scale accurately and avoid uneven ners are required to cope with this challenge and potentially will
micro-pearl distribution, we place a vibration motor (Newgen add further sensors. For example, another sensing domain not
Medicals Vibration Plate) below the jugs and apply a constant yet examined on larger material databases is the electric suscep-
20 Hz vibration before each manual reading of the scale as tibility x, which is very characteristic for electrically non-con-
shown in Fig. 8. The overall displacement procedure has ductive materials.
major advantages compared to infrared or stereoscopic
approaches. First, the volume of concave objects or hole-con-
IV. FEATURES FOR MATERIAL CLASSIFICATION
taining ones can be determined, and second, the volume of
highly reflective objects can be measured. We define and use a set of tactile as well as non-tactile fea-
tures based on the multi-modal data traces introduced in the
C. Recordings following section. Note that several parameters of the features
are the result of feature optimization and were iteratively
The TUM Department of Architecture holds the largest col- determined.
lection of building materials in all German universities [45]. We
scanned most of the available material samples along with mate-
A. TacTUM Tactile Features
rials from our previous work [28] to set up the largest public hap-
tic database (to the best of our knowledge). In our current version Fishel et al. [16] propose a set of representative tactile fea-
of the database, two human operators scanned each material tures which we understand as a more detailed description of
(five times in total) and stored the data traces as. mat files which the five major tactile dimensions from Okamoto et al. [14]. To

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 411

TABLE II
MACROTEXTURE NOTATION OVERVIEW

parameter which can be made adaptive to the speed in future


work. We add the standard deviation sðvabs Þ and the mean
of vabs to obtain an adjusted array to compensate amplitude
variations for different recordings of one material sample
and calculate
ð100Þ
vth ¼ ðsðvabs Þ þ vabs Þ  1 þ vabs (2)

vabs;D ¼ vabs  uðvabs  vth Þ (3)

which is computed using the unit step function uðxÞ that


returns 1 element-wise if the corresponding array entry is
greater than 0, and 0 otherwise. We then define the vibro-tac-
tile spikiness (VS) component as

VS ¼ log 10 ð1 þ vabs;D Þ (4)

We further consider friction signal variations (FV) during a


slide as another component for MaTXTUM . Rapid changes in
the friction force signal ffr;t indicate a change in the material’s
underlying physical structure, e.g., while sliding over a coarse
or bumpy surface profile.
Fig. 9. Example data traces of one material entry (polyurethane). Overall,
nine sensors record twenty different data traces, which are the source for our ð100Þ
features. FV ¼ sðffr;t  ffr;t Þ (5)

the best of our knowledge, there is no explicit mathematical As a third component, the IR sensor’s reflectance signal
definition of these features publicly available. Hence, and as a data is related to the height profile of a material surface. Dur-
major contribution, we propose the equations for the calcula- ing a scan, variations in the reflectance signals capture the
tion of these features in the following. change of the current surface-sensor distance. We calculate
the moving average rð100Þ and consider the standard deviation
1) Macroscopic Roughness: The macrotexture (MaTXTUM ), of the difference r  rð100Þ as our reflectance variation (RV)
macrotexture coarseness (MaCOTUM ), and macrotexture regu- value
larity (MaRGTUM ) describe the macroscopic roughness features
of a material surface. RV ¼ sðr  rð100Þ Þ (6)
Finally, MaTXTUM is calculated as a linear combination
a) Macrotexture MaTXTUM: We define a surface- 
1 MaTX TUM ¼ 1  VS þ 2  FV þ 3  RV (7)
dependent macrotexture feature based on our macroscopic
roughness strength feature from [28]. Several sensing using the weighting parameters 1;2;3 which are determined
dimensions determine the physical existence of surface during feature optimization (Section V-A2) as (3, 2, 4).
structural patterns. We believe three different signals to be
appropriate for calculating this feature value. Table II shows b) Macrotexture Coarseness MaCOTUM: The existence
the auxiliary functions which we use for the MaTXTUM of visual patterns on a surface indicates whether the material
calculation. surface is coarse or not. Consequently, we apply the Harris
The first component is related to acceleration spikes in the Corner Detector (HCD) [52] to identify the n most character-
vibro-tactile signals. We define a spike detection algorithm istic image corner points in our surface images Idisp . We use
on the absolute values of the vibro-tactile sliding signals the speeded-up robust feature (SURF) extractor [53] on these
with N being the length of vabs . We first compute a 100- corner pixels to extract and sort the scale property sc in a
ð100Þ
point window simple moving average array vabs of vabs . descending order array sc (using the Matlab function
The moving average window length is an optimization detectSURFFeaturesðÞ). We consider the visual

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
412 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

component of macrotexture coarseness as the mean value of


the n largest scale factors sc1...n identified in Idisp multiplied
by the spectral spread of the image histogram (introduced as
contrast CðÞ by Tamura et al. [54]). CðIdisp Þ is hence the
image contrast of the material images; low-contrast images
are more likely to be of completely flat material surfaces,
e.g., plastics or glasses. We then multiply the result by the
mean value of the absolute vibro-tactile signal which trans- Fig. 10. Image spectrum before (left) and after (right) applying the inverted
forms this image feature into a tactile feature. Otherwise, Gaussian smoothing filter.
painted structures on flat materials potentially distort the fea-
ture value. Hence, vibro-tactile signal energy and the mean value of all pixels
in Xfilt as

2 MaCOTUM ¼ vabs  CðIdisp Þ  sc1...n (8)
with n being 10 in this paper resulting from feature optimiza-   filt  CðIdisp ÞÞ
3 MaRGTUM ¼ vabs  ðX (12)
tion (Section V-A2).
c) Macrotexture Regularity MaRGTUM: Visible texture We subtract the image contrast for two reasons. First, homoge-
regularity is related to the intensity and arrangement of the neous images with low contrast are regular; secondly, high-
image spectral components. Regular structures reveal specific contrast irregular surface images should lead to a very small
frequency peaks, whereas irregular image spectra show an MaRGTUM value. The multiplication of vabs follows a similar
incoherent noisy pattern. The real part of the image spectrum motivation as for MaCOTUM ; we need to prevent that irregu-
! larly painted structures on actually flat (homogeneous) surfa-
1 X N1
M1 X
0
X ðk; lÞ ¼ Re pffiffiffiffiffiffiffiffiffiffiffiffiffi Iðm; nÞ  e2pjðmk
M þnl
N Þ
(9) ces distort the feature value.
M  N m¼0 n¼0
2) Microscopic Roughness: The microtexture roughness
is calculated to obtain a spectral representation which is then (MiROTUM ) and microtexture coarseness (MiCOTUM ) describe
centered using Xðk; lÞ ¼ X 0 ðk  M2 ; l  N2 Þ. Thereafter, a the microscopic roughness features of a material surface.
threshold t is defined
a) Microtexture Roughness MiROTUM: The wavelet
t ¼ a  |ffl{zffl}
max Xðk; lÞ
decomposition of a vibro-tactile signal vt;s (during a slide
1kK
1lL using the stainless steel tool tip recordings) leads to approxi-
mation and detail signals which change comparably for differ-
with a being 0.75. Consequently, all values smaller than 75% ent recording speeds. Our microscopic roughness feature from
of the maximum value are removed using [28] is based on the difference of two of such different detail
 levels by calculating
Xðk; lÞ if Xðk; lÞ t
X thresh ðk; lÞ ¼ (10)
0; otherwise. 
4 MiROTUM ¼ log 10 ð1 þ dÞ (13)

The remaining peaks represent the most characteristic spatial with


frequencies of the surface image. Irregular structures have
several peaks around the center of the real part of the image D1 ðvt;s Þ
spectrum, whereas regular images reveal repetitions up to d ¼ D1 ðvt;s Þ   D5 ðvt;s Þ: (14)
larger spatial frequencies. That is why an inverted [200 x 200] D ðvt;s Þ
5

pixel Gaussian smoothing filter G0 ðk; lÞs ¼ 1:0  Gðk; lÞs ¼


k2 þl2 Dk ðvÞ contains the absolute values of the kth detail level of v.
1:0  1
e
2ps 2
with s = 25 is applied on each element in
2s 2
The used wavelet type is a Daubechies 8 (db8) wavelet result-
the spectral domain of Xthresh to underweight closer peaks to ing from the feature optimization (Section V-A2).
the center using the Hadamard product b) Microtexture Coarseness MiCOTUM: We adapt the
spectral centroid definition from Fishel et al. [22] based on the
Xfilt ¼ Xthresh
G0s : (11) real part vibro-tactile signal spectra Vt;s
Fig. 10 visualizes this step for the image spectrum of a regu- PN
larly perforated epoxy plate (C3 S4 M2 P2 X1 ).
 i¼1 f i  jVt;s ðf i Þj
Spectra of irregular surface images scatter close to 5 MiCO TUM
¼ fs  PN (15)
i¼1 jVt;s ðf i Þj
the center, whereas regular surface images can reveal
specific peaks far from the center. The threshold t and the
filter dimensions are result of the feature optimization with fs ¼ 3000 Hz being the sample rate of the accelerometer.
(Section V-A2). Finally, the Macrotexture Regularity Smaller values indicate the presence of a fine (i.e., smooth)
(MaRGTUM ) is calculated as the product of the mean material surface, e.g., a polished steel surface.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 413

Fig. 12. Texplorer Unit 2 while being pressed on a deformable material sam-
ple. The normal force FSR (in red color) is used to infer the tactile compliance.
Fig. 11. Separation of normal and tangential (friction) forces for the calcula- Note that the indentation depth is constant (thanks to the attached hemisphere
tion of tactile stiction FSTTUM and sliding resistance FRSTUM . Note that the and TU2 casing) after the operator presses D x ¼ 4.0 mm into the deformable
increase of the friction force before the normal force increase may result from, material sample.
e.g., operator finger movements inside the TU2.

3) Friction: Tactile stiction (FSTTUM ) and sliding resis- noticeably adhesive. Secondly, the adhesion properties of
tance (FRSTUM ) can be considered equivalent to the static and any sensing device also differ from those of a human fin-
dynamic friction coefficient of a material surface. ger; a material surface scanned with a biomimetic sensor
made of silicone may result in a larger adhesive tack fea-
a) Tactile Stiction FSTTUM and Sliding Resistance ture value than expected. Thirdly, an operator can intui-
TUM
FRS : As long as the normal force ffr;n is about zero, the tively assess the stickiness of a material during several
operator did not noticeably press on the material. An increas- trials. Next, any oily or sticky surface might damage the
ing normal force then indicates the beginning of the static fric- device, and the detection of such surface conditions is still
tion phase. Normal and friction force ffr;t values are then used challenging for robotic systems. Lastly, the inclusion of
to model friction based on Coulomb’s law. the operator judgment represents another variable not
Since we can not directly measure the transition from stick- depending on any other sensing domain.
ing to sliding over the surface with TU2, the ending segment
of the static phase needs to be determined following another 4) Compliance: Operator-based material scanning
approach. We observe that the friction and normal forces approaches reveal an advantage over robot-based
increase differently whenever the operator starts touching the approaches for complex object structures in terms of hard-
material in order to perform a slide. We shift a 10 ms window ness estimation. Humans generally identify the most suited
over the signals to identify the maximum positive slopes locations for grasping and folding/pinching a material sam-
Dffr;max and Dfn;max and define tactile stiction as ple. We use the recorded normal FSR and Flex sensor sig-
nals during the pressing and folding actions of the operator
Dffr;max to infer hardness-related features. We first determine if a

6 FST TUM ¼ (16) material is deformable or not with regard to reasonable
Dfn;max
human interaction forces. The folding angle values a f of
the TU2 act as tie-breaker in this binary decision. We
Fig. 11 visualizes exemplary signal traces and the correspond- evaluate the expression sða af Þ < 1o prior to any compli-
ing slopes. ance feature calculation. If an object is rigid, all five com-
The following signals after the ending of the static segment pliance features are set to zero.
describe the sliding phase until the normal force again
decreases to zero. The FSR signals in the sliding segment are a) Tactile Compliance CCPTUM: Tactile compliance
denoted as ffr;FRS and fn;FRS , respectively. Consequently, slid- refers to the spring stiffness [37] of a material sample and is
ing resistance is calculated as defined as the ratio of the required normal force fn and
ffr;FRS achieved indentation depth Dx. Thanks to the design of our

7 FRS TUM ¼ (17) Texplorer2, we are faced with a fixed Dx and measure the cor-
fn;FRS
responding (normal) force as shown in Fig. 12.
b) Adhesive Tack ATKTUM: The effort required to break The tactile compliance then is
contact with a surface [15] is proportional to the adhesive 
forces of a surface. We define it as a quintary operator-decided 0 a f Þ < 1o
if sða

9 CCP TUM ¼ fn;max (19)
value Dx else
8
>
> 0 if not adhesive with
>
> fn;max ¼ max f n;i
< 0:25 if slightly adhesive 1in
(20)

8 ATK TUM ¼ 0:5 if uncertain (18)
>
> being the maximum force and D x = 4.0 mm measured with
>
> 0:75 if adhesive
: a caliper. If the material sample is deformed up to this inde-
1 if strongly adhesivecases
ntation depth, the additional force is compensated by the
There are several reasons not to define this feature based whole TU2 casing. Note that a 3D printed PLA or rubber
on measurements from the sensing device. First, the major- (Shore hardness scale value > 75) hemisphere are applicable
ity of naturally occurring material surfaces are not in this context.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
414 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

b) Local Deformation CDFTUM: We calculate the local


deformation around the measuring device as


10 CDF TUM ¼ sða
af Þ (21)

which describes how the object can be folded during free


interaction based on the achieved folding angle a f . Fig. 13. Texplorer Unit 2 while pinching a deformable material sample. The
normal force FSR and the Flex Sensor are shown in red color. Damping and
relaxation can be defined by the ratio Dff and vf ¼ Da
af =Dt.
c) Damping CDPTUM and Relaxation CRXTUM: Damp-
ing can be understood as the ratio of the counteracting force
and the current speed (Hartog et al. [55]). We measure the material sample to identify yielding correctly, which a human
temporal change of the folding angle array a f values and the operator can anticipate thanks to human knowledge about
required folding force ff;max . Note that these force values still materials. We hence redefine yielding as a quintary value
result from the normal force measurements, but are denoted 8
with subscript f to indicate the folding phase. Fig. 13 illus- >
> 0 if material recovers shape
>
>
trates how the TU2 implements this definition. < 0:25 if material likely recovers shape
The damping phase identification starts whenever the normal 
13 CYDTUM ¼ 0:5 if uncertain
>
>
>
> 0:75 if material unlikely recovers shape
force increases and ends at the maximum force value. This :
1 if material stays deformed
index range is defined as Dtdamp ¼ f1s  ðnmax;damp  nmin;damp Þ
with fs ¼ 3000 Hz being the sample rate of the DAQ. We (24)
extract the corresponding force difference ff;Dtdamp ¼ f f;nmax 
to account for the state uncertainty for such materials after
f f;nmin;1 which leads to
operator interactions. Only reasonable forces are considered
(
0 a f Þ < 1o
if sða during common human–surface interactions.

11 CDP TUM ¼ ff;Dtdamp (22) 5) Warmth: Observations from Bergmann Tiest et al. [56]
v else show that we can assume a heat dispersion model for material
f
Da surfaces. According to Newton’s continuous law of cooling,
with vf ¼ Dt f .
damp heated material surfaces cool down following the expression
We interpret tactile relaxation as the difference of the maxi-
mum occurring force and the sustained force value, i.e., the dT ðtÞ
last value fn;end of the force recordings using the normal force  ¼ r  ðT ðtÞ  Tamb Þ (25)
dt
FSR during pressing on the material sample.
with Tamb being the ambient and T ðtÞ the current temperature.

12 CRX TUM ¼ fn;max  fn;end (23) This differential equation can be solved to

T ðtÞ ¼ Tamb þ ðTmax  Tamb Þ  ert (26)


This definition is also only applied if the material sample is
deformable at all. Additionally, the majority of material sam-
assuming that surfaces cool down to ambient temperature after
ples sustain their maximum force. However, several samples
an indefinite amount of time. Note that r is an object-specific
slightly deform after the application of force and sustain only
constant containing the mass m, the touched area A, the ther-
a fraction of the maximum force value until the end of the
mal conductivity  and the thermal capacity c
recording.
A
d) Yielding CYDTUM: Yielding is defined as the degree r¼ (27)
to which a surface remains deformed after being pressed [15]. cm
We let the human operator identify this feature value without
using our Texplorer2 device. The motivation is similar to the and is, consequently, related to the object’s thermal cooling
adhesive tack feature definition. First, the majority of the after initial heating. We use our Texplorer unit 2 from Section
scanned materials are not deformable at all, and even if III to heat up the material surface and, subsequently, measure
deformable they completely recover to their initial shape. A the thermal cooling.
human operator can quickly assess this fact whereas a robot- a) Thermal Cooling TCOTUM: Our discrete thermal sig-
based system requires more time and possibly introduces nals t are sampled at fs ¼ 20Hz. Using
unnecessary scanning uncertainty. For example, a piece of
paper placed on a table is identified as rigid by a planar-work- tmax ¼ max1in ti (28)
ing robotic system, but the operator intuitively lifts it, recog-
and measuring, e.g., tamb ¼ 19
C during our recordings, we
nizes that it is not rigid, and also identifies that it might stay
infer the first thermal feature TCOTUM as
deformed after interaction depending on the applied force.
Additionally, a robotic system as well as a purely sensorized
manual approach needs to irreversibly change or destroy the 
14 TCOTUM ¼ tmax  tamb (29)

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 415

Insulating materials locally heat up to a significant tempera- TABLE III


OTHER FEATURES OVERVIEW
ture, whereas metals have almost no measurable difference
compared to their initial surface temperature. Note that the
ambient temperature needs to be recorded whenever a new
recording session starts.
b) Thermal Persistence TPRTUM: The calculation of
TUM
TPR needs to consider two relevant aspects. First, the
maximum temperature after five seconds (proposed by [16])
TCOTUM ¼ tmax;5 s  tamb and the temperature change at the leading to 93 hand-crafted image features. The following list
end of the recording tend  tamb indicate the thermal behavior. shows which features are used:
Secondly, we square both terms to emphasize the larger abso-  Color (Icol ), brightness (Ibr ), edginess (Iedg ), and aver-
lute temperature differences (compared to ambient/room tem- age values of 80%, and 85% of the largest peaks in the
perature) leading to real part of the normalized DFT2 (Idft1;2 )
 14 Haralick [58] (e.g., sum of entropy Ientr and image
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi complexity Icom ) and two Tamura [54] features: coarse-

15 TPRTUM ¼ ðtmax;5 s  tamb Þ2 þ ðtend  tamb Þ2 (30) ness (Icoa ) and contrast (Icon )
 Ten rotational-invariant local binary pattern (RLBP) [61]

of our recorded thermal signals t. A perfect insulator with sig-


D. Hand-Crafted Audio Features
nificant thermal persistence causes a large initial temperature
increase tmax and maintains a large value tend after a specific We use the Matlab implementation from Giannakopoulos
time as indicated by Fishel et al. [16], e.g., up to 10 seconds in et al. [62] to calculate a set of 35 well-known audio features
our recordings. like the spectral centroid, zero crossing rate, and the Mel fre-
quency cepstral coefficients. Note that we apply these on all
B. Tactile Features from Related Work finger and stainless steel tool tip audio signals af;s and at;s dur-
ing sliding, respectively, leading to 70 audio features.
Other researchers in the field have proposed suitable fea-
 Zero-crossing rate, mean signal energy, signal entropy
tures based on signal statistics. For example, Chu et al. [41]
 Spectral centroid, - spread, - roll-off, - entropy, - flux
have shown that the standard deviation and kurtosis of
 Harmonic ratio (HR) and fundamental period (F0)
recorded vibro-tactile signals are characteristic features, or
 13 MFCC
Fishel et al. [22] used features such as the average signal
 12 Chroma values
energy and spectral centroid to achieve a remarkable classifi-
cation accuracy of 95.4% on 117 textures. E. Other Features
C. Hand-Crafted Image Features Table III shows a set of common signal statistics and non-
tactile features which do not fall into any of the previously
Liu et al. give a comprehensive overview of the last mentioned feature domains, but reveal additional and indepen-
two decades of image-based texture classification in [57]. dent material classification abilities.
The description of all presented approaches goes beyond the
scope of our article. The main statement of [57] is that hand-
F. Discussion
crafted feature approaches as well as convolutional neural net-
work (CNN)-based approaches enable a robust classification The materials in the current database do not vary much
of the existing image texture databases. However, these data- along the two tactile dimensions CYDTUM (yielding) and
bases only contain a small set of all available materials ATK TUM (adhesive tack); almost all material samples retain
and their processing shapes. Based on our interdisciplinary their initial shape and have no adhesive surface. However,
naming convention, we consider our database with currently these features are completely independent from the other pro-
184 material labels as an appropriate candidate for further posed features thanks to their different sensing modality and
evaluation of image texture features for both hand-crafted potentially are relevant for haptic display of adhesive surfaces
and deep learning-based approaches consequently. and sustained surface deformation. Similarly, relaxation
Surface images reveal significant material information for CRX TUM has a limited descriptive performance on the current
classes like wood or stones. Consequently, we adapt the most database. Most of the samples are either rigid, or whenever
promising surface texture feature definitions from Haralick compliant, naturally maintain the maximum force while
et al. [58], Tamura et al. [54], and use the corresponding actively being pressed. However, if further materials were
Matlab implementations from Monzel et al. [59] and Sdhir added to the database, these features potentially enable a more
et al. [60]. We further work with features like the dominant subtle classification.
color (Icol ) or edginess (Iedg ) from our work in [27] and calcu- The features not falling in the audio, image, or tactile
late ten rotation-invariant local binary pattern (RLBP) descrip- domain are very characteristic for specific material subclasses.
tor values proposed by Ojala et al. [61]. In total, we calculate Notably, the density is the best feature to discriminate between
31 surface image features on Idisp , Imacro , and Iillu respectively, the variety of woods, or, the combination of the metal,

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
416 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

Fig. 14. Example boxplot of feature CCPTUM of class C2 (biodegradables


like furs or leather). The feature value is always set to zero for stiff materials
(if sðaaf Þ < 1o ), whereas it has a good distinction performance for soft
materials. Fig. 15. Explained variance over the principal components.

dominant color, and magnetic feature works well for metal feature space. Larger correlations can be expected for sub-
classification. The surface reflectivity is also highly character- dimensions belonging to the same major tactile dimension.
istic and independent of operator exploratory motions since For example, FSTTUM and FRSTUM belong to the dimension
the data is captured during static touch. of friction and are intuitively correlated.

V. SUPERVISED MATERIAL CLASSIFICATION 2) Feature Optimization: Several features from Section IV


The primary goal of this work is the classification of the dif- contain parameters that need to be optimized in terms of fea-
ferent material samples from the database recordings. ture quality. For example, the optimal wavelet type for the fea-
ture MiROTUM needs to be determined by testing different
A. Data Preparation wavelet types followed by an evaluation of how the feature
quality and its correlation to the other features changes. Our
This section shows how the features are evaluated and used optimization approach follows the same idea of high feature
to set up the feature space before we train a classifier. quality (FQ) and low correlation between the features. The
corresponding feature space columns need to be recalculated
1) Feature Quality and Correlation: The most important using different parameter sets. We ran a grid search of poten-
ingredients of machine learning are content-based, descriptive, tial parameters to identify the maximum FQ and minimum
and independent features [63]. Our features are heavily correlation values. For example, the weighting parameters
inspired by the major tactile dimensions from Okamoto 1...3 of MaTXTUM result in a set of (3, 2, 4) indicating that the
et al. [14] and Fishel et al. [16], and hence, we believe that influence of the reflectance sensor data is most relevant and
they fulfill the first condition. also causally independent since the reflectance sensor is not
Since the individual material scans are assumed to be inde- used for any other TacTUM feature.
pendent, one-way analysis of variance (ANOVA) is conducted
on each feature to assess its descriptive quality based on the 3) Raw Feature Space Creation: Our initial feature space
Kruskal-Wallis test. Boxplots visualize the feature values, as X0 is composed of 193 features (columns f of X0 ) before fea-
shown for CCPTUM and a subset of materials in Fig. 14. ture selection takes place. The last additional five columns are
Post-hoc analysis based on the ANOVA statistics performs reserved for the responses y (true labels) based on our naming
pair-wise comparisons between all class means (Matlab mult- convention Cx Sx Mx Px Xx . For example, coarse classifica-
compare implementation, alpha = 5% and Tukey-Kramer pro- tion yC considers the eight major classes C1 – C8 regardless
cedure for balanced one-way ANOVA with equal sample of the sub-class or material identifier. Five observations per
size). Larger p-values indicate low discrimination between material compose the instances (rows) of X0 . We split our fea-
two classes, and hence, we define the feature quality (FQ) as ture space based on k-fold cross-validation (five folds) using
the function cvpartition() from Matlab.
FQ ¼ 100%  ð1  pÞ (31)
4) Feature Space Scaling: In order to make the features
based on the average of all p-values p resulting from the pair- comparable to each other [63], we create a normalized (feature
wise comparisons NðN 1Þ with N ¼ 184 being the number of
2 values between 0 and 1) feature space X01 . Each feature vector
f i fmin
material classes used in this article. A descriptive feature entry, i.e. feature fi , is normalized by applying f 0i ¼ fmax fmin .
shows a small intra-class and a large inter-class variance, We denote the raw feature space X01 as X in the following.
and hence, has a large FQ value. Note that this metric is
also used as one target parameter for feature optimization in 5) Feature Space Reduction and Selection: It is general
Section V-A2. practice to keep the number of features as low as possible to
Thirdly, the features ideally are as weakly correlated as pos- avoid over-fitting while using a sufficient number of descriptive
sible to guarantee feature independence. We consider the features to avoid under-fitting as highlighted by Domingo
Spearman rank correlation rspearman between two features as et al. [65]. One approach to reduce the initial feature space is
meaningful metric. Note that the Pearson’s correlation coeffi- dimensionality reduction. We first performed a principal compo-
cient is less appropriate for this task due to the potential nent analysis (PCA) on the initial feature space to assess if signif-
non-linearity in the data [64]. Whenever the absolute value of icant linearly uncorrelated variables could be identified.
correlation indicates stronger correlation, i.e., rspearman 0.5, However, the three major principal components only explain
one of the two compared features should be excluded from the about 30% of the variance, and the first ten about 62%, as shown

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 417

TABLE IV TABLE V
BEST FEATURES FROM DIFFERENT TESTED DOMAINS CRITICAL CORRELATIONS OF FINAL TACTUM FEATURES. NOTE THAT THE
FEATURES RELATED TO STIFFNESS ARE CORRELATED DUE TO THE FACT THAT
THE DATABASE CONTAINS MUCH MORE RIGID THAN DEFORMABLE MATERIALS,
AND HENCE, THE CORRESPONDING STIFFNESS FEATURE VALUES ARE SET TO 0

in Fig. 15, indicating that a large number of principal components


would be required to allow for successful material classification.
Consequently, PCA is not applied on the raw feature space.
We instead followed a feature filtering approach according to
[64] and used our FQ metric to identify the n best features in C. Results and Discussion
the feature space. Also note that other feature reduction The accuracy of a classification engine is an intuitive metric
approaches such as the supervised linear discriminant analysis whenever classes are balanced [63]. We additionally calculate
(LDA) can instead be used as a classifier on a feature space the F1 scores (averaged per class) for each classifier, because
based on manual feature selection. we face a class imbalance on a coarse and medium classifica-
We iterate through these features and check if a feature cor- tion level. Figs. 16 and 17 show the accuracies (top) and F1
relation exceeds the critical threshold and either accept or scores (bottom) for the feature spaces with ten selected fea-
reject it for the reduced feature space. For the sake of compari- tures X10 from different domains, e.g., purely image-based
son with the three features traction, roughness, fineness from features.The F1 score is specifically more descriptive when-
Fishel et al. [22], we choose to extract the three best features ever faced with class imbalance, which is visible in Fig. 16. If
to define X3 as well as the ten best features in a feature space fine material classification based on all 184 material classes is
X10 . Table IV shows the feature names of the best features considered, this imbalance is not present anymore.
from different modalities. We further evaluate only the three best features of the tactile
Table V shows the final correlations between the TacTUM domain as well as from our previous work in [28] and the fea-
features defined in this article. For example, TCOTUM and tures (traction, roughness, fineness) presented by Fishel
TPRTUM are strongly correlated, and hence, cannot both be et al. [22]. Note that these features are only mathematically
selected for the final feature space. Similar to adhesive tack adapted from their original definition from [22] using our defini-
and yielding, both features presumably are highly relevant for tions FRSTUM for traction, the logarithmic band-filtered signal
haptic display as they are related to the initial and persistent power for roughness (as in [22]), and MiCOTUM for fineness.
thermal drain of energy, though. We observe that three features lead to an accuracy of about 30%
to 65% shown in Fig. 18. We can thereby confirm that the fea-
B. Hand-Crafted Features Classifier Evaluation tures proposed by Fishel et al. [22] and the three best TacTUM
Our model is trained and tested on the different sets of the features show a noticeable classification performance. The rea-
five-fold partitions. We further evaluate our classifiers based son why the best three (best-3 from Table IV) features even out-
on three different classification levels (coarse, medium, fine) perform these results is caused by the best feature, the material
and two metrics (accuracy and F1 score). Coarse classification density. We conclude that whenever the density can be extracted
refers to the discrimination of C, i.e., how the classification reliably, it significantly improves the classification performance
engine is able to distinguish between the major classes C1 – of any material scanning system and even outperforms human
C8 regardless the sub-class S or the material M. Medium clas- classification performance. This does not come as a surprise. For
sification further includes the sub-class S, e.g., hardwood or example, the densities of woods are remarkably different
softwood, as shown in Table I. Fine classification additionally whereas visual and tangible information can be easily confused
considers the material M, such as aluminum, ABS, or oak. We even by humans examining such material samples.
apply common classifiers such as support-vector machines
D. Deep Learning-Based Material Classification
(SVM), linear discriminant analysis (LDA), k-nearest neighbor
(k-NN), neural network (NN), bagged trees, also known as ran- Convolutional neural networks (CNNs) have been applied for
dom forest (RF), and Naive Bayes (NB). The neural network is handwritten number classification by LeCun et al. [67]. Kriz-
trained using 60% training, 20% validation, and 20% test data hevsky et al. [68] further have proven the success of using
based on 500 hidden layers (tested with different numbers of CNNs for image-based object recognition, and ever since,
layers) and a scaled conjugate gradient back-propagation. several adaptations like VGGM by Cimpoi et al. [69] or

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
418 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

Fig. 17. Results of the fine classification with X10 . The dotted line shows the
Fig. 16. Results of the coarse classification with X10 (top: accuracy, bot- best accuracy of 90.2%  1.2% and F1 score of 0.90  0.01.
tom: F1 score) for different classifiers. The x-axis shows the sets of features
used to train the classifiers.

Fig. 19. DL network based on AlexNet [68] adapted to perform a fine classi-
fication of the 184 material classes in this article. Each convolutional layer is
followed by a batch normalization and rectified linear unit (ReLU) layer. If
not explicitly stated, the stride was set to 2 and the padding to “same”.

Fig. 18. Results of the fine classification with X3 (top: accuracy, bottom: F1 fix size of 768 x 330 x 3) and achieved a comparable result of
score) using only the three best features from different domains and related work. 95%. The actual bottleneck is the amount of correctly labeled
data to learn from. Existing deep learning-based networks
VGGVD [70] have been proposed for surface texture classifica- almost perfectly classify the available texture image databases
tion. Bello et al. [71] extensively compared deep networks with as summarized in [57]. Consequently, we consider our extensive
hand-crafted image features. Deep networks outperform other image database with the new material labeling taxonomy as
approaches in terms of texture classification whenever non-sta- highly suitable to run further validations. We sliced the captured
tionary textures and the presence of multiple acquisition condi- RGB images Idisp from Section III into [500 x 500] RGB patches
tions changes is considered. By contrast, hand-crafted features leading to 175 images per material class with no pixel overlap
are superior in distinguishing stationary textures under steady between slices. We rotated each image by 90o to augment this
imaging conditions and have proven to be more robust than data-set, leading to 350 images per material class. We applied a
CNN-based features for variation in the image rotation [71]. split of 80% training and 20% test data. Fig. 20 shows a few ran-
Due to the rapid progress of deep learning, a comparison to the domly selected patches from our DL database which we denote
hand-crafted feature approach is inevitable in this context. as TexTUM database.
We use the following hyper-parameters for training.
1) Network Architecture and Results: Liu et al. [57] sum-  Solver: stochastic gradient descent (SGDM) with
marize that the network structures AlexNet, VGGM, VGGVD, momentum of 0.9, alternatively: adam optimizer
and TCNN literally solved the texture classification problem on  Mini batch size of 128
the largest existing texture data-sets such as CUReT [72], KTH-  Initial learning rate of 0.01
TIPS2 [73], or ALOT [74] by achieving a classification perfor-  L2 regularization of 0.0001,
mance of 99% according to Cimpoi et al. [69]. That is why we  dropout layers with 40% dropout
use our database images based in these previously defined net- Overall, the DL network achieves a classification accuracy
work structures and do not go further into the network design of 90.7%  1.0% after 40 epochs of training, and hence,
details. Fig. 19 shows the DL network which we use in this arti- slightly outperforms all hand-crafted approaches.
cle. It is transfer-learned from AlexNet [68] which has proven to
work excellently for object recognition. We also cross-validated 2) Misclassification Comparison of Deep Learning and
the adapted network with the ALOT database (images sliced to a Hand-Crafted Features: Overall, the image-based DL network

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 419

using more content-based features. Until recently, extensive


feature design filled above 90% of the engineering work [65].
Deep learning-based approaches further confirm to be a valid
and promising alternative to the proposed hand-crafted feature
approaches [57], [68] in this article. Beyond purely image-
based approaches, deep neural networks have also been trained
on haptic-related data as shown by Gao et al. [75] and in our
Fig. 20. Thirty exemplary [500 x 500] RGB material image patches for the previous work [76], [77], e.g., on spectrograms of vibro-tactile
DL network. Overall, 64,400 of these texture patches are used to train (75%) signals to learn hidden features from the tactile domain. We
and test (25%) the DL network in this article to perform the fine classification
of the 184 materials constituting the largest taxonomy-based image texture
believe that a hybrid approach of prior knowledge (based on
database to date (TexTUM database). feature design) and deep neural networks will ultimately com-
bine the advantages of both methodologies, which already
slightly outperforms the hand-crafted feature approach in terms show high standalone classification performance. Up to this
of classification accuracy. However, one critical point arises point, even larger amounts of haptic-related data is required
considering content-based material classification. Several rele- which supports the fact that versatile object scanners (i.e.,
vant material characteristics, like the friction coefficient or the robot-based or operator-based) are crucial to capture the data
compliance, cannot be learned by such an image-based deep based on an appropriate material taxonomy. In this context we
neural network, and hence, the system has a poor understanding would like to emphasize the relevance of taxonomies like the
of haptic material properties. This observation can be proven by one proposed; materials need to be unambiguously catego-
considering the misclassified labels based on the true labels y rized, notably according to their actual processing shape, to
and the predictions y^. In order to account for this observation, ensure potential merging of haptic databases in the future.
we count the overall misclassified labels N (with y^ ¼ 6 y) and
compute the ratio rC of the misclassified major class labels VI. CONCLUSION
y^C ¼
6 yC . We extract the class identifier y^C from each misclassi- This article presented an improved material property acquisi-
fication y^ and calculate tion device (Texplorer2), a novel material taxonomy and a new
PN multi-modal feature space definition. A complete pipeline was
n¼1 ð1  dy^C;n ;yC;n Þ proposed that can be extended with further materials due to the
rC ¼ (32) flexible recording setup and the taxonomy. We defined a set of
N
tactile features based on human exploratory procedures, and
with dij being the Kronecker delta, which is equal to one if the together with image and audio features, common classifiers
two integer arguments i and j (the predicted and true labels) are were trained to successfully identify the materials. The compar-
the same. We consider it critical whenever the predicted class ison with image-based deep learning showed that they outper-
label y^C is not equal to the true class label yC and then increase form hand-crafted approaches in terms of accuracy, but led to
the numerator by one. The denominator denotes the number of obvious misclassifications in terms of haptic similarity. Ongo-
all misclassifications N based on the fine classification material ing research in this field now needs to collect more data to rep-
labels. The best hand-crafted (HC) approach using the random resent all different processing shapes of the materials, and also
forest classifier leads to rC;HC ¼ 0.05, and the trained DL net- should prove if the fifteen tactile features are descriptive enough
work leads to rC;DL ¼ 0.24. Despite a better classification accu- for haptic display. We intend to use the fifteen-dimensional fea-
racy (i.e., less misclassifications), the labeling error of the DL ture space as input parameters for haptic display. A separate
network is still larger than the one of the hand-crafted approach. manuscript will present how the features generate the control
Intuitively, the image-based DL network faces a critical chal- signals for a novel haptic display device. Subjective experi-
lenge while distinguishing between hard and soft materials and ments will be carried out to show the effectiveness in rendering
has no notion of the characteristic microscopic roughness or the tactile dimensions. Additionally, the Texplorer2 will be
surface friction. By contrast, hand-crafted tactile features repre- equipped with a speed measurement unit, e.g., contact-less by
sent such features more robustly, but tend to confuse the materi- mounting permanent magnets and a magnetometer. These
als among each other. That is why we believe that both measurements are required to constrain the operator motion
approaches have their advantages and right of existence, and into a specific range as well as to provide vibro-tactile signals
ideally are fused into a hybrid approach in future work. of different scan speed for potential haptic display.

E. Discussion ACKNOWLEDGEMENT
We can draw the conclusion that a selection of features We would like to thank the Chair of Building Construction and
from different modalities performs best when combined with Material Science (TUM Department of Architecture) for access
popular classifiers like Naive Bayes or random forest, which to the collection of building materials area, and specifically, Dipl.
also have been applied successfully to comparable classifica- Ing. Johann Weber for providing extensive information about the
tion tasks. Ten features from different domains already lead to materials. In addition, the authors would like to thank the anony-
a satisfying performance on different classification label detail mous reviewers for their constructive comments which helped to
levels, and we observe even larger accuracies and F1 scores improve this article.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
420 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

APPENDIX

Fig. 21. Overview of all 184 material classes in the database.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 421

TABLE VI
MATERIAL CATEGORIZATION CONVENTION TABLE. FOR A COMPLETE LIST PLEASE REFER TO OUR WEBSITE [42]

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
422 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

REFERENCES [12] J. M. Yau, J. B. Olenczak, J. F. Dammann, and S. J. Bensma€ıa,


“Temporal frequency channels are linked across audition and touch,”
[1] S. J. Lederman and R. L. Klatzky, “Haptic perception: A tutorial,” Atten- Current Biol., vol. 19, no. 7, pp. 561–566, 2009.
tion, Perception, Psychophys., vol. 71, no. 7, pp. 1439–1459, 2009. [13] J. A. Fishel and J. E. Colgate. Haptics symposium 2018 cross-cutting
[2] S. Lacey, C. Campbell, and K. Sathian, “Vision and touch: Multiple or challenge: Haptic dimensions of surfaces. (2019). [Online]. Available:
multisensory representations of objects?,” Perception, vol. 36, no. 10, http://2018.hapticssymposium.org/ccc2
pp. 1513–1521, 2007. [14] S. Okamoto, H. Nagano, and Y. Yamada, “Psychophysical dimensions
[3] Y. Vardar, C. Wallraven, and K. J. Kuchenbecker, “Fingertip interaction of tactile perception of textures,” IEEE Trans. Haptics, vol. 6, no. 1,
metrics correlate with visual and haptic perception of real surfaces,” in pp. 81–93, First Quarter 2013.
Proc. IEEE World Haptics Conf., 2019, pp. 395–400. [15] SynTouch. (2019). Syntouch Inc. website. [Online]. Available: https://
[4] K. Eilers, “D. Katz, Der Aufbau der Tastwelt (book review),” www.syntouchinc.com/en/
Kant-Studien, vol. 31, p. 104 ff, 1926. [16] J. A. Fishel, G. E. Loeb, B. Matulevich, and R. Davoodi, “Method and
[5] S. J. Lederman, “Tactile roughness of grooved surfaces: The touching applications for measurement of object tactile properties based on how
process and effects of macro-and microsurface structure,” Perception they likely feel to humans,” U.S. Patent 1 479 6647, 2016.
Psychophys., vol. 16, no. 2, pp. 385–395, 1974. [17] K. J. Kuchenbecker, J. M. Romano, and W. McMahan, “Haptography:
[6] R. L. Klatzky, S. J. Lederman, C. Hamilton, M. Grindley, and Capturing and recreating the rich feel of real surfaces,” in Robot.
R. H. Swendsen, “Feeling textures through a probe: Effects of probe Research. Berlin, Germany: Springer, 2011, pp. 245–260.
and surface geometry and exploratory factors,” Perception Psychophys., [18] J. M. Romano, T. Yoshioka, and K. J. Kuchenbecker, “Automatic filter
vol. 65, no. 4, pp. 613–631, 2003. design for synthesis of haptic textures from recorded acceleration data,”
[7] S. J. Bensma€ıa and M. Hollins, “The vibrations of texture,” Somatosen- in Proc. IEEE Int. Conf. Robot. Autom., 2010, pp. 1815–1821.
sory Motor Res., vol. 20, no. 1, pp. 33–43, 2003. [19] H. Culbertson, J. Unwin, and K. J. Kuchenbecker, “Modeling and
[8] S. J. Bensma€ıa and M. Hollins, “Pacinian representations of fine rendering realistic textures from unconstrained tool-surface inter-
surface texture,” Perception Psychophys., vol. 67, no. 5, pp. 842– actions,” IEEE Trans. Haptics, vol. 7, no. 3, pp. 381–393, Jul.–Sep.
854, 2005. 2014.
[9] W. M. Bergmann Tiest and A. M. Kappers, “Haptic and visual [20] H. Culbertson and K. J. Kuchenbecker, “Importance of matching
perception of roughness,” Acta Psychol., vol. 124, no. 2, pp. 177– physical friction, hardness, and texture in creating realistic haptic virtual
189, 2007. surfaces,” IEEE Trans. Haptics, vol. 10, no. 1, pp. 63–74, Jan.–Mar.
[10] W. M. Bergmann Tiest, “Tactual perception of material properties,” 2017.
Vision Res., vol. 50, no. 24, pp. 2775–82, 2010. [21] N. Jamali and C. Sammut, “Majority voting: Material classification by
[11] C. E. Schroeder, J. Smiley, K. G. Fu, T. McGinnis, M. N. O’Connell, tactile sensing using surface texture,” IEEE Trans. Robot., vol. 27,
and T. A. Hackett, “Anatomical mechanisms and functional implications no. 3, pp. 508–521, Jun. 2011.
of multisensory convergence in early cortical processing,” Int. J. Psy- [22] J. A. Fishel and G. E. Loeb, “Bayesian exploration for intelligent identi-
chophysiol., vol. 50, no. 1–2, pp. 5–17, 2003. fication of textures,” Frontiers Neurorobot., vol. 6, p. 4, 2012.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
STRESE et al.: HAPTIC MATERIAL ANALYSIS AND CLASSIFICATION INSPIRED BY HUMAN EXPLORATORY PROCEDURES 423

[23] J. M. Romano and K. J. Kuchenbecker, “Methods for robotic tool- [48] H. Culbertson, J. M. Romano, P. Castillo, M. Mintz, and
mediated haptic surface recognition,” in Proc. IEEE Haptics Symp., K. J. Kuchenbecker, “Refined methods for creating realistic haptic vir-
2014, pp. 49–56. tual textures from tool-mediated contact acceleration data,” in Proc.
[24] J. Hoelscher, J. Peters, and T. Hermans, “Evaluation of tactile feature IEEE Haptics Symp., 2012, pp. 385–391.
extraction for interactive object recognition.,” in Proc. Humanoids, [49] E. Meier. The wood database. (2019). [Online]. Available: https://www.
2015, pp. 310–317. wood-database.com/
[25] M. Kaboli and G. Cheng, “Robust tactile descriptors for discriminating [50] Material Archiv. (2019). [Online]. Available: http://www.materia-
objects from textural properties via artificial robotic skin,” IEEE Trans. larchiv.ch/
Robot., vol. 34, no. 4, pp. 985–1003, Aug. 2018. [51] A. Burka and K. J. Kuchenbecker, “Handling scan-time parameters in
[26] A. Burka et al., “Proton: A visuo-haptic data acquisition system for haptic surface classification,” in Proc. IEEE World Haptics Conf., 2017,
robotic learning of surface properties,” in Proc. IEEE Int. Conf. Multi- pp. 424–429.
sensor Fusion Integrat. Intell. Syst., 2016, pp. 58–65. [52] C. Harris and M. Stephens, “A combined corner and edge
[27] M. Strese, C. Schuwerk, A. Iepure, and E. Steinbach, “Multimodal detector.” in Proc. Alvey Vision Conf., Citeseer, 1988, vol. 15, no. 50.
feature-based surface material classification,” IEEE Trans. Haptics, pp. 1–5.
vol. 10, no. 2, pp. 226–239, Apr.–Jun. 2017. [53] H. Bay, T. Tuytelaars, and L. Van Gool, “Speeded up robust features
[28] M. Strese, Y. Boeck, and E. Steinbach, “Content-based surface (SURF),” in Proc. Eur. Conf. Comput. Vision., Springer, 2006, pp. 404–
material retrieval,” in Proc. IEEE World Haptics Conf., 2017, 417.
pp. 352–357. [54] H. Tamura, S. Mori, and T. Yamawaki, “Textural features correspond-
[29] S. Sato, S. Okamoto, Y. Matsuura, and Y. Yamada, “Wearable finger ing to visual perception,” IEEE Trans. Syst., Man, Cybern., vol. 8, no. 6,
pad sensor for tactile textures using propagated deformation on a side of pp. 460–473, Jun. 1978.
a finger: Assessment of accuracy,” in Proc. IEEE Int. Conf. Syst., Man, [55] J. P. Den Hartog, Mechanical Vibrations. Courier Corporation, p. 55 ff,
Cybern., 2015, pp. 892–896. 1985.
[30] Y. Vardar, B. G€ uçl€
u, and C. Basdogan, “Effect of waveform on tactile [56] W. M. Bergmann Tiest and A. M. Kappers, “Tactile perception of
perception by electrovibration displayed on touch screens,” IEEE Trans. thermal diffusivity,” Attention Perception Psychophys., vol. 71, no. 3,
Haptics, vol. 10, no. 4, pp. 488–499, Oct.–Dec. 2017. pp. 481–489, 2009.
[31] Y. Visell and Y. Shao, “Learning constituent parts of touch stimuli [57] L. Liu, J. Chen, P. Fieguth, G. Zhao, R. Chellappa, and M. Pietik€ainen,
from whole hand vibrations,” in Proc. IEEE Haptics Symp., 2016, “From BoW to CNN: Two decades of texture representation for tex-
pp. 253–258. ture classification,” Int. J. Comput. Vision, vol. 127, no. 1, pp. 74–109,
[32] R. Grigorii, M. Peshkin, and J. E. Colgate, “High-bandwidth tribometry 2019.
as a means of recording natural textures,” in Proc. IEEE World Haptics [58] R. M. Haralick, K. Shanmugam, and I. H. Dinstein, “Textural features
Conf., Munich, Germany, 2017, pp. 629–634. for image classification,” IEEE Trans. Systems, Man, Cybern., vol.
[33] S. Shin and S. Choi, “Geometry-based haptic texture modeling and SMC-3, no. 6, pp. 610–621, Nov. 1973.
rendering using photometric stereo,” in Proc. IEEE Haptics Symp., [59] R. Monzel. Haralick texture features Matlab. (2019). [Online]. Avail-
San Francisco, CA, USA, 2018, pp. 262–269. able: https://de.mathworks.com/matlabcentral/fileexchange/58769-hara-
[34] M. Dulik and L. Ladanyi, “Surface detection and recognition using licktextu refeatures
infrared light,” in Proc. IEEE Elektro, 2014, pp. 159–164. [60] S. Sornapudi. Tamura features. (2019). [Online]. Available: https://
[35] T. Aujeszky, G. Korres, and M. Eid, “Measurement-based thermal github.com/Sdhir/TamuraFeatures
modeling using laser thermography,” IEEE Trans. Instrum. Meas., [61] T. Ojala, M. Pietikainen, and T. Maenpaa, “Multiresolution gray-scale
vol. 67, no. 6, pp. 1359–1369, Jun. 2018. and rotation invariant texture classification with local binary patterns,”
[36] H. Choi, S. Cho, S. Shin, H. Lee, and S. Choi, “Data-driven thermal ren- IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 7, pp. 971–987,
dering: An initial study,” in Proc. IEEE Haptics Symp., San Francisco, Jul. 2002.
CA, USA, 2018, pp. 344–350. [62] T. Giannakopoulos and A. Pikrakis. Introduction to audio analysis.
[37] C. Basdogan and M. A. Srinivasan, “Haptic rendering in virtual (2019). [Online]. Available: https://de.mathworks.com/matlabcentral/
environments,” Handbook Virtual Environ., vol. 1, pp. 117–134, fileexchange/45831-matlab-audio- analysis-library
2002. [63] P. Flach, Machine Learning: The Art and Science of Algorithms That
[38] A. M. Okamura, J. T. Dennerlein, and R. D. Howe, “Vibration feedback Make Sense of Data. Cambridge, U.K.: Cambridge University Press,
models for virtual environments,” in Proc. IEEE Int. Conf. Robot. Auto- 2012.
mat., 1998, vol. 1, pp. 674–679. [64] I. Guyon and A. Elisseeff, “An introduction to variable and feature
[39] K. J. Kuchenbecker, J. Fiene, and G. Niemeyer, “Improving contact real- selection,” J. Mach. Learn. Res., vol. 3, no. Mar., pp. 1157–1182, 2003.
ism through event-based haptic feedback,” IEEE Trans. Visual. Comput. [65] P. Domingos, “A few useful things to know about machine learning,”
Graph., vol. 12, no. 2, pp. 219–230, Mar./Apr. 2006. Commun. ACM, vol. 55, no. 10, pp. 78–87, 2012.
[40] J. Sinapov and V. Sukhoy, “Vibrotactile recognition and categorization [66] M. Strese, C. Schuwerk, and E. Steinbach, “Surface classification
of surfaces by a humanoid robot,” IEEE Trans. Robot., vol. 27, no. 3, using acceleration signals recorded during human freehand move-
pp. 488–497, Jun. 2011. ment,” in Proc. IEEE World Haptics Conf., Evanston, IL, USA, 2015,
[41] V. Chu et al., “Robotic learning of haptic adjectives through physical pp. 214–219.
interaction,” Robot. Auton. Syst., vol. 63, pp. 279–292, 2015. [67] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based
[42] M. Strese, C. Schuwerk, R. Chaudhari, and E. Steinbach, “Haptic texture learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11,
database,” [Online]. Available: https://zeus.lmt.ei.tum.de/downloads/ pp. 2278–2324, Nov. 1998.
texture/ [68] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification
[43] H. Culbertson, J. J. Lopez Delgado, and K. J. Kuchenbecker, “One with deep convolutional neural networks,” in Proc. Adv. Neural Inf.
hundred data-driven haptic texture models and open-source methods Process. Syst., 2012, pp. 1097–1105.
for rendering on 3d objects,” in Proc. IEEE Haptics Symp., 2014, [69] M. Cimpoi, S. Maji, and A. Vedaldi, “Deep filter banks for texture
pp. 319–325. recognition and segmentation,” in Proc. IEEE Conf. Comput. Vision
[44] W. Hassan, A. Abdulali, M. Abdullah, S. C. Ahn, and S. Jeon, “Towards Pattern Recognit., 2015, pp. 3828–3836.
universal haptic library: Library-based haptic texture assignment using [70] K. Simonyan and A. Zisserman, “Very deep convolutional networks for
image texture and perceptual space,” IEEE Trans. Haptics, vol. 11, large-scale image recognition,” 2014, arXiv:1409.1556.
no. 2, pp. 291–303, Apr.–Jun. 2018. [71] R. Bello-Cerezo, F. Bianconi, F. Di Maria, P. Napoletano, and
[45] F. Musso and J. Weber, TUM collection of building materials. (2019). F. Smeraldi, “Comparative evaluation of hand-crafted image descriptors
[Online]. Available: https://www.ar.tum.de/en/ebb/collection-of-build- vs. off-the-shelf CNN-based features for colour texture classification
ing-materials/ under ideal and realistic conditions,” Appl. Sci., vol. 9, no. 4, p. 738, 2019.
[46] H.-J. Bargel and G. Schulze, Werkstoffkunde. Berlin, Germany: [72] K. J. Dana, B. Van Ginneken, S. K. Nayar, and J. J. Koenderink,
Springer-Verlag, 2008. “Reflectance and texture of real-world surfaces,” ACM Trans. Graph.,
[47] N. Landin, J. M. Romano, W. McMahan, and K. J. Kuchenbecker, vol. 18, no. 1, pp. 1–34, 1999.
“Dimensional reduction of high-frequency accelerations for haptic ren- [73] P. Mallikarjuna, A. T. Targhi, M. Fritz, E. Hayman, B. Caputo, and
dering,” in Proc. Int. Conf. Human Haptic Sensing Touch Enabled J.-O. Eklundh, The KTH-TIPS2 Database. Stockholm, Sweden: Compu-
Comput. Appl., Springer, 2010, pp. 79–86. tational Vision and Active Perception Laboratory (CVAP), 2006.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.
424 IEEE TRANSACTIONS ON HAPTICS, VOL. 13, NO. 2, APRIL-JUNE 2020

[74] J. M. Geusebbroek, Amsterdam library of textures (ALOT). (2019). Jonas Kirsch received the Bachelor of Science
[Online]. Available: http://aloi.science.uva.nl/public/alot/ degree in electrical and computer engineering from
[75] Y. Gao, L. A. Hendricks, K. J. Kuchenbecker, and T. Darrell, “Deep the Technical University of Munich and is currently
learning for tactile understanding from visual and haptic data,” in Proc. working toward the Master of Science degree. In the
IEEE Int. Conf. Robot. Automat., 2016, pp. 536–543. summer of 2018, he completed his bachelor thesis in
[76] M. Ji, L. Fang, H. Zheng, M. Strese, and E. Steinbach, “Preprocessing- the field of haptics and tactile data acquisition and
free surface material classification using convolutional neural networks has been working ever since at the Chair of Media
pretrained by sparse autoencoder,” in Proc. IEEE 25th Int. Workshop Technology.
Mach. Learn. Signal Process., 2015, pp. 1–6.

[77] H. Zheng, L. Fang, M. Ji, M. Strese, Y. Ozer, and E. Steinbach, “Deep
learning for surface material classification using haptic and visual
information,” IEEE Trans. Multimedia, vol. 18, no. 12, pp. 2407–2416,
Dec. 2016. Eckehard Steinbach (M’96–SM’08–F’15) received
the degree in electrical engineering from the Univer-
Matti Strese received the degree in electrical engi- sity of Karlsruhe (Germany), University of Essex,
neering from the Technical University of Munich, Britain, and ESIEE in Paris. From 1994 to 2000,
Germany. He received the Master of Science degree he was a member of the Research Staff of the Image
in July 2014. After graduating, he joined the Chair of Communication Group, University of Erlangen-
Media Technology, Technical University of Munich, Nuremberg, Germany, where he received the engi-
in September 2014, where he is a member of the neering doctorate in 1999. From February 2000 to
Research Staff and working toward the Ph.D. degree. December 2001, he was a Postdoctoral Fellow with
His current research interests include the analysis of the Information Systems Laboratory of Stanford Uni-
haptic texture signals, surface classification, and arti- versity. In February 2002, he joined the Department
ficial surface synthesis devices. of Electrical and Computer Engineering, Technical University of Munich,
Germany, where he is currently a Full Professor of Media Technology. His
current research interests include the area of haptic and visual communication,
Lara Brudermueller received the Bachelor of Sci- indoor mapping, and localization.
ence degree in management and technology with spe-
cialization in electrical engineering from the
Technical University of Munich, Germany. She is
currently working toward Master of Science degree
in robotics, cognition, and intelligence from the
Technical University of Munich. In the summer of
2018, she wrote her bachelor’s thesis in the field of
haptics and object surface recognition at the Chair of
Media Technology.

Authorized licensed use limited to: SLUB Dresden. Downloaded on June 02,2023 at 07:51:49 UTC from IEEE Xplore. Restrictions apply.

You might also like