1. Introduction
The diversity of objects is reflected in the different attributes of the objects and the difference in the strength level of the same attributes. When modeling the diversity of objects based on deep learning methods, such as the image classification of visual data [1] and the speech recognition of audio data [2], it is often necessary to pay attention to the differences among objects. In the field of haptics, the difference in haptic signals often reflects the difference in the properties of measured objects, which is an important basis for classifying the input into discrete categories.
In recent years, with the development and application of deep learning theory, many scholars have taken into account the unique advantages of tactile information and have begun to devote themselves to the fields of tactile perception and attribute recognition by using deep learning methods [3]. The tactile sensor can provide the robot with feedback information, such as the interaction force [4]: for example, whether or not the sensor is in contact with the object [5], whether there is sliding [6], and what the physical properties of the object are, such as its temperature [7], roughness, and texture [8,9].
Because objects with different material properties deliver different feedback to the tactile sensor during the interaction, the task of many deep learning methods is to classify tactile data into discrete categories to identify objects [10,11]. Robots can collect this feedback information through dynamic exploration programs [12,13], for example, [14,15] letting volunteers touch objects to feel their physical characteristics to learn the meaning of tactile attribute adjectives, and establishing a tactile data set by designing an exploration program, but the authors only require each volunteer to give a binary label (yes/no). The data set cannot evaluate the attributed strength level of objects. In addition, many scholars have conducted in-depth research on the identification of the object’s temperature, surface texture, softness, and hardness. For example, Ref. [14] used a convolutional neural network (CNN) to analyze the time series data of pressure and six-axis acceleration sensors to obtain tactile texture information. The static friction coefficient of an object can be estimated by measuring the normal force and tangential force during the initial slip [16]. In [17], a variety of deep learning models are used to explore surface roughness, and the feature extraction capabilities of models with different roughness ranges are introduced in detail. In [18], an artificial nail with a three-axis acceleration sensor is used to recognize the surface of vibration when scratching the surface of different materials, and proves that good performance can be obtained by using a variety of exploration behaviors. In [19], according to the signal amplitude caused by vibration, the k-nearest neighbor method is used to distinguish different surfaces. In [20], a CNN is used to identify the tactile information of human or non-human objects in disaster scenes to protect people. In [21], a CNN is used to predict object attributes, but the feature extraction of each attribute requires a classifier, and the amount of network parameters is large.
Many researchers try to teach robots to recognize adjectives concerning an object’s tactile attributes from raw data, try to mark objects by using adjectives concerning binary tactile attributes (for example, hard or not hard), and map the characteristics of tactile data. However, the above method has the following shortcomings:
In the existing public data set [14,15], experimenters are only required to give each object attribute a binary label (yes/no). The binary label does not include the knowledge of the strength level of the object’s delicate attribute.
In real scenes, humans who have a rich tactile perception system will obtain a delicate sensory feedback by touching an object and will have a specific cognition of the strength level of the object’s attribute. For networks, only using binary tactile labels to describe objects will simplify object attributes to binary space and the network has a very rough understanding of the level of the strength of an object’s attribute [22].
In order to solve the above problems, first of all, this paper established a tactile-based object attribute data set. Samples in the data set contain the elasticity and hardness characteristic information of objects, and each attribute is divided into 10 levels to describe the object more delicately. Compared with other tactile data sets in this field, the unique contribution of the tactile data set established in this paper is to solve the problem of lack of relevant data about the intensity difference of object attributes and to add information about the intensity difference of object attributes in the data set. The data set lays a foundation for more refined cognition and operation of objects through robot touch. Second, a method for identifying the attribute strength level of a convolutional neural network by using multi-scale features is proposed, which can identify the difference in the strength of elasticity and hardness of different objects in the data set. The main contribution of this method is that, in each feature extraction process, the multi-scale features of the original data are comprehensively superimposed, which makes the network understand the data more comprehensively. Compared with other methods, the unique advantage of this method is that the integrity and sufficiency of information are fully considered in each step of feature extraction, which means that this method is sensitive to data changes. The elasticity strength is regarded as the strength of the feedback to the experimenter when the object returns to its original state after deformation, and the hardness is regarded as the difficulty of the deformation of the object under the same force.
The structure of this paper is as follows: in Section 2, the latest developments related to this topic are introduced. In Section 3, the establishment process of the tactile data set is introduced. Section 4 introduces the recognition algorithm for the strength level of object attributes of a multi-scale convolutional neural network. Section 5 introduces the experimental results of the recognition algorithm on the tactile data set. Finally, Section 6 provides the discussion and Section 7 provides the conclusion.
2. Related Work
The robot’s tactile object recognition capability is achieved by processing haptic signals, which can be represented as a continuous signal, a set of discrete measurements, or a series of images [20]. How the data are processed may be affected by different data structures. This paper reviews the related studies on the use of haptic data to identify their attributes from the perspective of the data structure of haptic signals.
At present, the common haptic signals are force signals and vibration signals. If tactile sensors provide a pressure value, the best form of haptic signals may be a time-varying pressure curve, which can be used to detect contact or sliding events between contact surfaces [23,24]. Some studies consider the static information of the pressure images [25]. Although the object is pressed several times, each pressure image only contains the pressure distribution information caused by the shape of the object and does not contain the time relationship between the images [26]. Therefore, some studies have represented each image as a matrix of pressure readings at an instant, which can contain the physical properties of an object’s changes in information over time. For a haptic image, machine learning methods, such as the k-nearest neighbor method [19], Bayesian method [27], and the traditional method based on images [28,29], were used to identify features. However, this did not mean that these corresponding methods must be used in the extraction process of the haptic feature.
There has been some progress in deep learning using haptic data to identify objects. In [30,31], a tactile recognition system using a deep learning method was proposed. This system can recognize objects by grasping objects, but it cannot recognize the physical properties of objects. In [32], a deep learning multi-class and multi-label model was designed. The model can identify targets by learning four tactile features, including hardness, thermal conductivity, roughness, and texture from haptic images. Considering the good performance of convolutional neural network processing to extract spatial features, the use of convolutional neural networks to process tactile images is widely used [21,33]. In addition, some researchers use the advanced processing of haptic pressure images to classify and recognize objects [34]. For example, the pressure images obtained during extrusion and release are connected into a tensor that can be used to classify objects in 3D CNN.
Regardless of the form of the data structure of the haptic signal, one way to obtain richer information about the haptic signal is for humans to classify haptic samples using discrete categories that are more detailed than a binary label. The simplest partitioning task is to classify objects according to their similarity and select one or more dimensions for analysis. In [35], the results of the free sorting of different material samples were analyzed by multidimensional scaling, and the tactile material space was calibrated by the physical measurements of compressibility and roughness. Similarly, Ref. [36] discussed the main dimensions of tactile surface perception. Roughness, smoothness, and softness proved to be important orthogonal dimensions, and it was concluded that elasticity may correspond to the third main dimension. However, in a later study [37], the third major dimension of tactile perception was identified as viscosity or slippage. In addition, in [14,15], a number of different researchers were selected to give binary labels to adjectives of different objects, but the antisense relationships between adjectives were not taken into account. Therefore, Ref. [22] complements this aspect of the study by confirming the antonym pairs of hard/soft, rough/smooth, and cold/warm, and more tactile information than the binary label is analyzed.
3. Materials
The haptic data sets Penn Haptic Adjective Corpus-1 (PHAC-1) and Penn Haptic Adjective Corpus-2 (PHAC-2) are proposed in [14,15]. However, each volunteer was only required to give the binary label (yes/no) of each object attribute, but the binary label cannot judge the strength of the object attribute. Therefore, this article uses its own tactile data collection platform to establish a tactile data set of the strength levels of the object.
3.1. Robot Platform
As shown in Figure 1, the desktop Kinova robotic arm equipped with SynTouch’s NumaTac tactile sensor is selected as the haptic data acquisition platform in this paper.
The Kinova arm has seven degrees of freedom, including a manipulator and a two-finger gripper. Two NumaTac tactile sensors are mounted on each finger of the two-finger gripper. The original data collected include the DC pressure signal (PDC) at a sampling frequency of 100 Hz and the AC pressure vibration signal (PAC) at a sampling frequency of 2200 Hz. The DC pressure value and AC pressure value per unit area can be calculated with Formulas (1) and (2):
DC = (PDC − offset) × 12.94 Pa/bit(1)
AC = (PAC − offset) × 0.13 Pa/bit(2)
In Formula (1), offset is the DC pressure signal value of the NumaTac tactile sensor signal under atmospheric pressure, and PDC is the instantaneous DC pressure signal value obtained by the NumaTac tactile sensor when the two-finger gripper interacts with the object. The resolution of 12.94 is obtained from the official documentation of the NumaTac tactile sensor. In Formula (2), offset is the AC pressure vibration signal value of the NumaTac tactile sensor under atmospheric pressure. PAC is the instantaneous AC pressure vibration signal value obtained by the NumaTac tactile sensor when the two-finger gripper interacts with the object, and the resolution of 0.13 is obtained from the official document of the NumaTac tactile sensor. The rest of the relevant information about the NumaTac tactile sensor can be found on SynTouch’s official website.
The tactile data acquisition platform uses the robot operating system (ROS) as the software interface. ROS has a series of libraries and tools to help developers write robot software programs [38], which is an ROS node. When a complex task is completed through a series of programs, ROS creates a network that connects all nodes, which is an ROS graph. Through the interaction between the ROS graph, other nodes can obtain the information published.
In this paper, the nodes included in the ROS graph mainly include the Kinova node, the SynTouch node, and the process node. The registration, communication, and parameter server of the nodes are managed by the ROS master, as shown in Figure 2.
-
The ROS master is the core of ROS. It registers the names of nodes, services, and topics and maintains a parameter server.
-
The Kinova node plans a reasonable path to move the end of the manipulator to a fixed point by manipulating the Kinova manipulator, then manipulating the gripper to open and close the object. The gripper continues to close after contacting the object. The gripper stops closing when the fixed force threshold is reached.
-
The SynTouch node is used to publish the tactile data generated by the interaction between the tactile sensor and the object. During the execution of a complete exploration action to generate tactile data, the two tactile sensors following the two-finger gripper physically interact with the object to obtain the tactile sensor data and publish data to the ROS network at a frequency of 100 Hz.
-
The judge is used to judge whether the tactile sensor is in contact with the target object. When the ratio of the PAC signal value at the next moment to at the previous moment is greater than 1.1 or less than 0.9, it is considered that the initial touch occurs and the data will be published to the process node. The values 1.1 and 0.9 are experimentally measured thresholds that can determine whether the initial contact is generated.
-
The process node subscribes to the haptic data released by the SynTouch node and judges whether an initial contact occurs. After satisfying the conditions, it intercepts the data released by the SynTouch node and superimposes them into a dual-channel haptic sample. Since the Kinova robotic arm needs to reach the preset position before performing the exploration action, the NumaTac returns useless data in this process. In order to solve this problem, a sub-thread is established. First, it continuously receives tactile data in the main thread and judges whether the tactile sensor has initial contact with the object. When the initial contact is made, the sub-thread waits for 3 s until all the sample data in the main thread are received, where 3 is the experimentally measured time required to collect a complete sample. The data are then intercepted to create haptic samples.
3.2. NumaTac Haptic Dataset
This paper uses the abovementioned robot platform to physically interact with some common objects in daily life repeatedly in order to collect a large number of tactile samples. Then, we let four experimenters physically interact with the objects to provide the elasticity strength level and hardness strength level of each object as the object labels. The elasticity and hardness properties of each object are coupled together. These constitute the NumaTac haptic data set.
3.2.1. Objects
Considering the foam material on the surface of the NumaTac tactile sensor and the movement limitation of the Kinova, 30 different objects were selected after excluding the objects with dangerous characteristics, such as sharpness, high temperature, humidity, and unsuitable scale. Objects contain a variety of material properties representing a wide range of physical properties, as shown in Figure 3. The 30 objects contain a variety of shapes and materials, which can meet our needs to realize attribute strength difference recognition.
3.2.2. Data Collection Detail Settings
Humans can use a range of grasping movements when evaluating objects. Similarly, the motions of the Kinova to grasp objects in this paper are identified as squeeze, static grip, and release. Since the interaction between the Kinova and the object is a continuous process, the above Kinova actions are combined into a complete process to collect tactile data. The specific ROS graph is shown in Figure 2.
Kinova performs the abovementioned grasping action to obtain the haptic data of the object. By adjusting the position of the object, the object is kept on the mid-axis plane of the two fingers, and the slight posture change is set to simulate the uncertainty of touch. When the squeezing action occurs, the two fingers touch the object almost at the same time, preventing the object position offset from interfering with the two tactile sensors and reducing unnecessary changes between experiments.
During the gripping process, Kinova remains in a fixed position, and the whole process is just one grasping motion. The tiny robotic arm’s jitter information is ignored and only the signal generated by the tactile sensor is considered. When the grasping action begins, the gripper closes at a constant speed until the NumaTac makes initial contact with the object. At the beginning of the squeeze phase, the NumaTac continuously deforms to obtain tactile signals until it reaches the force threshold of 3493.8 pa, and the PDC value of one NumaTac reaches 270 bit (the offset is 238). The callback function in the node sends a stop command to the gripper. Since the signal transmission takes time, the gripper will continue to squeeze the object for a short period of time before receiving the stop command, and will continuously send data back. When the stage of squeezing the object ends, the gripper enters the stage of static gripping the object, the deformation of the NumaTac remains unchanged, and the tactile signal is continuously received for a period of time after the static grip phase is over. Then, the gripper opens at a constant speed and enters the release phase, where the NumaTac restores its initial shape and state and generates a tactile signal.
A total of 30 objects were grasped and each object was grasped 50 times by performing the above continuous grasping motion using a tactile data acquisition platform. All PDC and PAC were recorded.
3.2.3. Data Preprocessing
The time required for a complete data collection action is 3 s, and the frequency at which the NumaTac haptic data are returned to the ROS network is 100 Hz. Therefore, the length of each sample data is set to 300 discrete data points. Since the process is a continuous process, all initial samples are more than 300 discrete points in length. For samples with more than 300 discrete points, appropriate data segments are intercepted.
Considering the symmetry of the mechanical structure when the robot’s dexterous hand grasps the object and the integrity of the tactile information of the measured object, we increased the feature quantity of the sample’s tactile information, and superimposed the tactile data of the left and right fingers of each sample into a dual-channel sample.
(3)
In Formula (3), is the normalized data, x is the original data, μ is the mean over the feature dimension of x, and σ is the standard deviation in the feature dimension of x. In order to eliminate the initial value of the influence of atmospheric pressure on the tactile sensor, the normalization function in Formula (3) is used to eliminate the influence of the initial value of the tactile sensor when there is no object contact.
3.2.4. Labels
To match an object’s haptic data with its own real-world label, four experimenters were given eye patches and headphones, and they were instructed to use two fingers to simulate the grasping action of a gripper touching the object and compare the strength levels of the object’s properties. All experimenters were asked to evaluate the elasticity strength level and the hardness strength level separately for each object, and the levels were limited to 10 levels from 1 to 10. The label dimension of each object is 1 × 20. The first 10 grades represent elasticity, and the last 10 grades represent hardness. For example, the label of a mineral water bottle is (0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0) in Figure 4.
Regarding the question of whether the data set is balanced, since the focus is on the properties of the elasticity and hardness of the object, each sample contains these two kinds of information. The data set has the same sample size of each strength level of objects’ tactile attributes, and each strength level has 150 samples, so the data set is a balanced data set.
For the problem of different strength levels given by different experimenters, the experimenters can vote on all the strength levels given, and the experimenters could agree with their own or others’ judgment of the object. This process is completely decided by the internal discussion of the four experimenters, without the influence of other external personnel. The experimenters are required to give unique elasticity and hardness strength levels for each object that will not be changed.
4. Method
4.1. Multi-Scale Convolutional Neural Network Architecture
The recognition of the tactile attribute strength level of objects is an important issue. One way to obtain rich tactile information with human perception is to use more discretized categories to represent the similarity of object attributes and the difference of attribute strength levels. The feature extraction of tactile information is the basis of object attribute strength level recognition. Considering the good ability of CNN feature extraction, this work used a multi-scale convolutional neural network to extract the tactile information features of objects. The network structure is shown in Figure 5.
The multi-scale convolutional neural network consists of an input layer, a block module, and a fully connected module. The block feature extraction module contains three scales: the original feature map, single-channel feature map, and multi-channel feature map.
-
The original feature map: the obtained feature map is subjected to a max-pooling operation with a kernel size of two to correspond to the size of the other two convolutional layers, and no convolution operation is performed to preserve the original information.
-
The single-channel feature map: the output of the previous step is subjected to a convolution operation through conv1_1, conv2_1, and conv3_1, and a max-pooling operation with a kernel size of two is performed, and then multiplied by a weight coefficient of a convolutional layer. The is used for the weight coefficient of training single-channel output features in the network.
-
The multi-channel feature map: the output feature map of the previous step is subjected to a convolution operation through conv1_2, conv2_2, and conv3_2, and a max-pooling operation with a kernel size of two is performed, and then multiplied by the weight coefficient . The is the weight coefficient of training multi-channel feature maps in the network.
4.2. Multi-Scale Convolutional Neural Network Parameters
Before starting training, the parameters of the convolutional layer and the fully connected layer are initialized to a normal distribution, with an average value of 0 and a standard deviation of 1. The specific parameters and outputs of each layer of the multi-scale convolutional neural network are shown in Table 1.
4.3. Calculation Process
In the process of the forward propagation and back propagation of the multi-scale convolutional neural network, the specific operation process is shown in Formulas (4)~(7).
(4)
Formula (4) is the sigmoid activation function used in the neural network, and x is the input feature data of the activation function.
(5)
In Formula (5), is the feature tensor value of the latent space output by the i + 1 block module, and σ is the sigmoid activation function. Max-pooling is the dimensionality reduction operation, is the weight parameter of the single-channel feature map, and is the neuron weight parameter value contained in the convolutional single-channel feature layer in the i block module. is the neuron bias parameter value contained in the convolutional single-channel feature layer in the i block module. is the weight parameter of the convolutional multi-channel feature layer. is the neuron weight parameter value contained in the convolutional multi-channel feature layer in the i block module, and is the neuron bias parameter value contained in the convolutional multi-channel feature layer in the i block module. High-dimensional latent features are concatenated in the channel dimension by using “cat”.
(6)
In Formula (6), is the output feature tensor value of the fully connected layer of j + 1 in the fc module. are the weight parameter and bias parameter of the neuron of the j + 1 layer in the fc module, respectively, and is the final output value of the network.
(7)
In Formula (7), is the true label value of n samples, is the network prediction value, and σ is the sigmoid activation function.
In the process of backpropagation, the gradient descent method is used to find the partial derivative of the loss function to w, b, and k, which are ,, and . W, b, and k are optimized by using the Adam optimizer with a learning rate of 0.001.
5. Experiment and Results
5.1. Object Attribute Strength Level Recognition Experiment
In this paper, PyTorch, a deep learning library based on Python, is used to build a neural network, and all haptic samples in the NumaTac haptic data set are trained and tested. For some deep networks, as resnet-152, the 23 in the data length (23 × 300 × 2) cannot carry out a very deep convolution operation. Therefore, considering the existing networks and our own needs, the multi-scale convolutional neural network is compared with a variety of models, which are mnasnet1_0 [39], resnet18 [40], shufflenet_v2 [41], and mobilenet_v2 [42]. The four models used in this paper are mature and advanced. They have been tested and the rationality of their network structure fully proved. The different evaluation criteria of accuracy, precision, recall, and f1_score are shown in Table 2.
5.2. Analysis of Results
In Table 2, the effects of the five models under the four evaluation criteria on the elasticity and hardness strength levels of 1–10 are counted. The four evaluation criteria are accuracy, precision, recall, and f1_score.
5.2.1. Accuracy
The accuracy is the ratio of the number of correctly classified samples to the total number of samples. It represents the overall prediction degree. Only when the actual value is completely consistent with the predicted output is the predicted output considered to be accurate. As can be seen from Table 2, the accuracy of the multi-scale convolutional neural network with regard to the elasticity strength level of objects is 9%, 4%, 5%, and 1%, respectively, higher than mnasnet1_0, resnet18, shufflenet_v2, and mobilenet_v2. The accuracy of the multi-scale convolutional neural network concerning the hardness strength level of objects is 14%, 14%, 10%, and 5%, respectively, higher than mnasnet1_0, resnet18, shufflenet_v2, and mobilenet_v2. This means that the difference of the multi-scale convolutional neural network between the output and the real value is the smallest, and the prediction effect of the multi-scale convolutional neural network is better in the object prediction process.
5.2.2. Precision
The precision is the ratio of all the predicted positive samples to actually positive samples in the prediction result. Table 2 shows the precision of the sample strength level of all models. In fact, the worst precision of the multi-scale convolutional neural network only achieves 80% at elasticity strength level and 87% at hardness strength level, but it is still higher than other models, that is, the precision of mansnet1_0 at the ninth elasticity strength level is 35%, and the precision at the third hardness strength level is 35%. The precision of resnet18 is 65% at the second elasticity strength level and 39% at the seventh hardness strength level. The precision of shufflenet_v2 at the first elasticity strength level is 76%, and the precision at the second hardness strength level is 63%. The precision of mobilenet_v2 at the third elasticity strength level is 71%, and the precision at the sixth hardness strength level is 68%. The above proves that the prediction of the multi-scale convolutional neural network for all strength levels is more stable and can be maintained at a higher recognition rate.
5.2.3. Recall
The recall is the ratio of the actual positive samples to predicted positive samples. As can be seen from Table 2, the worst recall of the multi-scale convolutional neural network is 74% at the third elasticity strength level and 81% at the eighth hardness strength level. Meanwhile, the worst recall of mansnet1_0 at the second elasticity strength level is 77%, and at the eighth and ninth hardness strength level is 75%. Resnet18 has a worst recall of 73% at the tenth elasticity strength level and 75% at the third hardness strength level. Shufflenet_v2 has a worst recall of 71% at the eighth elasticity strength level and 75% at the seventh hardness strength level. The worst recall of mobilenet_v2 at the fourth elasticity strength level is 69% and, at the sixth hardness strength level, is 77%. The recognition rate of the other nine levels of the multi-scale convolutional neural network remains at a higher level than other models.
5.2.4. F1_Score
F1_score is an indicator that comprehensively considers the precision and the recall. Therefore, in order to measure the quality of the model in a more comprehensive and balanced manner, the quality of the model is evaluated by the harmonic average of the precision and the recall. As can be seen from Table 2, f1_score of the multi-scale convolutional neural network is 98% at the fifth elasticity strength level and 99% at the tenth hardness strength level, which are higher than the other models. Meanwhile, the f1_score of mansnet1_0 is 89% at the fourth elasticity strength level and 91% at the seventh hardness strength level. The f1_score of resnet18 is 94% at the first elasticity strength level and 89% at the ninth hardness strength level. The shufflenet_v2 has an f1_score of 89% at the fourth elasticity strength level and 94% at the ninth hardness strength level. The f1_score of mobilenet_v2 is 97% at the fifth elasticity strength level and 97% at the tenth hardness strength level. The above proved that the multi-scale convolutional neural network’s predictions for all strength levels remain at a higher recognition rate.
To summarize, compared with other models, the multi-scale convolutional neural network is more accurate and it has a significant improvement in terms of the recognition rate. From the point of view of precision, in the prediction results of the positive examples in the predicted value, the true value has a higher proportion of positive examples. The prediction effect of positive examples in the true value is better from the perspective of the recall. Finally, the performance of the recognition effect is more stable and the recognition rate is higher from the perspective of the f1_score. It can be concluded that the recognition of attribute strength level of the multi-scale convolutional neural network performs better.
6. Discussion
There are abundant tactile sensors on the human palm, so it is an interesting idea to carry tactile sensors on the robotic arm. By designing the control and algorithm of the robotic arm, the robot’s intelligent perception level of the target can be enhanced.
The purpose of this paper is to enable robots to understand the property strength levels of objects in a more detailed manner. To this end, this paper uses the Kinova manipulator and the NumaTac tactile sensor to establish a data set of tactile properties of objects. The signals in the tactile samples of objects are force signals and vibration signals that change over time. These two signals can reflect the properties of elasticity and hardness of objects. The data set is a balanced data set for the two properties of sample elasticity and strength.
In this paper, a multi-scale convolutional neural network method for recognizing the attribute strength levels of objects is proposed and, in order to verify the effectiveness of the method, the method was trained and tested using a haptic data set. Compared with the comparison model, experiments show that this method can better extract the tactile data features of objects and can identify the elasticity strength level and hardness strength level of the objects. We believe that improving the speed of multi-scale convolutional neural networks is a challenge. In the current research trend, the speed of the network is a key issue, which is also the focus of our next research. On the basis of this work, using robots to identify objects online is a good application direction, and this part of the work may be carried out in future studies.
7. Conclusions
In this paper, a robotic arm platform is used. The platform includes a robotic arm and a robotic gripper. Each of the two fingers on the gripper is equipped with a NumaTac tactile sensor. Through this operating platform, we established a tactile data set of object attribute strength levels, selected two properties of the object (elasticity and hardness) for analysis, and proposed a multi-scale CNN to identify the object’s elasticity strength level and hardness strength level. The experimental results show that the network performs better in terms of prediction accuracy and prediction stability.
In future work, we propose to focus on the relationship between tactile dimensions and a more delicate tactile attribute strength level classification method to identify the objects’ strength levels. Besides, compared with some well-known data sets, our data set is not large and we plan to expand the data set in our future work. Finally, how to apply it to practical robotics applications, such as product identification in the industrial field, is also a problem that should be considered.
Conceptualization, G.Y.; methodology, G.Y.; software, G.Y.; validation, G.Y.; formal analysis, G.Y. and X.W.; investigation, G.Y.; resources, P.Z. and D.S.; data curation, G.Y.; writing—original draft preparation, G.Y.; writing—review and editing, P.Z., D.S., Z.C., and X.W.; visualization, G.Y.; supervision, P.Z. and D.S. All authors have read and agreed to the published version of the manuscript.
The work is supported in part by Shandong Province’s Major Scientific and Technological Innovation Project, Grant No. 2019JZZY010444 and No. 2019TSLH0315, the Project of 20 Policies to Facilitate Scientific Research at Jinan Colleges, Grant No. 2019GXRC063, and the Natural Science Foundation of Shandong Province of China, Grant No. ZR2020MF138.
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 3. A total of 30 objects in the NumaTac haptic data set (from left to right and from top to bottom; the objects are a towel, garbage bag, paper cup, medium draw paper, earphone box, bubble wrap, mineral water bottle, hand cream, square sponge, jelly, glue, mask, tennis, double layer foam, can, square foam, ham sausage, triangular bandage, metal column, rubber, black bandage, book, glass bottle, orange, plastic box, dental cylinder box, small draw paper, white thread group, foam ball, and soap box).
Figure 4. Horizontal graph of the elasticity and hardness strength level of 30 objects in the NumaTac haptic data set.
Parameters of each layer of the multi-scale convolutional neural network.
Layers | Parameters | Next Operation | Output | |
---|---|---|---|---|
Size, Stride, Padding | ||||
Input | Input |
2 × 2, 0, 0 | 23 × 300 × 2 |
|
Conv1_1 | conv |
3 × 3 × 1, 1, 1 |
BN | 23 × 300 × 1 |
Conv1_2 | conv |
3 × 3 × 5, 1, 1 |
BN | 23 × 300 × 5 |
Cat_layer1 | Leaky_relu | 11 × 150 × 8 | ||
Conv2_1 | conv |
3 × 3 × 1, 1, 1 |
BN | 11 × 150 × 1 |
Conv2_2 | conv |
3 × 3 × 15, 1, 1 |
BN | 11 × 150 × 15 |
Cat_layer2 | Leaky_relu | 5 × 75 × 24 | ||
Conv3_1 | conv |
3 × 3 × 1, 1, 1 |
BN | 5 × 75 × 1 |
Conv3_2 | conv |
3 × 3 × 32, 1, 1 |
BN | 5 × 75 × 32 |
Cat_layer3 | Leaky_relu | 2 × 37 × 57 | ||
Fc1 | (4218,512) | BN | 512 | |
Fc2 | (512,20) | BN | 20 |
Accuracy, precision, recall, and f1_score of elasticity and hardness strength levels of the five models, and the best or worst result shown in bold.
Net | Level | Elasticity Level_Score (%) | Hardness Level_Score (%) | ||||||
---|---|---|---|---|---|---|---|---|---|
accu | prec | reca | f1_score | accu | prec | reca | f1_score | ||
Multi-scale convolutional neural network | 1 | 90 | 95 | 77 | 85 | 95 | 95 | 89 | 92 |
2 | 85 | 98 | 91 | 87 | 97 | 92 | |||
3 | 80 | 74 | 77 | 93 | 95 | 94 | |||
4 | 85 | 89 | 87 | 93 | 99 | 95 | |||
5 | 98 | 99 | 98 | 96 | 100 | 98 | |||
6 | 98 | 82 | 89 | 99 | 95 | 97 | |||
7 | 94 | 96 | 95 | 95 | 96 | 95 | |||
8 | 82 | 97 | 89 | 98 | 81 | 89 | |||
9 | 93 | 95 | 94 | 97 | 99 | 98 | |||
10 | 95 | 94 | 94 | 99 | 99 | 99 | |||
Mnasnet1_0 | 1 | 81 | 98 | 79 | 87 | 81 | 99 | 83 | 91 |
2 | 98 | 77 | 87 | 100 | 81 | 90 | |||
3 | 95 | 78 | 86 | 35 | 99 | 52 | |||
4 | 98 | 81 | 89 | 99 | 82 | 90 | |||
5 | 99 | 79 | 88 | 99 | 76 | 86 | |||
6 | 99 | 79 | 88 | 100 | 78 | 88 | |||
7 | 99 | 77 | 86 | 99 | 83 | 91 | |||
8 | 99 | 78 | 87 | 98 | 75 | 85 | |||
9 | 35 | 99 | 52 | 100 | 75 | 85 | |||
10 | 100 | 80 | 89 | 99 | 81 | 89 | |||
Resnet18 | 1 | 86 | 98 | 91 | 94 | 81 | 100 | 76 | 86 |
2 | 65 | 99 | 78 | 83 | 89 | 86 | |||
3 | 94 | 75 | 83 | 93 | 75 | 83 | |||
4 | 83 | 91 | 87 | 96 | 77 | 86 | |||
5 | 80 | 96 | 87 | 98 | 79 | 87 | |||
6 | 99 | 81 | 89 | 100 | 77 | 87 | |||
7 | 93 | 83 | 87 | 39 | 99 | 56 | |||
8 | 80 | 84 | 82 | 100 | 79 | 88 | |||
9 | 87 | 82 | 85 | 100 | 81 | 89 | |||
10 | 100 | 73 | 85 | 98 | 81 | 88 | |||
Shufflenet_v2 | 1 | 85 | 76 | 81 | 78 | 85 | 88 | 86 | 87 |
2 | 85 | 93 | 89 | 63 | 90 | 74 | |||
3 | 90 | 81 | 86 | 90 | 81 | 85 | |||
4 | 88 | 91 | 89 | 86 | 82 | 84 | |||
5 | 85 | 89 | 87 | 86 | 93 | 90 | |||
6 | 88 | 87 | 88 | 94 | 85 | 90 | |||
7 | 82 | 84 | 83 | 83 | 75 | 79 | |||
8 | 79 | 71 | 75 | 89 | 82 | 85 | |||
9 | 91 | 83 | 87 | 93 | 95 | 94 | |||
10 | 85 | 88 | 86 | 93 | 85 | 89 | |||
Mobilenet_v2 | 1 | 89 | 96 | 87 | 91 | 90 | 90 | 95 | 92 |
2 | 97 | 95 | 96 | 99 | 91 | 95 | |||
3 | 71 | 95 | 81 | 96 | 77 | 86 | |||
4 | 97 | 69 | 83 | 99 | 88 | 93 | |||
5 | 97 | 97 | 97 | 97 | 85 | 90 | |||
6 | 99 | 83 | 91 | 76 | 95 | 85 | |||
7 | 97 | 90 | 93 | 68 | 97 | 80 | |||
8 | 73 | 92 | 81 | 97 | 93 | 95 | |||
9 | 94 | 90 | 92 | 98 | 87 | 92 | |||
10 | 97 | 93 | 95 | 99 | 95 | 97 |
References
1. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. ImageNet: A large-scale hierarchical image database. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition; Miami, FL, USA, 20–25 June 2009; pp. 248-255.
2. Bai, Z.X.; Zhang, X.L. Speaker recognition based on deep learning: An overview. Neural Netw.; 2021; 140, pp. 65-99. [DOI: https://dx.doi.org/10.1016/j.neunet.2021.03.004] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33744714]
3. Li, Q.; Kroemer, O.; Su, Z.; Veiga, F.F.; Kaboli, M.; Ritter, H.J. A Review of Tactile Information: Perception and Action through Touch. IEEE Trans. Robot.; 2020; 36, pp. 1619-1634. [DOI: https://dx.doi.org/10.1109/TRO.2020.3003230]
4. Lederman, S.J.; Klatzky, R.L. Extracting object properties through haptic exploration. Acta Psychol.; 1993; 84, pp. 29-40. [DOI: https://dx.doi.org/10.1016/0001-6918(93)90070-8]
5. Komeno, N.; Matsubara, T. Tactile perception based on injected vibration in soft sensor. IEEE Robot. Autom. Lett.; 2021; 6, pp. 5365-5372. [DOI: https://dx.doi.org/10.1109/LRA.2021.3075664]
6. Zhang, Y.; Kan, Z.; Tse, A.Y.; Yang, Y.; Wang, M.Y. Fingervision tactile sensor design and slip detection using convolutional lstm network. arXiv; 2018; arXiv: 1810.02653
7. Siegel, D.; Garabieta, I.; Hollerbach, J.M. An integrated tactile and thermal sensor. Proceedings of the IEEE International Conference on Robotics & Automation; San Francisco, CA, USA, 7–10 April 1986; pp. 1286-1291.
8. Tsuji, S.; Kohama, T. Using a convolutional neural network to construct a pen-type tactile sensor system for roughness recognition. Sens. Actuators A-Phys.; 2019; 291, pp. 7-12. [DOI: https://dx.doi.org/10.1016/j.sna.2019.02.032]
9. Kaboli, M.; Cheng, G. Robust Tactile Descriptors for Discriminating Objects from Textural Properties via Artificial Robotic Skin. IEEE Trans. Robot.; 2018; 34, pp. 985-1003. [DOI: https://dx.doi.org/10.1109/TRO.2018.2830364]
10. Fishel, J.A.; Loeb, G.E. Bayesian exploration for intelligent identification of textures. Front. Neurorobotics; 2012; 6, 4. [DOI: https://dx.doi.org/10.3389/fnbot.2012.00004]
11. Spiers, A.J.; Liarokapis, M.V.; Calli, B.; Dollar, A.M. Single-Grasp Object Classification and Feature Extraction with Simple Robot Hands and Tactile Sensors. IEEE Trans. Haptics; 2016; 9, pp. 207-220. [DOI: https://dx.doi.org/10.1109/TOH.2016.2521378]
12. Lepora, N.F. Biomimetic Active Touch with Fingertips and Whiskers. IEEE Trans. Haptics; 2016; 9, pp. 170-183. [DOI: https://dx.doi.org/10.1109/TOH.2016.2558180]
13. Okamura, A.M.; Cutkosky, M.R. Feature detection for haptic exploration with robotic fingers. Int. J. Robot. Res.; 2001; 20, pp. 925-938. [DOI: https://dx.doi.org/10.1177/02783640122068191]
14. Chu, V.; McMahon, I.; Riano, L.; McDonald, C.G.; He, Q.; Perez-Tejada, J.M.; Arrigo, M.; Fitter, N.; Nappo, J.C.; Darrell, T. et al. Using robotic exploratory procedures to learn the meaning of haptic adjectives. Proceedings of the 2013 IEEE International Conference on Robotics and Automation; Karlsruhe, Germany, 6–10 May 2013; pp. 3048-3055.
15. Chu, V.; McMahon, I.; Riano, L.; McDonald, C.G.; He, Q.; Perez-Tejada, J.M.; Arrigo, M.; Darrell, T.; Kuchenbecker, K.J. Robotic learning of haptic adjectives through physical interaction. Robot. Auton. Syst.; 2015; 63, pp. 279-292. [DOI: https://dx.doi.org/10.1016/j.robot.2014.09.021]
16. Okatani, T.; Takahashi, H.; Noda, K.; Takahata, T.; Matsumoto, K.; Shimoyama, I. A Tactile Sensor Using Piezoresistive Beams for Detection of the Coefficient of Static Friction. Sensors; 2016; 16, 718. [DOI: https://dx.doi.org/10.3390/s16050718] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/27213374]
17. Lin, W.J.; Lo, S.H.; Young, H.-T.; Hung, C.-L. Evaluation of Deep Learning Neural Networks for Surface Roughness Prediction Using Vibration Signal Analysis. Appl. Sci.; 2019; 9, 1462. [DOI: https://dx.doi.org/10.3390/app9071462]
18. Sinapov, J.; Sukhoy, V.; Sahai, R.; Stoytchev, A. Vibrotactile Recognition and Categorization of Surfaces by a Humanoid Robot. IEEE Trans. Robot.; 2011; 27, pp. 488-497. [DOI: https://dx.doi.org/10.1109/TRO.2011.2127130]
19. Feng, J.H.; Jiang, Q. Slip and roughness detection of robotic fingertip based on FBG. Sens. Actuators A-Phys.; 2019; 287, pp. 143-149. [DOI: https://dx.doi.org/10.1016/j.sna.2019.01.018]
20. Gandarias, J.M.; Gómez-de-Gabriel, J.M.; García-Cerezo, A.J. Enhancing perception with tactile object recognition in adaptive grippers for human-robot interaction. Sensors; 2018; 18, 692. [DOI: https://dx.doi.org/10.3390/s18030692]
21. Lu, J. Multi-Task CNN Model for Attribute Prediction. IEEE Trans. Multimed.; 2015; 17, pp. 1949-1959.
22. Richardson, B.A.; Kuchenbecker, K.J. Learning to predict perceptual distributions of haptic adjectives. Front. Neurorobotics; 2020; 13, 116. [DOI: https://dx.doi.org/10.3389/fnbot.2019.00116]
23. Melchiorri, C. Slip detection and control using tactile and force sensors. IEEE/ASME Trans. Mechatron.; 2000; 5, pp. 235-243. [DOI: https://dx.doi.org/10.1109/3516.868914]
24. Su, Z.; Hausman, K.; Chebotar, Y.; Molchanov, A.; Loeb, G.E.; Sukhatme, G.S.; Schaal, S. Force Estimation and slip detection/classification for grip control using a biomimetic tactile sensor. Proceedings of the 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids); Seoul, Korea, 3–5 November 2015; pp. 297-303.
25. Luo, S.; Mou, W.X.; Althoefer, K.; Liu, H. Novel Tactile-SIFT Descriptor for Object Shape Recognition. IEEE Sens. J.; 2015; 15, pp. 5001-5009. [DOI: https://dx.doi.org/10.1109/JSEN.2015.2432127]
26. Gandarias, J.M.; Gomez-de-Gabriel, J.M.; Garcia-Cerezo, A. Human and Object recognition with a high-resolution tactile sensor. Proceedings of the 2017 IEEE Sensors; Glasgow, UK, 29 October–1 November 2017; pp. 981-983.
27. Pastor, F.; Garcia-Gonzalez, J.; Gandarias, J.M.; Medina, D.; Closas, P.; Garcia-Cerezo, A.J.; Gomez-De-Gabriel, J.M. Bayesian and Neural Inference on LSTM-Based Object Recognition from Tactile and Kinesthetic Information. IEEE Robot. Autom. Lett.; 2021; 6, pp. 231-238. [DOI: https://dx.doi.org/10.1109/LRA.2020.3038377]
28. Okamura, A.M.; Turner, M.L.; Cutkosky, M.R. Haptic exploration of objects with rolling and sliding. Proceedings of the International Conference on Robotics and Automation; Albuquerque, NM, USA, 25–25 April 1997; pp. 2485-2490.
29. Dong, S.Y.; Yuan, W.Z.; Adelson, E.H. Improved gelsight tactile sensor for measuring geometry and slip. Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Vancouver, BC, Canada, 24–28 September 2017; pp. 137-144.
30. Schmitz, A.; Bansho, Y.; Noda, K.; Iwata, H.; Ogata, T.; Sugano, S. Tactile object recognition using deep learning and dropout. Proceedings of the 2014 IEEE-RAS International Conference on Humanoid Robots; Madrid, Spain, 18–20 November 2014; pp. 1044-1050.
31. Yuan, W.; Zhu, C.; Owens, A.; Srinivasan, M.A.; Adelson, E.H. Shape-independent hardness estimation using deep learning and a GelSight tactile sensor. Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA); Singapore, 29 May–3 June 2017; pp. 951-958.
32. Han, D.; Nie, H.; Chen, J.; Chen, M.; Deng, Z.; Zhang, J. Multi-modal haptic image recognition based on deep learning. Sens. Rev.; 2018; 38, pp. 486-493. [DOI: https://dx.doi.org/10.1108/SR-08-2017-0160]
33. Hui, W.; Li, H. Robotic tactile recognition and adaptive grasping control based on CNN-LSTM. Yi Qi Yi Biao Xue Bao/Chin. J. Sci. Instrum.; 2019; 40, pp. 211-218.
34. Pastor, F.; Gandarias, J.M.; García-Cerezo, A.J.; Gómez-De-Gabriel, J.M. Using 3D convolutional neural networks for tactile object recognition with robotic palpation. Sensors; 2019; 19, 5356. [DOI: https://dx.doi.org/10.3390/s19245356]
35. Tiest, W.M.B.; Kappers, A.M.L. Analysis of haptic perception of materials by multidimensional scaling and physical measurements of roughness and compressibility. Acta Psychol.; 2006; 121, pp. 1-20. [DOI: https://dx.doi.org/10.1016/j.actpsy.2005.04.005] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/16055070]
36. Holliins, M.; Faldowski, R.; Rao, S.; Young, F. Perceptual dimensions of tactile surface texture: A multidimensional scaling analysis. Percept. Psychophys.; 1993; 54, pp. 697-705. [DOI: https://dx.doi.org/10.3758/BF03211795]
37. Hollins, M.; Bensmaïa, S.; Karlof, K.; Young, F. Individual differences in perceptual space for tactile textures: Evidence from multidimensional scaling. Percept. Psychophys.; 2000; 62, pp. 1534-1544. [DOI: https://dx.doi.org/10.3758/BF03212154]
38. Zhang, Z.H.; Mei, X.S.; Bian, X.; Cai, H.; Ti, J. Development of an intelligent interaction service robot using ROS. Proceedings of the 2016 IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC); Xi’an, China, 3–5 October 2016; pp. 1738-1742.
39. Tan, M.X.; Chen, B.; Pang, R.M.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-aware neural architecture search for mobile. Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); Long Beach, CA, USA, 15–20 June 2019; pp. 2815-2823.
40. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA, 27–30 June 2016; pp. 770-778.
41. Ma, N.N.; Zhang, X.Y.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. Proceedings of the European Conference on Computer Vision—ECCV; Munich, Germany, 8–14 September 2018; pp. 122-138.
42. Sandler, M.; Howard, A.; Zhu, M.L. MobileNetV2: Inverted residuals and linear bottlenecks. Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510-4520.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In order to solve the problem in which most currently existing research focuses on the binary tactile attributes of objects and ignores identifying the strength level of tactile attributes, this paper establishes a tactile data set of the strength level of objects’ elasticity and hardness attributes to make up for the lack of relevant data, and proposes a multi-scale convolutional neural network to identify the strength level of object attributes. The network recognizes the different attributes and identifies differences in the strength level of the same object attributes by fusing the original features, i.e., the single-channel features and multi-channel features of the data. A variety of evaluation methods were used for comparison with multiple models in terms of strength levels of elasticity and hardness. The results show that our network has a more significant effect in accuracy. In the prediction results of the positive examples in the predicted value, the true value has a higher proportion of positive examples, that is, the precision is better. The prediction effect for the positive examples in the true value is better, that is, the recall is better. Finally, the recognition rate for all classes is higher in terms of f1_score. For the overall sample, the prediction of the multi-scale convolutional neural network has a higher recognition rate and the network’s ability to recognize each strength level is more stable.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 School of Electrical Engineering and Automation, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China;
2 School of Mechanical Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China;
3 School of Control Science and Engineering, Shandong University, Jinan 250061, China;