Next Article in Journal
Identification of Epileptic EEG Signals Using Convolutional Neural Networks
Next Article in Special Issue
LSTM and Bat-Based RUSBoost Approach for Electricity Theft Detection
Previous Article in Journal
The Influence of Wave Nonlinearity on Cross-Shore Sediment Transport in Coastal Zone: Experimental Investigations
Previous Article in Special Issue
A Survey on Theories and Applications for Self-Driving Cars Based on Deep Learning Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Simpler Learning of Robotic Manipulation of Clothing by Utilizing DIY Smart Textile Technology

IDLab-AIRO, Electronics and Information Systems Department, Ghent University, 9052 Zwijnaarde, Belgium
*
Author to whom correspondence should be addressed.
Submission received: 19 May 2020 / Revised: 3 June 2020 / Accepted: 11 June 2020 / Published: 13 June 2020
(This article belongs to the Special Issue Artificial Intelligence for Smart Systems)

Abstract

:

Featured Application

Learned robotic manipulation skills for clothing on a low-cost robot by building a DIY smart textile and using a single CPU core.

Abstract

Deformable objects such as ropes, wires, and clothing are omnipresent in society and industry but are little researched in robotics research. This is due to the infinite amount of possible state configurations caused by the deformations of the deformable object. Engineered approaches try to cope with this by implementing highly complex operations in order to estimate the state of the deformable object. This complexity can be circumvented by utilizing learning-based approaches, such as reinforcement learning, which can deal with the intrinsic high-dimensional state space of deformable objects. However, the reward function in reinforcement learning needs to measure the state configuration of the highly deformable object. Vision-based reward functions are difficult to implement, given the high dimensionality of the state and complex dynamic behavior. In this work, we propose the consideration of concepts beyond vision and incorporate other modalities which can be extracted from deformable objects. By integrating tactile sensor cells into a textile piece, proprioceptive capabilities are gained that are valuable as they provide a reward function to a reinforcement learning agent. We demonstrate on a low-cost dual robotic arm setup that a physical agent can learn on a single CPU core to fold a rectangular patch of textile in the real world based on a learned reward function from tactile information.

1. Introduction

Industry and society benefit greatly from autonomous robots that are able to perform manipulation skills on complex objects. Recent work shows that it is possible to learn grasping [1,2] and manipulation [3,4] skills on everyday objects. Although prior work has been undertaken [1] including deformable objects in the set of daily household objects, most robotic manipulation tasks are limited to objects with considerable stiffness, such as rubber ducks. Highly deformable objects—for example, cables or clothes—are often left out of the set due to the many challenges associated with handling the infinite amount of configurations these objects can be in [5].
Prior work [6,7,8] has focused on scripting solutions for the robotic manipulation of deformable objects. However, the engineering effort to make this work is considerable, and the designed solutions often break down when the environment changes. Recent work [9,10] has leveraged learning methods that are able to generalize to a wider range of deformable objects. Often, these methods require the definition of the task in simulation in order to generate rollouts in the environment. Unfortunately, simulating cloth is both challenging and computationally expensive. Furthermore, there is no guarantee that the acquired skills in simulation are transferable to the real world.
A typical approach for the robotic folding of textile relies on the use of vision in order to detect grasping points and to perform texture segmentation and pose estimation [7,8,11], as well to estimate the state [9] and—in the case of reinforcement learning (RL)-based approaches—the reward function [12]. However, while vision is instrumental for recognizing and localizing objects, touch and force measurements become important once contact occurs, and the object is explored using the end-effector [13]. Recent work [14] has illustrated this idea by applying a touch-based control method for a ball repositioning task, rolling a dice and the deflection of a stick. In line with other authors [14,15], we believe the use of tactile input and its fusion with other sensory information is crucial to learn complex robot manipulation tasks in an efficient way.
In this work, we deviate from the wider focus on using vision for state and pose estimation. Instead, we propose a cost-effective solution for the robotic folding of textile. Our approach is based on the use of the reinforcement learning of a shallow neural controller for a low-cost setup of a dual robotic arm, visible in Figure 1. Instead of engineering the reward function, we developed a smart textile piece that is able to detect whether it has been folded.

2. Related Work

When calculating the required trajectories for deformable object manipulation, the deformations need to be taken into account [5]. One approach to incorporating this is by using simulation. In [16], for example, deformations of a linear flexible object are approximated in a simulation in order to generate trajectories for folding in the air. In the case of clothing, the prior work presented in [7,8] dealt with the high-dimensional clothing state by building folding processes in which grasping points are manually chosen so that wrinkles are minimized. Next, the executed manipulation trajectories are hard-coded or calculated by matching polygonal models to contours. However, these approaches can accumulate errors across stages. For example, the type of textile can be mislabelled early in the process, leading to subsequent incorrect pipeline operations [8]. Additionally, engineered robotic pipelines for the folding task can take up to 24 min to execute [7], with the grasp point detection phase being the largest bottleneck. In [17], a faster process was implemented which could grasp and classify clothing in one minute. There were two overhead cameras in their setup, but the view was occluded when the arm engages. Consequently, the grasping procedure had to be executed repeatedly at different heights in order to detect whether an isolated garment was grasped.
To avoid highly-engineered pipelines and the associated loss of information between the stages, more recent methods utilize learning methods in order to grasp or manipulate deformable objects. For example, in [9], deep reinforcement learning is used to train a neural network to learn to manipulate a towel. The controller is trained in simulation and transferred to the real platform using domain randomization. In their approach, the agent learns from the pixel input only, while our method leverages tactile sensing to attain a more informational representation of the cloth. In [18], a forward dynamics model of the cloth using a latent space representation is learned. This architecture is used in [19] to bring cloth into the desired configuration. However, transferring this to the real world is challenging due to the difficulty in generating high-fidelity rollouts in simulation [9,19]. In our work, the policy of the robot is learned to avoid sim-to-real transfer issues. Other work tries to overcome the many training demonstrations required for training by utilizing learning from demonstration methods in order to speed up the training process [20,21]. This was illustrated in [22], who used teleoperated demonstrations to teach a low-cost robotic setup multiple tasks. However, general-purpose, dual-armed robots typically contain many degrees of freedom (e.g., the humanoid dual-arm Baxter® robot contains seven degrees of freedom per arm [23]), making teleoperation difficult.
Defining a reward function for the robotic folding task is difficult because it requires the state estimation of the cloth. Previous work [8,24] has extracted the contour of the textile using color segmentation. In [20], a marked towel is tracked, allowing the calculation of the distance between points on the towel from the training sample and the example demonstrations so that it can be used as a reward for the agent. Their method requires prior information about the shape of the object in order to reconstruct the missing market points. More recent work in using (deep) reinforcement learning for robotic folding also used vision-based methods to define the reward function for the agent [9,12]. However, relying solely on visual inputs and marker clues does not scale well. Another approach to find a reward function was explored in [25,26], where inverse reinforcement learning (IRL) was used to learn a reward function from expert demonstrations. Because IRL learns a reward function and a policy separately, it requires RL to be run in an inner loop, making it computationally expensive to train. Our approach learns the reward function offline in a supervised way, making it feasible to train on a real robot.

3. Learning to Fold Textile with Low-Cost Hardware

In this work, we propose a shallow neural network controller that is able to learn control policies for a dual-arm robot using reinforcement learning with the aim of folding a textile piece. An overview of our setup is given in Figure 1.

3.1. An Inexpensive Setup for Textile Folding

Our setup consists of two robot arms that have to fold a piece of textile. In order to demonstrate the possibility of applying our methodology in a low-cost setting, we constructed the BCN3D Moveo robot arms (https://github.com/BCN3D/BCN3D-Moveo). These robot arms have five degrees of freedom each and are low-cost and open source. They can be fabricated by rapid prototyping techniques and off-the-shelf components. The textile piece was attached to the grippers by means of a hook-and-loop fastener.

3.2. Fitted Q-Learning

In order to train a controller for robotic folding, we use fitted Q-learning [27], a model-free reinforcement learning method in which the expected discounted reward of state–action pairs are calculated based on rollouts in the environment. We utilize experience replay [28] and approximate the Q-function with a shallow multilayer perceptron (MLP). We define a sparse reward function that gives a large reward upon folding the cloth. However, this requires the detection of the state of the cloth—a non-trivial task given the high amount of cloth configurations. Hand-engineered heuristics to estimate the cloth state have been researched [8] but are highly complex to reproduce and are prone to error. Visual cues such as colored surfaces or fiducial markers [11,12] can be attached to highly deformable materials but require the many occlusions caused by the self-collision of the cloth to be handled.

4. Smart Textile

In order to solve the problems associated with using vision for the state estimation of cloth, we developed an accessible, low-cost smart textile that can identify its actual pose. By integrating a flexible tactile sensor grid into a flat textile piece, a pose estimation model can be trained. Consequently, the textile is able to sense and react to its surrounding environment, making it self-aware or making it an active smart textile [29].
The layers in the smart textile can be seen in Figure 2. The textile consists of piezoresistive rubber (commercially available as Velostat®) that is sandwiched between two layers of conductive threads that act as electrodes [30]. The electrodes are made of copper strips that are insulated with a polyimide film. This results in a matrix of 12 by 7 tactile cells. The signal conditioning for measuring the pressure applied over a tactile cell is based on the voltage divider principle. More specifically, the voltage is measured between the tactile cell, which acts as a variable resistor, and a fixed resistor of 5.6 kOhm. The voltage was measured and processed by a standard microcontroller. The resulting textile can be seen in Figure 3.
Based on the output of the 12 by 7 tactile cells, a linear classifier is trained to predict the garment state: folded or unfolded. Therefore, we first collected data from the tactile cells by bringing the cloth into multiple configurations while hand-labeling the associated state. We define the folded state of the cloth with some slack because of the way in which the robots are positioned in order to avoid a collision; this is illustrated in Figure 4. The collected data were split into a training set of 1418 samples and a test set of 511 samples. In Figure 5, the average voltage and standard deviation of the tactile cells for both garment states are shown. One can observe that there is a clear voltage peak in the center column, implying that the tactile cells in the center of the textile experience higher pressure due to the folded state.
The classifier was trained using logistic regression with l2-norm regularization on 1418 training samples in a five-fold cross-validation setting. The regularization parameter was optimized by performing a random search and by using the F1-score as a performance metric. In order to verify the usability of the classifier, we tested it on an unseen test set of 511 samples which resulted in an average F1-score of 0.97 . The full results of the test set are summarized in Table 1.
We are able to classify different shape configurations for the textile piece with high accuracy, making the smart textile useful for learning purposes. Other desired features such as wrinkle detection are hard to measure with this technology given the limited sensitivity of the piezoresistive rubber used.

5. Results

As described in Section 3, the task consists of folding a rectangular textile piece by using a low-cost dual robotic arm. The arms are positioned in an opposing manner to avoid a collision. In order to avoid damaging the smart textile, we limited the movement of each robot arm to three degrees of freedom (DOF) instead of the available five DOF. Consequently, the movements of the robots are limited to one plane.
In order to find the control policy for folding, we formulated the problem as a fitted Q-learning problem. The state-space S = R 6 is defined by the joint space of the dual-arm robot, with six motor angels in total. For each joint, we define two actions: ± Δ with Δ equal to 4°. The reward function R is sparse and defined by
R = 100 on success 1 penalty for wandering
We approximate the Q-function with a simple MLP with six input neurons (one for each of the controlled joints), two hidden layers with 128 and 64 ReLu neurons, respectively, and 12 output (linear) neurons (one for each of the possible actions). In order to find a good set of parameters, we implemented a simple forward kinematic model of the robot arms in a reaching task setting. This allowed us to sweep across various hyperparameter settings. The optimizer we used is stochastic gradient descent with a learning rate of 0.01 and batch size of 64 samples. The target network is updated each 1000 steps. Exploration is done using an ϵ -softmax policy with 5000 exploration steps. Thus, the agent will first execute fully random motor actions and gradually start choosing actions that maximize the expected q-values.
After the selection of the hyperparameters, we trained our dual robotic arm setup from scratch in the real world. As shown in Figure 6a, after approximately 60 episodes (17,000 steps, or 8 h), the robot successfully learned to complete the folding task. Figure 6b shows the number of steps the robot needs to fold the cloth per episode. The low dimensional state space of the agent and the learned reward function from tactile information using supervised learning allows the agent to learn to fold the cloth after relatively few training examples.
An example trajectory of the learned policy is shown in Figure 7. The action probabilities show that at the start of the episode, the agent primarily moves the shoulder joints, as they lead to the largest movement of the end-effector. This behavior has the benefit that it minimizes the penalty associated with wandering or other suboptimal movements. As the episode progresses, the agent has a higher probability of actuating the elbow and end-effector motors for more refined movement. The end-effector trajectories, shown in yellow, are not smooth given that the policy is an action probability function calculated over the Q-values of the neural network. Because there are multiple reaching points in the plane to fold the cloth, the agent has multiple options to move during the first part of the episode, leading to jagged end-effector trajectories.
As our experiment illustrates, it is possible to avoid the need for complex vision-based state estimation to learn difficult tasks such as the folding of clothing with low-cost robots. The use of other input modalities such as a smart textile allows a reward function to be learned with simple linear models and is applicable to use for learning on real robotic platforms.

6. Conclusions and Future Work

Deformable objects are omnipresent in our daily environment and industry but little studied in deep reinforcement learning. In this paper, we proposed a cost-effective and computationally efficient solution for learning robotic manipulation skills for deformable objects. We have shown that an inexpensive dual robotic platform can learn to fold a textile piece in the real world with neural fitted Q-learning and without reward engineering. This is done by including tactile sensing in the learning process. We have shown that the high sensitivity of the proposed tactile cells allows the capture of arbitrary cloth configurations. We demonstrated that these tactile data can be used to train a simple logistic regression classifier to detect the cloth state; thus, we create a smart textile piece that is able to accurately detect whether it is folded.
The process proposed in this paper is feasible to scale up to other instances of cloth manipulation tasks given the low computational requirements and the tactile sensitivity of the sensor matrix. The accessible fabrication procedure of the smart textile allows the implementation of this technology for any type of clothing, such as shirts and trousers. The influence on the deformable properties of the cloth can be reduced by miniaturizing the electronics, by using conductive thread instead of copper wires, and by only applying tactile cells in crucial areas. Future work includes exploiting the sensitivity of the tactile sensor grid to classify more complex cloth configurations and using more powerful, non-linear classifiers compared to the logistic regression model used in this work. Future research should be directed towards fusing the tactile information with visual data in order to generate a database of self-labeled images of arbitrary cloth configurations. These data can be produced by manipulating the cloth while recording its state with cameras. In order to detect and grasp the cloth and train policies that minimize wrinkles, visual feedback will have to be included. A suitable approach to integrate visual and tactile information is to learn a multimodal latent space, as shown in [15].
We argue that by integrating a simple smart textile in the robot learning process, one can avoid the engineering of a reward function. In our approach, we avoided the use of vision for all steps in the learning process. Consequently, we can easily run the learning process in the real world using limited computational power. In our case, the state was purely based on the kinematics of the dual robot arms. We are aware that, in order to scale up our tasks to real clothing, we will have to integrate vision for the state estimation as well. Consequently, we do not advocate vision-free solutions, but we argue for an interdisciplinary approach in which vision, proprioception and tactile information is combined in order to find the simplest solution of state and reward function estimation.

Author Contributions

Conceptualization, A.V. and F.w.; methodology, A.V. and F.w.; software, T.H., R.P. and A.V.; validation, A.V., F.w.; formal analysis, A.V.; investigation, T.H., R.P. and C.D.H.; resources, A.V. and F.w.; data curation, T.H., R.P., C.D.H. and A.V.; writing—original draft preparation, A.V. and F.w.; writing—review and editing, A.V. and F.w.; visualization, A.V., T.H., R.P. and C.D.H.; supervision, F.w.; project administration, F.w.; funding acquisition, F.w. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Research Foundation Flanders (FWO) under grant number 1S54220N.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Levine, S.; Pastor, P.; Krizhevsky, A.; Ibarz, J.; Quillen, D. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Robot. Res. 2018, 37, 421–436. [Google Scholar] [CrossRef]
  2. Morrison, D.; Tow, A.W.; McTaggart, M.; Smith, R.; Kelly-Boxall, N.; Wade-McCue, S.; Erskine, J.; Grinover, R.; Gurman, A.; Hunn, T.; et al. Cartman: The low-cost cartesian manipulator that won the amazon robotics challenge. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 7757–7764. [Google Scholar]
  3. Agrawal, P.; Nair, A.V.; Abbeel, P.; Malik, J.; Levine, S. Learning to poke by poking: Experiential learning of intuitive physics. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 5074–5082. [Google Scholar]
  4. Gu, S.; Holly, E.; Lillicrap, T.; Levine, S. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 3389–3396. [Google Scholar]
  5. Foresti, G.L.; Pellegrino, F.A. Automatic visual recognition of deformable objects for grasping and manipulation. IEEE Trans. Syst. Man Cybern. (Appl. Rev.) 2004, 34, 325–333. [Google Scholar] [CrossRef]
  6. Saha, M.; Isto, P. Manipulation planning for deformable linear objects. IEEE Trans. Robot. 2007, 23, 1141–1150. [Google Scholar] [CrossRef]
  7. Maitin-Shepard, J.; Cusumano-Towner, M.; Lei, J.; Abbeel, P. Cloth grasp point detection based on multiple-view geometric cues with application to robotic towel folding. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–7 May 2010; pp. 2308–2315. [Google Scholar] [CrossRef]
  8. Doumanoglou, A.; Stria, J.; Peleka, G.; Mariolis, I.; Petrík, V.; Kargakos, A.; Wagner, L.; Hlaváč, V.; Kim, T.; Malassiotis, S. Folding Clothes Autonomously: A Complete Pipeline. IEEE Trans. Robot. 2016, 32, 1461–1478. [Google Scholar] [CrossRef]
  9. Matas, J.; James, S.; Davison, A.J. Sim-to-Real Reinforcement Learning for Deformable Object Manipulation. In Proceedings of the 2nd Conference on Robot Learning, Zurich, Switzerland, 29–31 October 2018; Volume 87, pp. 734–743. [Google Scholar]
  10. Seita, D.; Jamali, N.; Laskey, M.; Berenstein, R.; Tanwani, A.K.; Baskaran, P.; Iba, S.; Canny, J.F.; Goldberg, K. Robot Bed-Making: Deep Transfer Learning Using Depth Sensing of Deformable Fabric. arXiv 2018, arXiv:1809.09810. [Google Scholar]
  11. Bersch, C.; Pitzer, B.; Kammel, S. Bimanual robotic cloth manipulation for laundry folding. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 1413–1419. [Google Scholar] [CrossRef] [Green Version]
  12. Tsurumine, Y.; Cui, Y.; Uchibe, E.; Matsubara, T. Deep reinforcement learning with smooth policy update: Application to robotic cloth manipulation. Robot. Auton. Syst. 2019, 112, 72–83. [Google Scholar] [CrossRef]
  13. Billard, A.; Kragic, D. Trends and challenges in robot manipulation. Science 2019, 364, eaat8414. [Google Scholar] [CrossRef] [PubMed]
  14. Tian, S.; Ebert, F.; Jayaraman, D.; Mudigonda, M.; Finn, C.; Calandra, R.; Levine, S. Manipulation by Feel: Touch-Based Control with Deep Predictive Models. arXiv 2019, arXiv:1903.04128. [Google Scholar]
  15. Lee, M.A.; Zhu, Y.; Srinivasan, K.; Shah, P.; Savarese, S.; Fei-Fei, L.; Garg, A.; Bohg, J. Making sense of vision and touch: Self-supervised learning of multimodal representations for contact-rich tasks. In Proceedings of the 2019 IEEE International Conference on Robotics and Automation (ICRA), Montreal, QC, USA, 20–24 May 2019. [Google Scholar]
  16. Yamakawa, Y.; Namiki, A.; Ishikawa, M. Motion planning for dynamic folding of a cloth with two high-speed robot hands and two high-speed sliders. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 5486–5491. [Google Scholar] [CrossRef]
  17. Willimon, B.; Birchfield, S.; Walker, I. Classification of clothing using interactive perception. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 1862–1868. [Google Scholar] [CrossRef]
  18. Arnold, S.; Yamazaki, K. Fast and Flexible Multi-Step Cloth Manipulation Planning Using an Encode-Manipulate-Decode Network (EM*D Net). Front. Neurorobot. 2019, 13. [Google Scholar] [CrossRef] [PubMed]
  19. Tanaka, D.; Arnold, S.; Yamazaki, K. EMD Net: An Encode–Manipulate–Decode Network for Cloth Manipulation. IEEE Robot. Autom. Lett. 2018, 3, 1771–1778. [Google Scholar] [CrossRef]
  20. Balaguer, B.; Carpin, S. Combining imitation and reinforcement learning to fold deformable planar objects. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 1405–1412. [Google Scholar] [CrossRef]
  21. Laskey, M.; Powers, C.; Joshi, R.; Poursohi, A.; Goldberg, K. Learning Robust Bed Making using Deep Imitation Learning with DART. arXiv 2017, arXiv:1711.02525. [Google Scholar]
  22. Rahmatizadeh, R.; Abolghasemi, P.; Bölöni, L.; Levine, S. Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-to-End Learning from Demonstration. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 3758–3765. [Google Scholar] [CrossRef] [Green Version]
  23. Baxter SDK Information. Available online: https://sdk.rethinkrobotics.com/wiki/Arms (accessed on 2 June 2020).
  24. Miller, S.; van den Berg, J.; Fritz, M.; Darrell, T.; Goldberg, K.; Abbeel, P. A geometric approach to robotic laundry folding. Int. J. Robot. Res. 2012, 31, 249–267. [Google Scholar] [CrossRef]
  25. Abbeel, P.; Ng, A.Y. Apprenticeship learning via inverse reinforcement learning. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 1. [Google Scholar]
  26. Finn, C.; Levine, S.; Abbeel, P. Guided cost learning: Deep inverse optimal control via policy optimization. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 49–58. [Google Scholar]
  27. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  28. Lin, L.J. Self-improving reactive agents based on reinforcement learning, planning and teaching. Mach. Learn. 1992, 8, 293–321. [Google Scholar] [CrossRef]
  29. Stoppa, M.; Chiolerio, A. Wearable electronics and smart textiles: A critical review. Sensors 2014, 14, 11957–11992. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Drimus, A.; Kootstra, G.; Bilberg, A.; Kragic, D. Design of a flexible tactile sensor for classification of rigid and deformable objects. Robot. Auton. Syst. 2014, 62, 3–15. [Google Scholar] [CrossRef]
Figure 1. Schematic overview of the setup: dual robot arms hold a smart textile piece and have to learn to fold it from scratch using sensory feedback of the textile and proprioceptive input of the robotic arms.
Figure 1. Schematic overview of the setup: dual robot arms hold a smart textile piece and have to learn to fold it from scratch using sensory feedback of the textile and proprioceptive input of the robotic arms.
Applsci 10 04088 g001
Figure 2. Exploded-view drawing of the textile with integrated tactile sensing. Conductive threads organized in a 12 by 7 grid are applied on piezoresistive rubber, creating a variable resistor. The resulting voltage acts as a proxy for local forces applied on the cloth.
Figure 2. Exploded-view drawing of the textile with integrated tactile sensing. Conductive threads organized in a 12 by 7 grid are applied on piezoresistive rubber, creating a variable resistor. The resulting voltage acts as a proxy for local forces applied on the cloth.
Applsci 10 04088 g002
Figure 3. Our resulting implementation of a smart textile which is able to detect and translate local forces to a state configuration (for example, folded or unfolded). The smart textile is made using accessible off-the-shelf materials and tools.
Figure 3. Our resulting implementation of a smart textile which is able to detect and translate local forces to a state configuration (for example, folded or unfolded). The smart textile is made using accessible off-the-shelf materials and tools.
Applsci 10 04088 g003
Figure 4. Example of the folding task on the real platform. The robotic arms are positioned in such a way that they cannot collide, making it physically impossible to fold the cloth fully. Thus, we do not require a fully folded state of the textile from the agent.
Figure 4. Example of the folding task on the real platform. The robotic arms are positioned in such a way that they cannot collide, making it physically impossible to fold the cloth fully. Thus, we do not require a fully folded state of the textile from the agent.
Applsci 10 04088 g004
Figure 5. The average and standard deviation of the collected data from the 12 by 7 tactile cells show a higher voltage in the center column—where the fold is—when the cloth is in the folded state. These figures qualitatively indicate that the tactile cells in the textile react to different folding configurations.
Figure 5. The average and standard deviation of the collected data from the 12 by 7 tactile cells show a higher voltage in the center column—where the fold is—when the cloth is in the folded state. These figures qualitatively indicate that the tactile cells in the textile react to different folding configurations.
Applsci 10 04088 g005
Figure 6. Results of training the real robot from scratch to learn how to fold a smart textile. (a) The folding success rate and exploration probability per episode. The success rate grows to 100 % after training for 8 h on the real robot. (b) The amount of steps required per episode to solve the task; the amount of motor commands needed to fold the cloth decreases after relatively few training examples.
Figure 6. Results of training the real robot from scratch to learn how to fold a smart textile. (a) The folding success rate and exploration probability per episode. The success rate grows to 100 % after training for 8 h on the real robot. (b) The amount of steps required per episode to solve the task; the amount of motor commands needed to fold the cloth decreases after relatively few training examples.
Applsci 10 04088 g006
Figure 7. Demonstration of a folding trajectory of the learned policy broken down into three steps: start, middle and end of the episode. The robot state on the figure is visible from the stance of the robot arms. The action probabilities are given in red next to each joint. The executed trajectory is shown with a yellow line.
Figure 7. Demonstration of a folding trajectory of the learned policy broken down into three steps: start, middle and end of the episode. The robot state on the figure is visible from the stance of the robot arms. The action probabilities are given in red next to each joint. The executed trajectory is shown with a yellow line.
Applsci 10 04088 g007
Table 1. Test set results of the linear classifier that detects whether the cloth has been folded (511 samples). The high classification performance means that the smart textile is suitable to be used as a reward function.
Table 1. Test set results of the linear classifier that detects whether the cloth has been folded (511 samples). The high classification performance means that the smart textile is suitable to be used as a reward function.
SamplesPrecisionRecallf1-Score
Unfolded3420.980.980.98
Folded1690.960.950.96

Share and Cite

MDPI and ACS Style

Verleysen, A.; Holvoet, T.; Proesmans, R.; Den Haese, C.; wyffels, F. Simpler Learning of Robotic Manipulation of Clothing by Utilizing DIY Smart Textile Technology. Appl. Sci. 2020, 10, 4088. https://0-doi-org.brum.beds.ac.uk/10.3390/app10124088

AMA Style

Verleysen A, Holvoet T, Proesmans R, Den Haese C, wyffels F. Simpler Learning of Robotic Manipulation of Clothing by Utilizing DIY Smart Textile Technology. Applied Sciences. 2020; 10(12):4088. https://0-doi-org.brum.beds.ac.uk/10.3390/app10124088

Chicago/Turabian Style

Verleysen, Andreas, Thomas Holvoet, Remko Proesmans, Cedric Den Haese, and Francis wyffels. 2020. "Simpler Learning of Robotic Manipulation of Clothing by Utilizing DIY Smart Textile Technology" Applied Sciences 10, no. 12: 4088. https://0-doi-org.brum.beds.ac.uk/10.3390/app10124088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop