What are you looking for?
51 Résultats pour : « Portes ouvertes »

L'ÉTS vous donne rendez-vous à sa journée portes ouvertes qui aura lieu sur son campus à l'automne et à l'hiver : Samedi 18 novembre 2023 Samedi 17 février 2024 Le dépôt de votre demande d'admission à un programme de baccalauréat ou au cheminement universitaire en technologie sera gratuit si vous étudiez ou détenez un diplôme collégial d'un établissement québécois.

Automated Manufacturing Engineering Research and Innovation Intelligent and Autonomous Systems CoRo – Control and Robotics Laboratory

Robots with a Sense of Touch

Robots with the sense of touch thanks to a silicone skin

The featured image was bought on Istock.com. Copyrights.

Editor’s Note

This article is based on a public presentation by Professor Vincent Duchaine of the École de technologie supérieure (ÉTS) in Montréal, as part of CBC’s La Sphère, hosted by Mathieu Dugal on October 1, 2017.

——————————–

Robots are traditionally made of metal and are very strong. Even when their strength is set to the lowest threshold, they’re unable to grasp delicate objects – like empty eggshells or Styrofoam cups – without crushing them, because they cannot sense the object. My team of researchers is working to equip these robots with an artificial skin that can measure pressure exerted at any point of contact. This silicone skin is not composed of discrete sensors; it’s more like a touch screen, except it’s also flexible and extendable.

Evaluating a Robot’s Grip with Pressure Imaging

The manufacturing industry requires robots that can pick up objects with different shapes and textures. But how can a robot be taught to evaluate its grip on an object? How can it detect whether it has successfully grasped an object, or whether the object is slipping away?

objects from Amazon to experiment robots sense of touch

Figure 1 Objects used for testing

To answer these questions, we ordered around 100 different objects from Amazon. One student had a robot pick up each of these objects about 1000 times. At each try, a pressure image of the robot’s grip was generated (see Figure 2 below) and the student noted whether or not the grip was stable. These images, combined with the student’s observations, constituted a dictionary that the robot could quickly refer to in order to judge the stability of its grip. This method is called sparse coding.

Pressure images are used to form a dictionary for sparse coding

Figure 2 Sparse coding method

The success rate of this method reached 93%, which is one of the best rates achieved so far. However, there’s still a lot of work to be done in order to integrate this system in factories where a near-100% success rate is necessary. One reason why we didn’t reach 100% is because the sense of touch alone is not enough to determine whether or not a grip is stable. In fact, human beings also use their vision to do this.

A sense of touch alone is not enough to evaluate a grip

Figure 3 Examples where eyesight is needed to properly assess grip

Determining with Sound if an Object Is Slipping

Another question the research team tried to answer is: how can a robot detect whether the object it holds is slipping? To answer this, they used vibration sensors, or devices that can detect sound. The detected vibrations can come from three sources:

  1. slippage of a gripped object against the robot’s hand (Figure 4, C1 image);
  2. slippage of a gripped object against another surface (Figure 4, C2 image);
  3. the robot’s movements (Figure 4, C4 image).
The sense of hearing is associated to the sense of touch

Figure 4 Sources of detectable vibrations

The bigger issue here is how to distinguish between different types of vibrations in order to isolate cases where an object is slipping against the robot’s hand. The sparse coding method was used again, except this time the dictionary contained sounds, in a way that’s similar to how Apple’s Siri system accomplishes voice-recognition.

Figure 5 Vibration Dictionary

Cases of vibrations caused by robot movements were fully catalogued, as were cases where no vibration is perceptible. Cases where objects slipped from the robot’s hand were identified at a rate of 84%.

Figure 6 Resulting vibrations classification

This result is very encouraging, but again, much work remains to be done before these robots can be implemented in factories.

 

About the authors
Vincent Duchaine is a professor in the Department of Automated Manufacturing Engineering. He specializes in robotics, mechatronics and touch sensors. He is also co-director of the surgical innovation program at ÉTS, McGill and Concordia, and of the Innovation and Entrepreneurship program at the UQAM management sciences (ÉSG) at UQAM.