Next Article in Journal
Resource Allocation in Multi-Carrier Multiplexed NOMA Cooperative System
Next Article in Special Issue
Visual Servoing Approach to Autonomous UAV Landing on a Moving Vehicle
Previous Article in Journal
A Hybrid Route Selection Scheme for 5G Network Scenarios: An Experimental Approach
Previous Article in Special Issue
Real-Time Short-Term Pedestrian Trajectory Prediction Based on Gait Biomechanics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling and Control of a Spherical Robot in the CoppeliaSim Simulator

by
Guelis Montenegro
1,†,
Roberto Chacón
2,†,
Ernesto Fabregas
3,
Gonzalo Garcia
4,
Karla Schröder
2,
Alberto Marroquín
2,
Sebastián Dormido-Canto
3 and
Gonzalo Farias
2,*
1
Departamento de Electrotecnia e Informática, Universidad Técnica Federico Santa María, Av. Federico Santa María 6090, Viña del Mar 2520001, Chile
2
Escuela de Ingeniería Eléctrica, Pontificia Universidad Católica de Valparaíso, Av. Brasil 2147, Valparaíso 2362804, Chile
3
Departamento de Informática y Automática, Universidad Nacional de Educación a Distancia, Juan del Rosal 16, 28040 Madrid, Spain
4
Ocean and Mechanical Engineering Department, Florida Atlantic University, 777 Glades Road EW 190, Boca Raton, FL 33431, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Submission received: 1 July 2022 / Revised: 2 August 2022 / Accepted: 10 August 2022 / Published: 12 August 2022
(This article belongs to the Special Issue Advanced Sensors Technologies Applied in Mobile Robot)

Abstract

:
This article presents the development of a model of a spherical robot that rolls to move and has a single point of support with the surface. The model was developed in the CoppeliaSim simulator, which is a versatile tool for implementing this kind of experience. The model was tested under several scenarios and control goals (i.e., position control, path-following and formation control) with control strategies such as reinforcement learning, and Villela and IPC algorithms. The results of these approaches were compared using performance indexes to analyze the performance of the model under different scenarios. The model and examples with different control scenarios are available online.

1. Introduction

The field of robotics is very extensive with respect to robot design, as it is necessary to investigate and analyse the different types of mechanisms that a robot could eventually integrate. Mobile robots can move around their environment using different mechanisms, such as wheels, caterpillars and others [1,2,3,4]. Mobile robots include robots with a spherical shape, that roll to move, like a football. This type of robot represents a particular challenge because it must roll with all of its components inside [5,6]. Examples of this kind of robot can be found in the literature with different driving/steering mechanisms, including single wheels [7,8], pendulums [9,10,11], and omnidirectional wheel mechanisms [12].
It is well-known that, currently, advanced robots are very expensive and can be exposed to damage during laboratory experimentation. For this reason, simulators are very important in this field. Virtual laboratories using these simulators offer significant benefits for robotics education [13]. Using such virtual laboratories, students can test and gain an understanding of concepts that are not easy to follow in the classroom, at any time and pace, and from anywhere [14,15]. They can also design and test control strategies before implementing them in an actual robot in the laboratory without risk of damage to the physical device.
Currently, there are many simulators for different areas of robotics. For example, Argos [16], Webots [17], RFCSIM [18], and CoppeliaSim (formerly V-REP) [19], to mention only those most used. Some of these platforms have licenses that can be used free of charge for educational purposes. They have competitive functionalities with various components that can interact with each other and can be programmed with different programming languages.
The CoppeliaSim simulator deserves special attention for being one of the most widely used for pedagogical purposes today. This simulator provides a versatile and scalable framework for creating 3D simulations in a relatively short period of time. CoppeliaSim has an integrated development environment (IDE) that is based on a distributed and scripted architecture—each scene object can have a built-in script attached, all operating at the same time, in the form of threads. CoppeliaSim contains a wealth of examples, robot models, sensors, and actuators to create and interact with a virtual world at run-time. New models can be designed and added to CoppeliaSim to implement custom-designed simulation experiments [20,21]. CoppeliaSim has an extensive list of mobile robots, among which there is no spherical robot. Therefore, it would be interesting to add a model of a spherical robot so that it would be available and could be used by the community for experiments.
In our laboratory, we have recently developed a spherical mobile robot that can be easily reproduced with a 3D-printer and some basic electronic components. Our idea was to develop and test the simulation model for this robot, as we did in previous investigations [14,15] with a model of the Khepera IV robot in the CoppeliaSim simulator. In the model, the physical properties and components, such as mass, dimensions, and other variables were carefully taken into account. This model was tested with several control strategies under different scenarios. The results obtained with the model were very similar to those obtained with the actual Khepera IV robot with the same control algorithms and experimental conditions [22]. We expect the resulting new model to be very similar to the physical robot, based on our previous results obtained with the Khepera IV robot.
This paper presents the development and testing of a model of a spherical robot whose movement is based on an internal pendulum. The robot consists of a spherical-shaped cover that protects the pendulum and the internal circuitry and allows it to roll to move from one place to another. This type of morphology has a fundamental advantage in that there is no possibility of the robot tipping over, which gives it certain stability in movement. At the same time, it has certain disadvantages with respect to sliding on the surface and difficulties with the presence of obstacles or irregularities in the terrain [9,12,23].
The main contribution of this article is the modelling and control of a non-linear model of a spherical robot that does not exist in the CoppeliaSim simulator. This is a challenging task due to the complexity of the spherical robot model. The developed model is controlled under different scenarios with several control algorithms implemented by the authors in previous studies, including Villela [24], IPC (integral proportional controller) [25], and reinforcement learning (RL) [26,27]. The experiments undertaken to test the robot model included investigation of position control, path-following and formation control. As a result of this work, the model and some examples are available online for use by the community that works with mobile robots.
The article is organised as follows: Section 2 describes the model of the spherical robot and its design and implementation in CoppeliaSim. Section 3 describes the control laws and experiments implemented with the spherical robot. Section 4 shows the tests performed on position control, path-following and formation control. Finally, Section 5 presents the conclusions and discusses future work.

2. Spherical Robot Model

This section presents the details of the mathematical model of the spherical robot in different situations.

2.1. Robot Description

A spherical robot is a mobile machine that has a spherical casing with a single contact with the surface on which it rests. The housing not only allows the robot to be in a balanced state, but also enables the robot to roll from one place to another without sliding. Figure 1 shows a picture of the actual robot in the laboratory.
For practical purposes, the mechanism that moves the robot and maintains balance is based on a pendulum and three motors. One motor operates the pendulum and modifies the centre of gravity of the robot; the other motors are connected to the casing-sphere, which allows its movement using rotation. The design of this robot is based on [28]. Figure 2 shows a schematic view of the robot, where the right side is the front view and the left side is the lateral view. The components are as follows: (1) casing sphere, a 3D-printed shell which is the body of the robot; (2) bearings, the joins of the case with the shaft; (3) pendulum, which is a stick that allows the robot to turn; (4) counterweights, which are two body masses to improve the stability of the robot; (5) circuitry box, is a small space that contains all the electronic components; (6) shaft (fixed axle), which is the join between the circuit box and the case; and (7) articulation, which is the join between the pendulum and the shaft.

2.2. Horizontal Motion Equations

A simplified model of the robot can be seen in Figure 3, which is based on [28]. The model considers moments of inertia, the radius of the sphere, masses of the elements that make up the robot, angles, and the direction of linear velocity.
In this model, it is considered that the casing is always in contact with the surface, which allows the robot to roll without slipping. The dynamic model of the spherical robot for a horizontal movement considers the balance of potential energy, kinetic energy and rotational energy, respectively, as can be seen in Equations (1)–(3):
U 1 = 0 U 2 = M 2 g e cos ( θ 1 + θ 2 )
K 1 = 1 2 M 1 ( r w 1 ) 2 K 2 = 1 2 M 2 ( r ω 1 e cos ( θ 1 + θ 2 ) ( ω 1 + ω 2 ) ) 2
T 1 = 1 2 J 1 ω 1 2 T 2 = 1 2 J 2 ( ω 1 + ω 2 ) 2
where the main variables are the following:
U 1 : is the potential energy of the spherical housing with respect to the height of its centroid.
U 2 : is the potential energy of the pendulum with respect to the height of the centroid of the spherical casing.
K 1 : is the kinetic energy of the spherical shell.
K 2 : is the kinetic energy of the pendulum.
T 1 : is the rotational energy of the spherical shell.
T 2 : is the rotational energy of the pendulum.
e: is the distance between the centroid of the spherical shell and the pendulum.
M 1 : is the mass of the spherical shell.
M 2 : is the mass of the pendulum.
The Lagrange equations are calculated as follows:
L = K 1 + K 2 + T 1 + T 2 U 1 U 2
d d t L ω 1 L θ 1 = T + T f
d d t L ω 2 L θ 2 = T
where t is the independent variable of time, T is the torque applied between the casing and the circuitry, and T f is the torque that appears with the friction force that occurs between the casing and the ground. Solving the Equations (5) and (6), we obtain:
T = a 1 ( J 2 M 2 r e cos ( θ 1 + θ 2 ) + M 2 e 2 ) + a 2 ( J 2 + M 2 e 2 ) + M 2 g e sin ( θ 1 + θ 2 )
This last equation is useful to determine which motor may be used for the construction of the spherical robot and to consider its design.

2.3. Robot Turning Equations

The motion that allows changing the direction of the spherical robot is based on the mass of the pendulum using a CMG (control moment gyroscope). The calculation of this motion is based on a torsion pendulum, as shown in Figure 4.
Where T p is the torque of the pendulum; R p is the radius of the pendulum; ω p is the angular velocity of the pendulum; A c is the swing acceleration of the pendulum; and M 2 is the mass of the pendulum. The equation of the rotational motion of the pendulum is as follows:
T p = 4 R p M 2 A c
Figure 5 shows an angle θ corresponding to the angular position of a spherical mobile robot, whose orientation is controlled by varying the speed of a reaction wheel. The angular velocity of the reaction wheel, relative to the spherical mobile robot, can be varied by varying the voltage applied to the electric motor. This means that, if the motor rotates clockwise, the spherical robot will orient itself in the opposite direction. The effect is achieved by analysing the angular momentum at the axis of rotation; as the speed of the wheel varies, the speed of the mobile robot will also start to vary so that the momentum remains constant.
The equations governing this phenomenon are obtained after analysing the momentum of the spherical robot and the reaction wheel around the axis of rotation. As shown below:
θ ˙ = ω
ω ˙ = B 1 J 1 ω + B 2 J 1 Ω 1 J 1 τ m
Ω ˙ = B 2 J e q Ω + 1 J e q τ m
where the main variables are the following: θ is the angle of the spherical robot casing (robot angle); ω is the angular velocity of the spherical robot; Ω is the angular velocity of the pendulum wheel; J 1 is the moment of inertia of the spherical robot casing; J 2 is the moment of inertia of the pendulum wheel of the spherical robot; and J e q is the equivalent moment of inertia where: 1 J e q = 1 J 1 + 1 J 2 .

2.4. Building the Model of the Robot

The parts of the spherical robot were designed using the 3D design software Autodesk Fusion 360 [29] based on the actual robot shown in Figure 2. These parts were then imported into the CoppeliaSim simulator [30] working environment and the robot was assembled manually. Figure 6 shows the process of building the robot.
As can be seen, on the right side, the imported parts of the robot are shown (i.e., housing, pendulum motor, housing motors). On the left side, the result of the robot assembly is shown. The total diameter of the robot is 18 centimetres. Note that the motors and internal elements were built in the same software.

3. Experiments with the Spherical Robot

In this section, some tests/experiments implemented with the robot are presented and commented on.

3.1. Position Control

This experiment consisted of getting the robot to move from one point C (current position) to another T p (target point) in the most efficient way possible, which implies that it does so by following the shortest path to the destination point. Note that this robot can rotate without displacement, which means that its model is holonomic. To add complexity to the experiment, we imposed non-holonomic constraints on the model. This means that the robot has to move in order to rotate (it cannot rotate about its own position). Figure 7 shows a representation of this experiment.
As can be seen, the variables involved in this experiment were, on one hand, the pose C( x , y , θ ), which includes the position (x,y) and the orientation angle ( θ ) of the robot; on the other hand, the distance (d) and the angle ( α ) at which the target point is located. These variables are calculated as follows:
d = y p y c 2 + x p x c 2
α = tan 1 y p y c x p x c
The control law is calculated using the angular error as input ( α e = θ α ) and obtaining as outputs the linear velocity ( ν ) and the angular velocity of the robot ( ω ). Then the corresponding values for the housing motor and pendulum are calculated for the robot to move with these angular and linear velocities. As a result, the robot is positioned in a new pose C (x,y, θ ), which is used to recalculate the values described above. Figure 8 shows the block diagram of the control loop for this experiment. Where the block Compute implements Equations (12) and (13); while the Control Law block can be implemented in different ways, with artificial intelligence or conventional control law approaches, as is explained in the next subsections.
In the case of the spherical robot, the linear velocity ( ν ) is applied as a voltage to the servomotors of Figure 2, which allows movement of the pendulum backwards or forwards to change the centre of gravity of the robot and to make the robot move in one of those directions. The angular velocity ( ω ) is applied to the DC motor of the pendulum, which allows the robot to rotate clockwise or counter-clockwise.
In addition to this experiment, the model was tested with other approaches: (1) path-following and (2) formation control. In the first case, the robot must control its position by following a trajectory received as a reference. In the second experiment, more robots were added to the scenario. One of them acts as the leader and the rest as followers. The followers use the position of the leader to make a formation around the leader.

3.2. Reinforcement Learning Approach

Reinforcement learning is able to provide an optimal solution despite the complexity of the system. The system learns by acting on the environment while operating in real-time. This is an advantage of traditional optimization methods that rely on a mathematical model and are tuned backwards in time [31,32].
The reinforcement learning approach for this research, called Q-learning, is based on solving the Bellman equation, and the principle of optimality. This technique allows an optimal learning process to be carried out during regular operation, based on the robot’s dynamics, and continuous-time signals. In the limit, the Q-matrix captures a discretized version of the optimal action-state combination in terms of the highest long-term reward.
Given a system described by the dynamics x k + 1 = f ( x k , u k ) and a reward function σ ( x k , u k ) , where x k is the state of the system, and u k = π ( x k ) the control policy, a long-term reward can be defined by Equation (14):
k = 0 γ k σ ( x k , u k ) = k = 0 γ k σ ( x k , π ( x k ) )
where 0 < γ < 1 is a discount factor required to penalize future rewards and to ensure convergence of the summation. This expression represents the discounted accumulated rewards starting from the current state x 0 and the application of the policy π .
To apply Bellman’s optimality principle, the previous long-term reward expression (14) is redefined in terms of the function Q ( x k , u k ) , called action-value, which allows for the splitting of the reward assignment into two consecutive steps. This action-value function conveys the long-term reward by the contribution of the immediate reward due to applying an arbitrary action u k while in the state x k , and by the discounted accumulated reward continuing with the control policy π . This is as shown in (15) starting from x 0 :
Q π ( x 0 , u 0 ) = σ ( x 0 , u 0 ) + k = 1 γ k σ ( x k , π ( x k ) ) = σ ( x 0 , u 0 ) + γ k = 0 γ k σ ( x k + 1 , π ( x k + 1 ) )
The optimal value is obtained by maximizing the future rewards; using the optimal policy defined by π , a recursive equation is obtained:
Q ( x k , u k ) = σ ( x k , u k ) + γ max μ Q ( x k + 1 , μ )
This equation captures the optimal principle by stating that future optimal control actions are not specified by past optimal values, but, instead, only by the current state. The major advance in these calculations is the viability of forward-in-time learning, as opposed to a standard optimal search performed backwards-in-time. This method is also known as Q-learning. From (16), the following recursive equation can be devised that asymptotically converges to the fixed manifold Q [33,34]:
Q i + 1 ( x k , u k ) = Q i ( x k , u k ) + α ( σ ( x k , u k ) + γ max μ Q i ( x k + 1 , μ ) Q i ( x k , u k ) ) .
The term σ ( x k , u k ) + γ max μ Q i ( x k + 1 , μ ) Q i ( x k , u k ) is typically labeled temporal difference T D i ( x k , u k ) , or error between the target value σ ( x k , u k ) + γ max μ x k + 1 , μ and the current value Q i + 1 ( x k , u k ) , with 0 < α < 1 a learning rate. The expression (17) resembles a gradient descend numerical search. Another interpretation of (17) is the structure of a low-pass filter, by rearranging it as Q i + 1 ( x k , u k ) = α T D i ( x k , u k ) + ( 1 α ) Q i ( x k , u k ) . The learning rate α , or numerical search step size, establishes the effect of new information overriding previous information. A small value will reduce the rate of learning, while a larger value will rely more heavily on new data, despite what was previously learned.

3.3. Control Laws: Villela and IPC Approaches

As was mentioned before, the Control Law block of Figure 9 can be implemented with traditional control laws or with a machine learning approach. In this subsection, we show both control laws that will later be implemented in the robot. For example, the Villela control law [24], named after its author, was used previously with different kinds of robots with good results [15,18,22]. It calculates the linear velocity ( ν ) and the angular velocity ( ω ) of the robot, as shown in Equations (18).
ν = ν m a x i f d > k r d ν m a x k r i f d k r ω = ω m a x sin α e
where ν m a x is the maximum linear velocity, k r is the radius of a docking area (around the target point) and ω m a x is the maximum angular velocity of the robot.
Based on the Villela control law, in a previous study, we developed what we term an integral proportional controller IPC [25], which was compared with the Villela algorithm and was found to produce better results. The controller implements the velocities as follows in Equation (19):
v = min K v p α e d , v max ω = K p sin α e + K i 0 t α e d t
where p ( α e ) = 1 | α e | / π , for α e [ π , π ] and K v , K p , and K i are tuning parameters of the control law. We tested this control law with a differential wheeled mobile robot, so it is challenging to implement this controller with a spherical robot.

4. Results

In this section, the results of the implementation of the control experiments with the developed model are presented.

4.1. Reinforcement Learning Results

This subsection shows the simulation results for different tests in several iterations to build the Q-matrix. The Q-matrix was obtained in MATLAB during the learning stage and exported to Python (Spyder-Anaconda IDE). The CoppeliaSim software was connected to Spyder via remote API, which ensured it was compatible with Python programming. The experiments were performed with the CoppeliaSim simulator using the developed spherical robot model.
In the learning stage, the algorithm builds the Q-matrix to learn how to reach the destination point. To this end, the angle error ( α e ) is used to obtain the angular velocity ( ω ) in order to control the position of the robot. Note that, initially, the linear velocity ( ν ) is kept constant at its maximum value until the robot reaches the docking area.
The Q-matrix is composed of the sets (state, action), where the state is the angle error ( α e ), and the action is the angular velocity ( ω ). The criterion for obtaining the rewards of the Q-matrix is to penalize significant changes in the angle error and small changes in the angular velocity of the pendulum. In this case, the matrix Q has a size of 126 × 41, where 126 states are generated linearly spaced between π and π , and 41 actions linearly spaced between π / 2 and π / 2 . The array is made up of initial reward values. These initial values are adjusted according to the number of iterations of the algorithm based on a learning rate, a discount rate and a coefficient of relationship between exploration and use. They explore and use values to allow the robot to explore the space to complete knowledge of it and later use that knowledge.
Figure 9 shows the results of the position control experiment for different iterations of the RL algorithm (RL 500 m-500.000 iterations, RL 1M-1.000.000 iterations, and so on). The lines describe the trajectories followed by the robot for each value of iteration. The initial position of the robot is represented by the base of the red arrow at ( 0 ; 0 ), and the target point is represented by the red cross located at ( 5 ; 0 ). The direction of the arrow represents the initial orientation of the robot.
Figure 10 shows the distance to the destination point for these experiments. The y-axis represents the distance in meters and the x-axis represents the time in seconds. As can be seen, for all experiences, the time to arrive at the destination was similar, around 14 s. This would be expected given the similarity between the trajectories shown in the figure above.
The quality of each control algorithm can be evaluated using performance indexes. These indexes use the integral of the error, which is, in our case, the distance to the target point. The performance indexes considered in this work are the following: (1) integral square error (ISE), (2) integral absolute error (IAE), (3) integral time squared error (ITSE), and (4) integral time absolute error (ITAE). Note that the last two also include the time in the analysis [35]. Table 1 shows the performance indexes to compare the results of each algorithm. Note that all the indexes showed similar results, which is logical in view of the above results. The best performance was shown by the RL4M algorithm. For that reason, we selected this algorithm to compare with the other approaches.

4.2. Comparison between Different Approaches (RL, Villela and IPC)

To establish a basis for comparison of the results of the different control algorithms with the spherical robot, in addition to the RL, the Villela and the IPC algorithms were selected. In both algorithms, the parameters were selected based on our previous experience with the implementation of these experiments with the Khepera IV robot (see for example [25,27]). In the Villela algorithm, the parameters were the following: V m a x = 1 and ω m a x = π / 2 . For the IPC algorithm, the parameters were the following: K v = 0.15 , K p = 1.5 , K i = 0.000001 , V m a x = 1 and ω m a x = π / 2 .
Figure 11 shows the results of the position control experiment for the different algorithms (Villela, IPC, and RL). As in the previous case, the red arrow represents the initial orientation of the robot and the red cross represents the target point. The lines describe the trajectories followed by the robot for each control law. The initial position of the robot is represented by the base of the arrow at ( 0 ; 0 ), and the target point located at ( 5 ; 0 ).
As can be seen, the IPC and RL4M algorithms describe similar trajectories, while Villela’s approach shows the worst trajectory. In order to provide a better comparison, we can analyze the graph of distance vs. time. Figure 12 shows the distance to the destination point for these experiments. The y-axis represents the distance in meters and the x-axis represents the time in seconds.
As can be seen, the better performance was demonstrated by the RL4M algorithm, which took around 14 s to reach the destination point. Note that, in the previous figure, it appears that IPC had a similar trajectory to RL4M, but when the time is taken into account in the analysis, the differences are clearer. With the IPC algorithm, the robot took more than 20 s to reach the destination point. So the trajectory was similar but took more time and the performance was the worst of all, while RL4M showed the best behaviour.
Table 2 shows the performance indexes for all algorithms. As can be seen, as was expected, the best performance was shown by the RL4M algorithm, which confirms the previous results.

4.3. Path Following

To test the control strategies in a different scenario, we implemented a path-following example [36,37,38]. This experiment is widely known in the field of mobile robot control because it is used to demonstrate the behaviour of the implemented control algorithm. It consists of “dynamic” position control of the robot in which the reference point constantly changes to describe a trajectory by joining all points. The result is that the robot follows the points one by one to create the trajectory. Figure 13 shows the implementation of this experiment with the spherical robot in CoppeliaSim.
Figure 14 shows the trajectories described by the robot for different control algorithms: RL4M (red line), Villela (green line) and IPC (violet line). The dashed line represents the trajectory that the robot receives as a reference. As can be seen, the robot follows the trajectory with different behaviours for all the algorithms.
At first glance, it appears that the best performance was shown by the Villela algorithm. To perform a better comparison, we calculated the performance indexes for all algorithms. Table 3 shows these results.
As can be seen, the lowest values in all indexes were for the Villela algorithm, which means that, for this algorithm, the robot followed the trajectory better.

4.4. Multi-Agent Formation Control

This experiment was based on [22,39] and consisted of making a formation in a cooperative and decentralized way. One robot acted as the leader and the rest as followers. The positions of the follower’s robots were controlled as in the previous experiment. To make a formation, the followers’ robots have to reach a position using the leader position as the reference. Equation (20) shows how the velocity of the leader robot is calculated as a function of its own position error ( E p m ) and the followers’ errors in the formation ( E f ).
ν m ( t ) = K p E p m ( t ) K f E f ( t )
The values of K p and K f are manually adjusted to control the influence of each error in the velocity of the leader robot. If K f = 0 , the errors of the followers in the formation are not taken into account, and the control is made in a non-cooperative way because the leader robot does not consider the errors of the followers. Equation (21) shows how the formation error is calculated.
E f t = i = 1 N E p i t
Figure 15 shows this experiment in the CoppeliaSim simulator for the RL algorithm. For the leader robot, the reference is the target point at the left of the image (red semi-sphere) and, for the followers, the target points are their positions in the formation. In this case, the followers use the position of the leader to make a triangular formation around it. Both followers are situated at a fixed 4 m from the leader and 30 and − 30 behind it, respectively.
As can be seen, initially, the robots make a triangle using the leader robot as a reference. At the end of the experiment, the followers maintain the formation around the leader robot. Figure 16 shows the data for this experiment. The blue small circle represents the initial position of the leader robot and the green cross represents the destination point. The blue line represents the trajectory described by the leader robot and the red and orange lines represent the trajectories described by the following robots. As can be seen, the robots maintain the formation during the experiment.
Figure 17 shows the results of this experiment for all the algorithms. The y-axis represents the distance travelled by the robots. The leader robot is represented by the blue lines and they show the distance from the robot to its target point. The followers are represented by the red and orange lines, which show the distance between each follower and the leader.
As can be seen, at the beginning, the leader robot moves away from the target because, initially, the target is at its back. After a few seconds, the leader reaches the destination point, while the followers maintain a constant distance to the leader (4 m), which means that the formation is maintained during the experience.
By simple visual inspection, it can be observed that the RL algorithm showed better performance because the leader robot reached the destination point in less time and travelled the shortest distance. However, to be on the safe side, we calculated the performance indexes for each experiment to establish a more accurate comparison. Table 4, Table 5 and Table 6 show the performance indexes for each algorithm and robot in each experiment.
All experiences were generated with the same initial conditions, and only the control algorithm was changed in each case. We can then compare the results using Equation (21), which is shown in the row Sum in Table 4, Table 5 and Table 6. As can be seen, the least values in all cases for the Sum row were observed for the RL experiment, which confirms that the better performance was produced by this algorithm.

5. Conclusions

This article presents the design and implementation of a model of an actual spherical robot, the method of movement of which is based on an internal pendulum. The design of the model was developed using the 3D-design and modeling software, Autodesk Fusion 360. The model was incorporated piece-by-piece into the CoppeliaSim simulator where the hardware was assembled; the position control strategy was programmed in the LUA and Python programming languages to verify its operation. Different experiments, concerning position control, path-following, and multi-robot formation control were performed. The results obtained with the different control laws and experiments showed that the design and implementation of the robot model were satisfactory since its behaviour was similar to that previously obtained with a differential model of the Khepera IV robot. Future work will include performing these experiments with the actual robot in the platform previously implemented in our laboratory [22]. This is a challenging task due to the complexity of obtaining the absolute position of the robot in the platform.

Author Contributions

Conceptualization, G.M., R.C. and G.F.; methodology, G.M. and G.F.; software, G.M., G.G. and R.C.; validation, E.F., K.S. and A.M.; formal analysis, G.M.; investigation, E.F. and G.M.; resources, G.F. and S.D.-C.; data curation, G.G.; writing—original draft preparation, E.F., G.M. and G.G.; writing—review and editing, G.M., E.F., G.F. and S.D.-C.; visualization, E.F. and G.M.; supervision, G.F. and S.D.-C.; project administration, G.F.; funding acquisition, G.F. and S.D.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported, in part, by the Chilean Research and Development Agency (ANID) under Project FONDECYT 1191188, The National University of Distance Education under Project 2021V/-TAJOV/00, and the Ministry of Science and Innovation of Spain under Project PID2019-108377RB-C32.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jones, J.; Seiger, B.; Flynn, A. Mobile Robots: Inspiration to Implementation; CRC Press: Boca Raton, FL, USA, 1998. [Google Scholar]
  2. Siegwart, R.; Nourbakhsh, I.R.; Scaramuzza, D. Introduction to Autonomous Mobile Robots; MIT Press: Cambridge, MA, USA, 2011. [Google Scholar]
  3. Fankhauser, P.; Gwerder, C. Modeling and Control of a Ballbot. Ph.D. Thesis, Eidgenössische Technische Hochschule Zürich, Zürich, Switzerland, 2010. [Google Scholar]
  4. Rubio, F.; Valero, F.; Llopis-Albert, C. A review of mobile robots: Concepts, methods, theoretical framework, and applications. Int. J. Adv. Robot. Syst. 2019, 16. [Google Scholar] [CrossRef]
  5. Chen, W.H.; Chen, C.P.; Tsai, J.S.; Yang, J.; Lin, P.C. Design and implementation of a ball-driven omnidirectional spherical robot. Mech. Mach. Theory 2013, 68, 35–48. [Google Scholar] [CrossRef]
  6. Cook, G.; Zhang, F. Mobile Robots: Navigation, Control and Sensing, Surface Robots and AUVs; John Wiley & Sons: Hoboken, NY, USA, 2020. [Google Scholar]
  7. Bujňák, M.; Pirník, R.; Rástočný, K.; Janota, A.; Nemec, D.; Kuchár, P.; Tichý, T.; Łukasik, Z. Spherical Robots for Special Purposes: A Review on Current Possibilities. Sensors 2022, 22, 1413. [Google Scholar] [CrossRef] [PubMed]
  8. Alexey, B.; Alexander, K.; Yury, K.; Anton, K. Stabilization of the motion of a spherical robot using feedbacks. Appl. Math. Model. 2019, 69, 583–592. [Google Scholar] [CrossRef]
  9. Kilin, A.; Pivovarova, E.; Ivanova, T.B. Spherical robot of combined type: Dynamics and control. Regul. Chaotic Dyn. 2015, 20, 716–728. [Google Scholar] [CrossRef]
  10. Asiri, S.; Khademianzadeh, F.; Monadjemi, A.; Moallem, P. The design and development of a dynamic model of a low-power consumption, two-pendulum spherical robot. IEEE/Asme Trans. Mechatron. 2019, 24, 2406–2415. [Google Scholar] [CrossRef]
  11. Ma, L.; Sun, H.; Song, J. Fractional-order adaptive integral hierarchical sliding mode control method for high-speed linear motion of spherical robot. IEEE Access 2020, 8, 66243–66256. [Google Scholar] [CrossRef]
  12. Karavaev, Y.L.; Kilin, A.A. The dynamics and control of a spherical robot with an internal omniwheel platform. Regul. Chaotic Dyn. 2015, 20, 134–152. [Google Scholar] [CrossRef]
  13. Galán, D.; Fabregas, E.; Garcia, G.; Sáenz, J.; Farias, G.; Dormido-Canto, S.; Dormido, S. Online virtual control laboratory of mobile robots. IFAC-PapersOnLine 2018, 51, 316–321. [Google Scholar] [CrossRef]
  14. Peralta, E.; Fabregas, E.; Farias, G.; Vargas, H.; Dormido, S. Development of a Khepera IV Library for the V-REP Simulator. In Proceedings of the 11th IFAC Symposium on Advances in Control Education ACE, Bratislava, Slovakia, 1–3 June 2016; Volume 49, pp. 81–86. [Google Scholar] [CrossRef]
  15. Farias, G.; Fabregas, E.; Peralta, E.; Torres, E.; Dormido, S. A Khepera IV library for robotic control education using V-REP. IFAC-PapersOnLine 2017, 50, 9150–9155. [Google Scholar] [CrossRef]
  16. Pinciroli, C.; Trianni, V.; O’Grady, R.; Pini, G.; Brutschy, A.; Brambilla, M.; Mathews, N.; Ferrante, E.; Di Caro, G.; Ducatelle, F.; et al. ARGoS: A modular, parallel, multi-engine simulator for multi-robot systems. Swarm Intell. 2012, 6, 271–295. [Google Scholar] [CrossRef]
  17. Michel, O. Webots: Symbiosis between virtual and real mobile robots. In Proceedings of the International Conference on Virtual Worlds, Paris, France, 1–3 July 1998; Springer: Berlin/Heidelberg, Germany, 1998; pp. 254–263. [Google Scholar]
  18. Fabregas, E.; Farias, G.; Dormido-Canto, S.; Dormido, S. RFCSIM simulador interactivo de robótica móvil para control de formación con evitación de obstáculos. In Proceedings of the XVI Congreso Latinoamericano de Control Automático, Cancún, México, 14–17 October 2014. [Google Scholar]
  19. Rohmer, E.; Singh, S.P.; Freese, M. V-REP: A versatile and scalable robot simulation framework. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; IEEE: Piscataway, NY, USA, 2013; pp. 1321–1326. [Google Scholar] [CrossRef]
  20. Fabregas, E.; Farias, G.; Peralta, E.; Vargas, H.; Dormido, S. Teaching control in mobile robotics with V-REP and a Khepera IV library. In Proceedings of the 2016 IEEE Conference on Control Applications (CCA), Buenos Aires, Argentina, 19–22 September 2016; pp. 821–826. [Google Scholar] [CrossRef]
  21. Farias, G.; Torres, E.; Fabregas, E.; Vargas, H.; Dormido-Canto, S.; Dormido, S. Navigation control of the Khepera IV model with OpenCV in V-REP simulator. In Proceedings of the 2018 IEEE International Conference on Automation/XXIII Congress of the Chilean Association of Automatic Control (ICA-ACCA), Concepcion, Chile, 17–19 October 2018; IEEE: Piscataway, NY, USA, 2018; pp. 1–6. [Google Scholar]
  22. Farias, G.; Fabregas, E.; Peralta, E.; Vargas, H.; Dormido-Canto, S.; Dormido, S. Development of an easy-to-use multi-agent platform for teaching mobile robotics. IEEE Access 2019, 7, 55885–55897. [Google Scholar] [CrossRef]
  23. Zhan, Q.; Cai, Y.; Yan, C. Design, analysis and experiments of an omni-directional spherical robot. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; IEEE: Piscataway, NY, USA, 2011; pp. 4921–4926. [Google Scholar]
  24. Villela, V.J.G.; Parkin, R.; Parra, M.L.; González, J.M.D.; Liho, M.J.G. A wheeled mobile robot with obstacle avoidance capability. Inge. Mecánica. Tecnol. Desarro. 2004, 1, 159–166. [Google Scholar]
  25. Fabregas, E.; Farias, G.; Aranda-Escolástico, E.; Garcia, G.; Chaos, D.; Dormido-Canto, S.; Bencomo, S.D. Simulation and experimental results of a new control strategy for point stabilization of nonholonomic mobile robots. IEEE Trans. Ind. Electron. 2019, 67, 6679–6687. [Google Scholar] [CrossRef]
  26. Farias, G.; Garcia, G.; Montenegro, G.; Fabregas, E.; Dormido-Canto, S.; Dormido, S. Position control of a mobile robot using reinforcement learning. IFAC-PapersOnLine 2020, 53, 17393–17398. [Google Scholar] [CrossRef]
  27. Farias, G.; Garcia, G.; Montenegro, G.; Fabregas, E.; Dormido-Canto, S.; Dormido, S. Reinforcement Learning for Position Control Problem of a Mobile Robot. IEEE Access 2020, 8, 152941–152951. [Google Scholar] [CrossRef]
  28. Nagai, M. Control System for a Spherical Robot. Master’s Thesis, Luleå University of Technology, Luleå, Sweden, 2008. [Google Scholar]
  29. Autodesk Fusion 360. Available online: https://www.autodesk.com/products/fusion-360/overview (accessed on 30 September 2021).
  30. CoppeliaSim. Available online: https://www.coppeliarobotics.com/downloads (accessed on 30 September 2021).
  31. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  32. Sumanas, M.; Petronis, A.; Bucinskas, V.; Dzedzickis, A.; Virzonis, D.; Morkvenaite-Vilkonciene, I. Deep Q-Learning in Robotics: Improvement of Accuracy and Repeatability. Sensors 2022, 22, 3911. [Google Scholar] [CrossRef]
  33. Watkins, C.J.C.H. Learning from delayed rewards. Ph.D. Thesis, King’s College, Cambridge, UK, 1989. [Google Scholar]
  34. Watkins, C.J.C.H.; Dayan, P. Technical Note: Q -Learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  35. Sultan, A.J. Optimal load frequency control in a single area power system based Genetic Algorithm. Int. J. Sci. Eng. Res. 2014, 5, 2196–2200. [Google Scholar]
  36. Coelho, P.; Nunes, U. Path-following control of mobile robots in presence of uncertainties. IEEE Trans. Robot. 2005, 21, 252–261. [Google Scholar] [CrossRef]
  37. Lapierre, L.; Soetanto, D. Nonlinear path-following control of an AUV. Ocean. Eng. 2007, 34, 1734–1744. [Google Scholar] [CrossRef]
  38. Rayguru, M.; Elara, M.; Ramalingam, B.; Viraj, J.M.M.A.; Bhagya, P.S.S.M. A Path Tracking Strategy for Car Like Robots with Sensor Unpredictability and Measurement Errors. Sensors 2020, 20, 3077. [Google Scholar] [CrossRef]
  39. Lawton, J.R.; Beard, R.W.; Young, B.J. A decentralized approach to formation maneuvers. IEEE Trans. Robot. Autom. 2003, 19, 933–941. [Google Scholar] [CrossRef]
Figure 1. Actual spherical robot.
Figure 1. Actual spherical robot.
Sensors 22 06020 g001
Figure 2. Details of the spherical robot.
Figure 2. Details of the spherical robot.
Sensors 22 06020 g002
Figure 3. Simplified scheme of the horizontal movement of the robot (1 rotation angle of the ball, 2 rotation angle of the pendulum with respect to the ball).
Figure 3. Simplified scheme of the horizontal movement of the robot (1 rotation angle of the ball, 2 rotation angle of the pendulum with respect to the ball).
Sensors 22 06020 g003
Figure 4. Pendulum schema reaction wheel.
Figure 4. Pendulum schema reaction wheel.
Sensors 22 06020 g004
Figure 5. Top view of the robot for rotational movement.
Figure 5. Top view of the robot for rotational movement.
Sensors 22 06020 g005
Figure 6. Assembling the robot in CoppeliaSim.
Figure 6. Assembling the robot in CoppeliaSim.
Sensors 22 06020 g006
Figure 7. Position control experiment.
Figure 7. Position control experiment.
Sensors 22 06020 g007
Figure 8. Position control block diagram.
Figure 8. Position control block diagram.
Sensors 22 06020 g008
Figure 9. Obtained trajectories for different values of iteration in RL algorithm.
Figure 9. Obtained trajectories for different values of iteration in RL algorithm.
Sensors 22 06020 g009
Figure 10. Distance vs. time of all experiences/iterations.
Figure 10. Distance vs. time of all experiences/iterations.
Sensors 22 06020 g010
Figure 11. Obtained trajectories for each control algorithm (Villela, IPC and RL4M).
Figure 11. Obtained trajectories for each control algorithm (Villela, IPC and RL4M).
Sensors 22 06020 g011
Figure 12. Distance vs. time of all experiences (RL, Villela and IPC).
Figure 12. Distance vs. time of all experiences (RL, Villela and IPC).
Sensors 22 06020 g012
Figure 13. Path following of a Lissajous figure.
Figure 13. Path following of a Lissajous figure.
Sensors 22 06020 g013
Figure 14. Path following example for Villela, IPC and RL4M.
Figure 14. Path following example for Villela, IPC and RL4M.
Sensors 22 06020 g014
Figure 15. Formation control experiment.
Figure 15. Formation control experiment.
Sensors 22 06020 g015
Figure 16. Results of the formation control experiment.
Figure 16. Results of the formation control experiment.
Sensors 22 06020 g016
Figure 17. Positions of the robots for all algorithms.
Figure 17. Positions of the robots for all algorithms.
Sensors 22 06020 g017
Table 1. Performance indexes for each algorithm. In bold, the better case.
Table 1. Performance indexes for each algorithm. In bold, the better case.
IndexRL 500 mRL 1MRL 3MRL4MRL 5MRL 8MRL 10M
IAE46.8145.8446.2545.1945.6745.6645.52
ISE198.29195.68197.05191.79194.27194.47194.23
ITSE804.22781.07791.03755.07775.13776.11771.11
ITAE234.15222.85227.18217.66221.73221.49219.67
Table 2. Performance indexes for each algorithm in the position control experiment. In bold, the better case.
Table 2. Performance indexes for each algorithm in the position control experiment. In bold, the better case.
IndexVillelaRL4MIPC
IAE58.5945.3985.28
ISE273.33191.93388.68
ITSE1.334.0754.552.985.5
ITAE338.50220.96743.39
Table 3. Performance indexes for each algorithm in the path-following experiment. In bold, the better case.
Table 3. Performance indexes for each algorithm in the path-following experiment. In bold, the better case.
IndexRL4MVillelaIPC
IAE87.7985.12212.07
ISE162.91162.12384.11
ITSE4.18 × 10 3 3.78 × 10 3 2.34 × 10 4
ITAE2.25 × 10 3 2.02 × 10 3 1.32 × 10 4
Table 4. Performance indexes for the Villela algorithm in the formation control experiment.
Table 4. Performance indexes for the Villela algorithm in the formation control experiment.
RobotIAEISEITSEITAE
Leader57.97271.451.32 × 10 3 332.37
Follower 168.41268.242.34 × 10 3 599.01
Follower 268.90272.062.38 × 10 3 603.86
Sum195.28811.766049.301535.24
Table 5. Performance indexes for the RL4M algorithm in the formation control experiment.
Table 5. Performance indexes for the RL4M algorithm in the formation control experiment.
RobotIAEISEITSEITAE
Leader46.00197.80802.10224.79
Follower 158.41235.351.72 × 10 3 427.47
Follower 256.12217.321.57 × 10 3 408.16
Sum160.53650.484103.301060.42
Table 6. Performance indexes for the IPC algorithm in the formation control experiment.
Table 6. Performance indexes for the IPC algorithm in the formation control experiment.
RobotIAEISEITSEITAE
Leader95.52377.803.11 × 10 3 1.04 × 10 3
Follower 1129.12506.938.24 × 10 3 2.11 × 10 3
Follower 2129.44509.428.38 × 10 3 2.13 × 10 3
Sum354.091394.1719,747.705297.60
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Montenegro, G.; Chacón, R.; Fabregas, E.; Garcia, G.; Schröder, K.; Marroquín, A.; Dormido-Canto, S.; Farias, G. Modeling and Control of a Spherical Robot in the CoppeliaSim Simulator. Sensors 2022, 22, 6020. https://0-doi-org.brum.beds.ac.uk/10.3390/s22166020

AMA Style

Montenegro G, Chacón R, Fabregas E, Garcia G, Schröder K, Marroquín A, Dormido-Canto S, Farias G. Modeling and Control of a Spherical Robot in the CoppeliaSim Simulator. Sensors. 2022; 22(16):6020. https://0-doi-org.brum.beds.ac.uk/10.3390/s22166020

Chicago/Turabian Style

Montenegro, Guelis, Roberto Chacón, Ernesto Fabregas, Gonzalo Garcia, Karla Schröder, Alberto Marroquín, Sebastián Dormido-Canto, and Gonzalo Farias. 2022. "Modeling and Control of a Spherical Robot in the CoppeliaSim Simulator" Sensors 22, no. 16: 6020. https://0-doi-org.brum.beds.ac.uk/10.3390/s22166020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop