Next Article in Journal
A Machine Learning Approach to Detect Parkinson’s Disease by Looking at Gait Alterations
Previous Article in Journal
Deriving Fuzzy Weights from the Consistent Fuzzy Analytic Hierarchy Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Heuristic Construction Neural Network Method for the Time-Dependent Agile Earth Observation Satellite Scheduling Problem

1
School of Computational Science and Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
2
College of Systems Engineering, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Submission received: 26 August 2022 / Revised: 16 September 2022 / Accepted: 20 September 2022 / Published: 25 September 2022

Abstract

:
The agile earth observation satellite scheduling problem (AEOSSP), as a time-dependent and arduous combinatorial optimization problem, has been intensively studied in the past decades. Many studies have proposed non-iterative heuristic construction algorithms and iterative meta-heuristic algorithms to solve this problem. However, the heuristic construction algorithms spend a relatively shorter time at the expense of solution quality, while the iterative meta-heuristic algorithms accomplish a high-quality solution with a lot of time. To overcome the shortcomings of these approaches and efficiently utilize the historical scheduling information and task characteristics, this paper introduces a new neural network model based on the deep reinforcement learning and heuristic algorithm (DRL-HA) to the AEOSSP and proposes an innovative non-iterative heuristic algorithm. The DRL-HA is composed of a heuristic construction neural network (HCNN) model and a task arrangement algorithm (TAA), where the HCNN aims to generate the task planning sequence and the TAA generates the final feasible scheduling order of tasks. In this study, the DRL-HA is examined with other heuristic algorithms by a series of experiments. The results demonstrate that the DRL-HA outperforms competitors and HCNN possesses outstanding generalization ability for different scenario sizes and task distributions. Furthermore, HCNN, when used for generating initial solutions of meta-heuristic algorithms, can achieve improved profits and accelerate interactions. Therefore, the DRL-HA algorithm is verified to be an effective method for solving AEOSSP. In this way, the high-profit and high-timeliness of agile satellite scheduling can be guaranteed, and the solution of AEOSSP is further explored and improved.

1. Introduction

The rapid development of space technology has greatly improved the performance of earth observation satellites, enabling them to carry more payloads and have better maneuverability. As a new generation of imaging satellites, an agile satellite with the capabilities of roll, pitch, and yaw possesses greater mobility than traditional satellites. The observable range of the agile satellite is a zonal area with the centerline which is the sub-satellite track. Its roll angle and pitch angle jointly determine the observable range. In Figure 1, the observable range of the satellite is the area surrounded by the pitch boundary at the front and rear and the roll boundary at the left and right. Additionally, the upgrades of these instruments allow that the time shooting an observed target could be not completely constrained by satellite over-the-top time, thereby extending the visible time window for an observed target. In this way, the starting observation time can be chosen among a longer time interval. However, when converting to a new observation, the satellite requires some time, which we call transition time in this paper, to maneuver from a previous position to another one [1]. Any change in observation angles and positions between two consecutive tasks affects the starting time of the task, rendering this problem highly time-dependent. Thus, greater mobility creates more flexible observation angles and time-dependent features, magnifying the difficulty of the AEOSSP.
Many studies have proposed non-iterative heuristic construction algorithms and iterative metaheuristic algorithms to solve the AEOSSP. When designing these two types of heuristic algorithms, scholars should comply with some principles, such as the earliest start principle [2], the least possible conflict principle [3], and the maximum imaging quality principle [4], etc. With these principles, Lemaitre et al. [5] studied the Pleiades agile satellite launched by the French Space Agency and comprehensively described the AEOSSP for the first time. Based on Lemaitre’s research, numerous related algorithms have been proposed.
The heuristic construction algorithm contains a systematic structure and does not generate current status information in the solution process. With these algorithms, the accepted approximate solutions can be obtained within a short time. Many studies were made for developing the heuristic construction algorithms by establishing distinct models [6,7,8] with different theories and introducing indicators based on different task features [9]. However, because these algorithms’ systematic structures do not update the historical scheduling information in the solution process, global information features are lost and the AEOSSP is not solved effectively. For addressing these defects, the iterative metaheuristic algorithms that evaluate the local solution and update the historical scheduling information are tendered. They are similar to genetic and simulated annealing algorithms with certain randomness and generality. Because genetic algorithms are highly adaptable when combined with other models and algorithms, most of them belong to the class of improved genetic algorithms [10,11,12,13], raising the quality of the solutions. Furthermore, while some algorithms are designed for single-objective AEOSSPs, metaheuristic algorithms can also be utilized for multi-objective optimization [12,14,15]. With these iterative metaheuristic algorithms, the solutions are outstanding, but the surplus time and space complexity cause their inferior efficiency.
In conclusion, there are some inevitable weaknesses in both two types of algorithms. The heuristic construction algorithms spend a relatively shorter time at the expense of solution quality, while the iterative metaheuristic algorithms accomplish a high-quality solution with a lot of time. The traditional heuristic algorithms usually calculate the priority of each task according to the task characteristics and then schedules them based on the acquired priority. The static characteristics of the task are usually adopted in the heuristic functions. In general, when the observation target is determined, the constant parameters obtained through simulation calculation are considered as static characteristics, such as the return of the task, the start time of the visual time window (VTW), the duration time of the task, etc. During the whole scheduling process, these characteristics remain a fixed value. The flow of this type of algorithm is shown in Figure 2. Firstly, the scheduling scenario is initialized. Then, the priority sequence of the tasks to be observed is obtained through the heuristic function.
Generally, only static characteristics of the task are taken into consideration in the traditional heuristic algorithm. Since its solution process does not produce current state information, there is no historical scheduling information, which is significant to improve the effectiveness of the solution. In addition, most of the global information characteristics would be lost.
Based on the existing research about the development of artificial intelligence, the new neural network models provide new ideas for solving AEOSSP problems. The AEOSSP aims to rationally allocate limited resources to observation tasks for satisfying the users’ diversified demands and maximizing schedule profits, which is a classical combinational optimization problem. Many studies [16,17,18,19,20] were conducted to verify the effectiveness of neural network models for solving scheduling problems. Among them, the deep reinforcement learning utilized in the heuristic algorithm (DRL-HA) [21] was gradually developed by a series of research. Vinyals et al. [22] proposed a printer network (PN), applied to the sequence-to-sequence (Seq2Seq) structure with encoding and decoding as two recurrent neural networks (RNNs). PN applies an attention mechanism, which reaches a one-step decision by analyzing the outputs of encoding and decoding. Bello et al. [23] developed the attention mechanism and used the strategy gradient method and actor-critic algorithm (AC) for training the network model. Nazari et al. [24] simplified the encoding process of the pointer network by abandoning the RNN structure and adopted different attention mechanisms for the decoder. Compared to the previous research, the efficiency is greatly improved. However, there is one inevitable weakness of this model: one vehicle’s dynamic parameter remains an identical value for individual tasks, rendering the deviation of the solution. Kool [21] solved this defect and proposed an innovative attention model (AM) by introducing Transformer, a natural language process. Unlike the previous research, which abandons RNN, this study took Transformer encoder’s multihead-attention (MHA) as the model’s encoder. In this way, the embedding layer is able to gain the individual task characteristic information and enter it into the MHA layer for obtaining global variables of the current task set. Furthermore, this model interacts with the outputs of the encoder and decoder to get the matching degree of tasks, which is beneficial for selecting the suitable task. According to these studies, a neural model based on DRL-HA can be utilized to solve TSP, VRP, backpack, and other NP-hard problems. Therefore, based on these studies, this paper introduces the DRL-HA model to the AEOSSP problems and investigate its validity.
Because the AEOSSP is a NP-hard problem, for the traditional heuristic algorithms, the search difficulty and time to obtain the solution increase sharply with the increase of the problem scale. Thus, the traditional scheduling algorithms are not able to satisfy the requirements regarding the high-profit and high-timeliness of agile satellite scheduling, such as the ability to create rapid adjustments for emergency tasks. Meanwhile, in large-scale scheduling scenarios, the solution time of the traditional intelligent optimization algorithm is unacceptably lengthy. Additionally, in a specific scenario, the demand of user observation has a certain regularity, while the existing algorithms do not take this phenomenon into the consideration. With the continuous application of the DRL-HA in combinational optimization problems, the advantages of high efficiency and effective utilization of historical data are gradually discovered. Therefore, the research on agile satellite scheduling based on DRL-HA is consequential to improve the emergency response ability, rapid response ability and efficiency of satellite–earth observation system, which potentially possess values in theory and application. With this innovative algorithm, when agile satellites implement tasks such as environmental surveillance, disaster relief, and intelligence reconnaissance, the high-profit and high-timeliness of agile satellites tasks can be guaranteed.
From the analysis above, the traditional algorithms are not able to render time and quality to be satisfied simultaneously. When these traditional algorithms are applied, large-scale task scheduling scenarios usually require an unacceptable amount of time. Furthermore, neither the characteristic information of the task set nor historical scheduling scenarios are comprehensively taken into consideration. When switching to a new scheduling scenario, the algorithms search and iterate again, which wastes a large amount of time. Therefore, to rectify these disadvantages, the following problems should be solved:
(1)
How to comprehensively deploy dynamic task information and scheduling scenario information;
(2)
How to establish the DRL-HA model aiming to solve the AEOSSP;
(3)
How to utilize convoluted constraints for establishing an efficient model.
This study utilizes the DRL-HA model to comprehensively deploy the task characteristic information and historical scheduling information for solving AEOSSP with a high-quality solution and relatively less time.
The remainder of this paper is organized as follows. In Section 2, we put forward a mathematical description of the AEOSSP. In Section 3, a heuristic algorithm based on DRL-HA for AEOSSP is introduced in detail. We set up experiments to verify the effectiveness of the proposed algorithm in Section 4, and the experimental results are displayed in Section 5. The conclusions are drawn in Section 6.

2. Problem Analysis and Modeling

2.1. Problem Description and Assumptions

Agile satellites possess prominent capabilities for earth observation, but observation resources are still relatively scarce compared to the increasing observation requirements. The agile satellite scheduling problem (AEOSSP) is a comprehensive planning and scheduling problem, whose purpose is to develop an optimal observation plan for a satellite with the maximum observation profit under the current set of tasks to be observed. To study the ability to solve the AEOSSP by the DRL-HA model, when establishing the corresponding mathematical model, the following assumptions should be made:
(1)
It is necessary to decompose the complex tasks such as stereo imaging tasks and regional target tasks into point target tasks for processing because, in this way, the tasks considered in this paper are all point target tasks, which can be completed within the visible time window (VTW). Since the AEOSSP is a time-dependent problem, the VTW for each point target is significant for analysis and modeling.
(2)
This paper aims at the single-orbit AEOSSP and neglects the charging process in the sun area and the data download process. These assumptions were made because they are not the conditions of concern for the present research. These excessive conditions are unnecessary and bring dispensable costs.

2.2. Fundamental Objective Function

x i : the scheduled status of the task. When x i is equal to 1, the task is selected to be completed. Otherwise, it is not scheduled since it violates the constraint. t a s k i : the i-th task in the task set to be scheduled.
x i = 1 , t a s k i s c h e d u l e d 0 , o t h e r w i s e
The fundamental object of the AEOSSP is to maximize the total profit obtained from task requirements. The profit is measured by the sum of the profit of the individual observation task, expressed as:
M a x P = m a x i N p i x i
  • P: the total profit of the total satellite observation tasks.
  • p i : the profit of t a s k i .

2.3. Constraints

The observation time of each task should be within VTW of the task:
t i s t w i s 0
t w i e t i e 0
  • t w i : the VTW of t a s k i .
  • t w i s : the start time of t w i .
  • t w i e : the end time of t w i .
  • t i s : the actual start time of t a s k i .
  • t i e : the actual end time of t a s k i .
The actual observation of each task should meet the required duration of the observation:
t i e ( t i s + l i ) 0
  • l i : the duration time required for t a s k i to be observed.
The time interval between the two adjacent observation tasks t a s k i and t a s k j in the priority must be longer than the transition time t r a n s i j :
t j s ( t i e + t r a n s i j ) 0
  • t r a n s i j : the transition time from t a s k i to t a s k j where t a s k j is the next task of t a s k i .
The total storage occupied by the observation tasks should be within M.
M i N m i 0
  • M: the total storage of the satellite.
  • m i : the storage required for a successful observation t a s k i .
The total power consumed by the observation task should be within E.
E i N e i 0
  • E: the power of the satellite.
  • e i : the power required for a successful observation t a s k i .

2.4. Attitude Transition Angles and Time

In the agile satellites, the transition time is calculated based on attitude angles. In the actual scene, the agile satellite attitude transition is a multi-stage process from slowly accelerating to a constant speed, and lastly decelerating. The required stabilization time should be counted after the attitude control action is completed and the transition time is usually calculated based on the method of attitude angle synthesis. This paper utilizes a piece-wise linear function to approximate the time-dependent attitude transition time. When the total attitude angle changes with few degrees, the attitude maneuver time holds a fixed value. When the total attitude angle change reaches a certain threshold, the linear function is used to calculate the attitude maneuver time.
t r a n s i , j k = λ 0 , Δ θ i , j θ 1 Δ θ i , j v 1 + λ 1 , θ 1 < Δ θ i , j < θ 2 Δ θ i , j v 2 + λ 2 , θ 2 < Δ θ i , j < θ 3 Δ θ i , j v 3 + λ 3 , θ 3 < Δ θ i , j < θ 4 Δ θ i , j v 4 + λ 4 , θ 4 < Δ θ i , j
t w i s , t w i e , p i , l i , t i s , t i e , E , e i , t r a n s i , j 0
Δ θ i , j = | θ i r o l l _ s θ i p i t c h _ s | + | θ i r o l l _ e θ i p i t c h _ e |
  • λ 0 , λ 1 , λ 2 , λ 3 , λ 4 : the required stabilization time for satellite attitude transition at the different stages.
  • Δ θ i , j : the total attitude angle changes from t a s k i to t a s k j and the threshold value to calculate the attitude transition time.
  • θ 1 , θ 2 , θ 3 , θ 4 : the thresholds of attitude transition angles at the different stages.
  • v 1 , v 2 , v 3 , v 4 : the attitude transition speeds at the different stages.
  • θ i r o l l _ s : the start roll angle.
  • θ i r o l l _ e : the end roll angle.
  • θ i p i t c h _ s : the pitch angle.
  • θ i p i t c h _ e : the end pitch angle.
According to the research about GRILS [25], Table 1 displays these constants for attitude transition and task parameters.

3. Heuristic Scheduling Algorithm for Deep Reinforcement Learning

The traditional heuristic algorithm usually consists of two parts: calculating the priority of each task and scheduling them according to the priority of each task. According to the the characteristics of the traditional input task, this paper proposes an innovative heuristic algorithm based on the deep reinforcement learning model (DRL-HA).

3.1. DRL-HA

The heuristic scheduling algorithm based on the DRL-HA utilizes a neural network to parameterize the task priority. The framework of the algorithm is shown in Figure 3. Starting from the current scheduling scenario, we firstly initialize the decision point, sort the tasks in the task set through the trained heuristic construction neural network (HCNN) model, and then select the task with the highest priority.

3.1.1. HCNN Model

The HCNN is based on the sequence-to-sequence (Seq2Seq) structure, which includes an encoder, decoder, and global function.
(1)
Encoder: the encoder’s objective is to convert the static feature parameters of the to-be-scheduled tasks into a high-dimensional parameter eigenvector by one-dimensional convolution. In this way, the network is able to exhaustively exploit the parameter information of the task to make a reasonable decision. Because each imaging satellite task is independent, the output of the neural network is irrelevant to the sequence of the task data. Therefore, a simple embedding layer is necessary to solve this problem, which additionally reduces the time and space complexity of the encoder model. In this study, the features of a task are composed of the static feature s and the dynamic feature d. The static features are attribute features, including task observation time window t w i , profit p i , duration l i , attitude angle parameters θ i r o l l _ s , θ i r o l l _ e , θ i p i t c h _ s , θ i p i t c h _ e , and so on. The dynamic features are the to-be-scheduled status: the value is 0 or 1; 0 means available, while 1 means unavailable.
(2)
Global variables: since the eigenvector of an individual task is independent in the encoder, these variables are not able to indicate the interaction features between tasks. Therefore, this study applies an improved multi-head attention for extracting related features. Figure 4 displays the structure of the neural network. By inputting the eigenvector set of the task set, we get the variable G, which contains the historical scheduling information between tasks.
(3)
Decoder: unlike the encoder, each decoding step points to the input task. In this research, because selecting the current task is based on previous historical decision information, the HCNN model accomplishes the decoding process through this historical decision information. To obtain the historical decision variables, a neural network module with a memory function is required to realize the acquisition of historical decision information. Fortunately, the RNN network is a great choice for solving this problem. In this algorithm, the gated recurrent unit (GRU) network structure [26] is selected for achieving the extraction of the historical decision information.
(4)
Attention mechanism: the fundamental goal of the attention mechanism is to select the information that is the most compatible with the current target and has the highest degree of attention from numerous information. Based on the encoder-decoder structure, this mechanism utilizes the neural network model for completing feature interaction between the outputs from the encoder and decoder. In this way, the related distribution of the next task is obtained. At the softmax layer, this related distribution is able to be converted into the probability distribution diagram of task selection, which is greatly advantageous for selecting the next task.
The overall structure of the HCNN model is shown in Figure 5. The blue cells are the structural parts of this model; the dynamic process, where dynamic parameters are updated during the process, is indicated by the orange arrows and cells; there is one black arrow from the probability function to the embedding layer, representing the task which should be concealed; the grey arrows and cells demonstrate that all parameters are constant.
At the encoder layer, the static parameter eigenvector H s and the dynamic parameter eigenvector H d are obtained by embedding. At the decoder layer, the historical decision information h t is obtained by GRU.
In the HCNN model, this study designs two attention layers for selecting tasks. In the first layer, the total feature parameter M is obtained from the connection layer in formula (12). Furthermore, we utilize the Glimpse network structure [23] to complete feature interactions and attention for h t and M. By originally filtering the current task set, the weighted context variable c t of the current scene task set is obtained. It should be mentioned that DRL-HA uses a masking mechanism [19] to conceal tasks that have been ordered to ensure that there is no duplicate selection of tasks. In this figure, the only black arrow reflects that this model has been scheduled as s1 and this task should be concealed.
M = w 1 [ H s , H d ]
c t = G l i m p s e ( M , h t )
  • [ H s , H d ], [ g s i , t s i , c t ]: indicating the variables are joined together.
  • w 1 , v 1 i : the to-be-trained network parameters.
The second attention layer achieves feature interaction and attention for the global variable G s , the total feature parameter M and the context variable of the task set c t for the matching degree of the task. Lastly, all variables are normalized by the softmax layer and the selected probability of the to-be-selected task under the current decision situation is output.
u ˜ t i = v 1 i t a n h ( w 2 [ g s i , t s i , c t ] )
P ( y t + 1 | Y t , X t ) = s o f t m a x ( u ˜ t i )
To guarantee the high-quality solution and the generality of the HCNN model, this paper employs random and greedy strategies for training and testing.

3.1.2. Task Arrangement Algorithm

The task arrangement algorithm (TAA, Algorithm 1) is used to arrange the specific execution time of observation tasks after the task sequence is formed by the HCNN model. In order to decrease the training cost of DRL-HA, this paper adopts the following TAA for model evaluation and training. When the task is selected to be executed, the earliest VTW start time and end time are determined for the task, and this part of the selected time window resources is updated so that they cannot be used by other tasks. TAA mainly includes checking constraint conditions, selecting the task execution location, and updating the time window. The pseudocode of TAA is the following.
Algorithm 1 TAA
  • Input:
  •   Task priority List L = { t a s k i | i = 1 , 2 , . . . , l }
        The available time windows W;
  • Output:
  •   Task execution List L
        Initialize task execution List L
  •   for task in L do
  •    for time window in W do
  •       if task satisfy all constraints then
  •          L task // Select the most ahead position for task by the earliest VTW start time
  •       end if
        Update VTW
  •     end for
  •   end forreturn L
Constraint checking, which is used for determining whether the task can be executed, is the first part of TAA. The tasks will be handled in the observation selection process. Some tasks will be executed, while others will be deleted. In the task scheduling process shown in Figure 6, the task currently being scheduled has a task observable time window, and its length covers the task’s observation duration, so the task is capable of being executed.
In the task arrangement shown in Figure 7, there are two observable time windows for the task being arranged currently. However, the length of these two times windows cannot cover the whole task duration, indicating that the task is not feasible for execution and should be deleted from the to-be-scheduled task set.

3.2. Training Methods

The HCNN model is a decision model, and we apply a policy gradient method and the actor-critic (AC) algorithm (Algorithm 2) [23] to train the parameters of this decision model.
Algorithm 2 Actor-critic algorithm.
1:
Initialize actor network parameters θ
2:
Initialize critic network parameters θ c
3:
Initialize the priority list L
4:
for i t e r a r i o n 1 , 2 , do
5:
    reset gradients: d θ 0 , d θ c 0
6:
    generate N problem instances from ϕ
7:
    for  k = 1 , 2 , N  do
8:
        step counter t 0
9:
        while not terminated do
10:
           select next variable x t + 1 k according to the probability distribution P ( y t + 1 k | Y t k , X t k )
11:
           update new state X t + 1 k to X t k
12:
            y t + 1 is added to L
13:
            t t + 1
14:
        end while
15:
        calculate the reward R k by algorithm T A A to L
16:
    end for
17:
     d θ 1 N k = 1 N ( R k V ( X 0 k ; θ c ) ) θ log P ( Y k | X 0 k )
18:
     d θ c 1 N k = 1 N θ ( R k V ( X 0 k ; θ c ) ) 2
19:
    update θ using d θ and θ c using d θ c
20:
end for
The objective of reinforcement learning is to maximize the reward from the environment. However, the output of the HCNN model is a sequence of task priority, which is not an explicit reward. To evaluate the model and obtain a reward appropriately, the profit of the TAA after the HCNN model is selected to be the reinforcement reward function value.
During the training, the parameters θ of the HCNN model are updated in each iteration for approximating the optimal parameters θ * as closely as possible, which could render the reinforcement reward function value to become the optimal R * .
In the training process of each iteration, for each case, the task scheduling sequence is generated by the current HCNN network parameters. Then, the profit of this scheduling sequence is calculated by the TAA algorithm to get the reward value, which reinforces learning needs. The calculation process shown in step 17 is used to update the actor-network, where V ( X 0 k ; θ c ) is the output value of the critic with the current parameter θ c . After observing the difference between the estimated and the true value, the critic network is updated at step 18.

4. Experiment Study

4.1. Experiment Setup

4.1.1. Dataset

The agile satellite scheduling problem (AEOSSP) is an arduous combinatorial optimization and complex engineering problem. The observation time window of each task and the acquisition of various parameters such as pitch value at each moment within the time window require complex dynamic equation calculations. Considering the lack of a benchmark for the AEOSSP and the requirements of the model training dataset, we designed simulation scenarios and generated a dataset based on Liu’s study [1]. Specifically, we applied the regional observation target used in [1] as our task to be scheduled, and all targets were spread according to a uniform random distribution. On the simulation day, we set eight orbits from the satellite and defined the O-ith to distinguish the ith orbit. The scene where the target can be observed on the first orbit (O-1) was considered the training dataset. A total of 200,000 scenes were generated as the training dataset. The validation dataset consisted of 1000 scenes generated independently.
The maximum pitch, roll, and yaw angles of the satellite were 45 , 45 , and 90 , respectively. The scheduling time range was from 00:00:00 to 24:00:00 on a certain day, a total of 24 h. The six orbital parameters of the satellite were the semimajor axis (a), eccentricity (e), inclination (i), perigee angle ( ω ), right ascension of ascending node ( Ω ), and angle of horizontal approach (m). The satellite orbit parameter setting is shown in the following Table 2.

4.1.2. Competitors

To verify the effectiveness of the algorithm proposed in this paper, we designed the following heuristic algorithms: Random, the earliest visual time window (VTW) start time first (WF), the profits first (PF), the profits/duration first (PDF), and the profits-window first (PWF) algorithm for comparisons. The details are as follows:
(1)
Random: Randomly sort all to-be-scheduled tasks, insert the scheduling plan in order, and check whether the constraints are satisfied until the scheduling plan cannot accommodate any task.
(2)
WF: Sort all to-be-scheduled tasks from the earliest start time of the task’s VTW to the latest start time of the task’s VTW, insert the scheduling plan in order, and check whether it satisfies constraints until the scheduling plan cannot accommodate any task.
(3)
PF: Sort all to-be-scheduled tasks from high profit to low profit, insert the scheduling plan in order, and check whether the constraints are satisfied until the scheduling plan cannot accommodate any task.
(4)
PDF: Before sorting, divide the profits of all to-be-scheduled tasks by their respective observation duration time to obtain the unit observation time profits of the tasks. Sort these tasks from high unit observation time profit to low unit observation time profit, insert the scheduling plan in order, and check whether the constraints are satisfied until the scheduling plan cannot accommodate any task.
(5)
PWF: Sort all to-be-scheduled tasks from high profit to low profit, and when the tasks possess the same profit, sort them from the earliest start time of the task’s VTW to the latest start time of the task’s VTW. Insert the scheduling plan in order, and check whether the constraints are satisfied until the scheduling plan cannot accommodate any task.
In addition, to verify the practicability of the heuristic construction neural network (HCNN) model as the initial solution provider of the meta-heuristic algorithm, we utilized two meta-heuristic representative iterative search algorithms, the tabu search (TS) algorithm [27,28,29,30,31,32], simulated annealing (SA) algorithm [33,34] and greedy randomized iterated search (GRILS) [25] for testing. SA and TS are classical algorithms for solving combinatorial optimization problems but are not designed particularly for solving AEOSSP. GRILS is a state-of-the-art algorithm for AEOSSP since it introduces the domain structure and INSERT operator according to the characteristics of the AEOSSP. The TS, SA and GRILS with the trained HCNN model as the initial solution are named TS-HCNN, SA-HCNN and GRILS-HCNN, respectively. Similarly, the initial solutions with other heuristic algorithms mentioned above are considered “competitors”: the selected algorithms that are compared with the algorithm proposed in this study. Since the other three algorithms are not prevailing and classical, WF and PF are selected as the representative competitors. For example, the TS algorithm with the WF output was selected as the initial solution (TS-W), and the SA algorithm with the PF output was selected as the initial solution (SA-P). In TS and SA, new solutions are both generated by the neighborhood operation, which reverses two random tasks in the current solution. Table 3 displays all competitors.
There are 12 algorithms in the comparison algorithm experiment, including HCNN, WF, PF, GRILS, TS-HCNN, TS-W, TS-P, SA-HCNN, SA-W, SA-P, G-W, and G-P, and the maximum time is set as 10 min. In the TS algorithm, the neighborhood size is equal to the quadruple scale scene, and the local optimum is not allowed in 5 iterations; the initial temperature of SA is 4000 , and the temperature decay rate is 0.999. The SA algorithm stops when the temperature is less than 0 . 001 .

4.2. Algorithm Parameters

HCNN employs the AC algorithm to train the model. The network model parameter settings of the actor in the AC algorithm are shown in Table 4. The network model parameter settings of the critic are shown in Table 5. There is a ReLU activation between two feed-forward layers.
This topic used the Adam optimizer [35] to train the network of participants and reviewers. The initial learning rate η was 0.0005, and the number of training epochs was 10. Due to the different scales of training data and GPU utilization, the corresponding batch training scale was used for different scene scales. All experiments in this paper used a single GPU RTX 2080-Ti, i9-9900k CPU, and 64 GB memory. The compiler language was Python 3.7 (Python Software Foundation: Wilmington, DE, USA); the deep learning framework used Pytorch 1.2.0 (Linux Foundation: San Francisco, CA, USA). They are both open sources and free.

5. Experiment Results

In this section, we designed three kinds of experiments to verify the performance of the deep reinforcement learning and heuristic algorithm (DRL-HA): (1) utilizing DRL-HA as a heuristic solving algorithm; (2) testing this algorithm’s data distribution sensitivity; (3) utilizing the HCNN as an initial solution of the meta-heuristic algorithm. The algorithm evaluation requires the following indicators:
(1)
The average task total profits (P). When an algorithm can obtain a better solution in a large number of scenarios, the superiority of high-quality solutions and the stability of the algorithm are verified.
(2)
The total running time of the algorithm (T). The running time of the algorithm is the required running time for obtaining the scheduling sequence. It fully reflects the timeliness performance of the algorithm for the agile satellite problem (AEOSSP). In addition, as the scale of the problem increases, the time complexities of the different algorithms become different, and the applicable scenarios of the algorithm will also change accordingly. It should be mentioned that the time of the DRL-HA algorithm refers to the time that the trained network runs on the test set. For the following data of T, its unit is second(s).

5.1. Results of DRL-HA

To validate the generalization capabilities of the DRL-HA, we trained 25-, 50-, 75-, 100-, 125-, and 150-size HCNN models for DRL-HA. These sizes of model are representative, and the tests with these sizes can comprehensively reflect the generalization capabilities. For the test data, we generate 100 instances with 25, 50, 75, 100, 125, 150, 175, and 200 tasks, respectively, according to the O-1 dataset, which is the same as the training data.
The average task total profits and total running time of DRL-HA and its competitors are shown in Table 6 and Table 7. In these two result tables, H-25 is used to represent the HCNN model trained under the scale of 25, and the HCNN models trained by other distinct scales are expressed in the same way.
From the results, it can be seen that the DRL-HA algorithm can obtain the best value in eight scenarios with different scales, and the best HCNN model is 9.15 % higher than the best competitor. The average results of 100 scenarios also fully reflect the stability and superiority of the DRL-HA to agile satellite scheduling scenarios. Figure 8 displays the average profits and total time for the HCNN at different size. The HCNN models trained by different scale datasets have similar performance when meeting the same scale test scenarios, indicating that there is no essential difference in the performances of HCNN models trained on different scales. Additionally, the Figure 9a shows the average profits of each algorithm in different scale scenarios. Obviously, the trained HCNN model can directly solve scheduling scenarios with different task scales and with great performance.
In terms of the running time, the experimental results of each algorithm in different scale scenarios are shown in Figure 9b. The data for “HCNN” are the average values from all HCNN models. Obviously, the running time of the DRL-HA algorithm matches the level of the magnitude and the calculation trend compared to other heuristic algorithms. Therefore, the DRL-HA algorithm does not increase the time complexity.
In summary, DRL-HA not only accomplishes better performance in terms of task profits but also has the advantage of running time, which demonstrates the high efficiency of the DRL-HA as a solving algorithm.

5.2. Results of Sensitivity Experiments

To validate the generalization capability of DRL-HA under different task distributions, we trained it at the scheduling scenarios with scales of 50and 100 in this paper, and 100 scheduling instances were generated for the other seven orbits (distinguished by the number of orbits, where O-1 is the test/training orbit). These scales of samples are representative from the small to large scale. Since this test was designed mainly for comparing competitors instead of performance on different scales, these samples are appropriate. The random algorithm, WF, PF, PDF, and PWF were compared with the HCNN.
In order to study the performance in a single-orbit with a large-scale task scenario, this experiment used a scale of 100 single orbit scheduling instances and generated 100 sets of scheduling instances for different orbits. The average task total profits and the total running time of different orbits in different scenarios are shown in Table 8 and Table 9, while Figure 10 displays the line trend charts. From these tables and figures, at every orbit and different scenario sizes, the HCNN obviously accomplishes more average profits than other algorithms. In addition, the HCNN works better with the increasing scenario sizes. For example, the HCNN achieves 24.74 % and 139.86 % better average profits than Random in G2 for the 25 scenario task and 100 scenario task, respectively. Therefore, the DRL-HA model achieves better performance than other algorithms and the best results in large-scale instances in different orbits, indicating that the model has great generalization ability for different scenario sizes and task distributions.

5.3. Results of HCNN as the Initial Solution

To validate the generalization capabilities of the HCNN as the initial solution of the iterative meta-heuristic algorithm, we conducted comparative experiments with TS and SA algorithms with task scales of 50, 100, 150, and 200. These scales of samples are representative and sufficient for comparing two algorithms. Since these algorithms possess randomness, we ran them ten times and calculated their average results. For the tabu search (TS), when the historical optimal solution is not updated for 100 iterations, this algorithm stops and outputs the historical optimal solution. Each iteration of the TS algorithm searches multiple domain solutions, but only one single algorithm is performed in the simulated annealing (SA) algorithm. Thus, we set the iteration to have no improvement for 1000 times for SA. In addition, due to the randomness of these meta-heuristic algorithms, we ran these algorithms 30 times and calculated the standard deviation (std.) to compare their stability. The parameters of the greedy randomized iterated search (GRILS) are almost identical with the study [25]. However, since this paper focuses on the AEOSSP for the single orbit, and the maximum number of completed tasks is much less than that of [25], we therefore set NumoFlterNolmp”, which is the number of interactions without the updated optimal solution, as 10.
The average task total profits, the standard deviation, and the total running time of HCNN and its competitors are shown in Table 10. From this table, it can be seen that the profit of the HCNN is obviously higher than those of the other competitors; for example, G-HCNN achieves more profits than G-W and G-P. Besides, the std. of the HCNN is mostly smaller than its competitors, indicating that the HCNN can maintain a stable and high-quality initial solution for different task scales and corresponding meta-heuristic algorithms. Figure 11 displays the average profits for all algorithms as an initial solution, showing that the performance of the algorithms with GRILS is better than that of other competitors, and the G-HCNN completes the most average profits. For example, in size-25, because G-W achieves more 17.65 % average profits than TS-W and G-P achieves more 20.67 % average profits than SA-P, the GRILS algorithm reveals its superior performance to the other two algorithms. Furthermore, in size-200, TS-HCNN achieves 23.33 % and 10.21 % improved performance compared to TS-W and TS-P, respectively, suggesting that the HCNN model as the initial solution can better ameliorate the performance of the iterative meta-heuristic algorithm than other two algorithms.
Although the results of three different GRILS algorithms are similar, the advantages of G-HCNN can still be discovered. To further explore the differences among these three algorithms, we count the average number of iterations of ten solution processes, shown in Figure 12. From this figure, it can be obtained that G-HCNN maintains the advantage of quality during the algorithm operations, because the profit quality at the early iterations of GRILS achieves the same level of the profit quality at the later iterations of other competitors, and the total number of iterations are relatively smaller. This finding demonstrates that HCNN can output the effective task priority sequence, select high-quality tasks, and provide high priority for observation, which accelerates the iterations of GRILS. Furthermore, during the initialization of the algorithms, the quality of the initial solutions by the three algorithms has been improved, which benefits from the INSERT operator of GRILS algorithm. The INSERT operator is able to arrange tasks for observation as much as possible but causes the algorithm to iterate slowly. In the later iterations, the profit gap among three algorithms gradually decreases, indicating the eminent optimization ability of GRILS in the AEOSSP.
In summary, the trained HCNN model as an initial solution of the iterative meta-heuristic algorithm is not only able to accomplish better profits than competitors but also costs a reduced number of iterations compared to others.

6. Conclusions and Future Directions

In this paper, a heuristic scheduling algorithm based on the deep reinforcement learning and heuristic algorithm (DRL-HA) is proposed based on the historical scheduling and task characteristic information to solve the agile earth observation satellite scheduling problem (AEOSSP) with high efficiency. The final scheduling sequence is obtained by combining the DRL-HA model trained by the heuristic construction neural network (HCNN) and the task arrangement algorithm (TAA). Three kinds of experiments verify that the algorithm has outstanding performance and generalization ability. Among them, experiments of instances with different scales show that (1) the profit of the DRL-HA algorithm is 9.15 % higher than that of the best heuristic algorithm, on average; (2) moreover, this algorithm model has a good generalization ability, demonstrating this algorithm process’ outstanding performance in the experiments with different task scales and task distributions; (3) the algorithm is capable of providing high-quality initial solutions with a smaller number of iterations compared to results from other competitors. The model trained under a fixed scale and task distribution is able to deal with other scenarios with eminent performance and maintains the superiority of the algorithm.
The DRL-HA model proposed in the paper exhaustively unitizes historical scheduling information and task feature information to obtain the prominent scheduling sequence. Therefore, the DRL-HA algorithm is verified to be an effective method for solving AEOSSP. In this way, the high profit and high timeliness of agile satellite scheduling can be guaranteed, and the solution of AEOSSP is further explored and improved. Although this model requires the support of hardware and much time for training, its great generalization capability renders it suitable for different scenarios, and producing a high-quality initial solution provides this model the ability to be combined with other algorithms. In future work, we will try to combine deep reinforcement learning with the single-satellite algorithm to accomplish refined task allocation for multi-satellite scheduling problems.

Author Contributions

Conceptualization, J.C., M.C. and X.L.; Data curation, J.C. and J.W.; Formal analysis, J.C. and L.H.; Funding acquisition, L.H.; Investigation, J.C. and M.C.; Methodology, J.C., M.C. and L.H.; Project administration, J.C.; Resources, J.C.; Supervision, X.L.; Validation, J.C.; Visualization, J.C.; Writing—original draft, J.C., M.C. and J.W.; Writing—review & editing, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Hunan Provincial Innovation Foundation for Postgraduate CX20210034 and the National Natural Science Foundation of China (72001212).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, X.; Laporte, G.; Chen, Y.; He, R. An adaptive large neighborhood search metaheuristic for agile satellite scheduling with time-dependent transition time. Comput. Oper. Res. 2017, 86, 41–53. [Google Scholar] [CrossRef]
  2. Wang, X.; Han, C.; Zhang, R.; Gu, Y. Scheduling multiple agile earth observation satellites for oversubscribed targets using complex networks theory. IEEE Access 2019, 7, 110605–110615. [Google Scholar] [CrossRef]
  3. Liu, Z.; Feng, Z.; Ren, Z. Route-reduction-based dynamic programming for large-scale satellite range scheduling problem. Eng. Optim. 2019, 51, 1944–1964. [Google Scholar] [CrossRef]
  4. Beaumet, G.; Verfaillie, G.; Charmeau, M. Feasibility of autonomous decision making on board an agile earth-observing satellite. Comput. Intell. 2011, 27, 123–139. [Google Scholar] [CrossRef]
  5. Lemaıtre, M.; Verfaillie, G.; Jouhaud, F.; Lachiver, J.M.; Bataille, N. Selecting and scheduling observations of agile satellites. Aerosp. Sci. Technol. 2002, 6, 367–381. [Google Scholar] [CrossRef]
  6. De Florio, S. Performances optimization of remote sensing satellite constellations: A heuristic method. In Proceedings of the 5th International Workshop on Planning and Scheduling for Space (IWPSS 2006), Baltimore, MD, USA, 22–25 October 2006. [Google Scholar]
  7. Wang, P.; Reinelt, G.; Gao, P.; Tan, Y. A model, a heuristic and a decision support system to solve the scheduling problem of an earth observing satellite constellation. Comput. Ind. Eng. 2011, 61, 322–335. [Google Scholar] [CrossRef]
  8. Han, C.; Wang, X.-W.; Chen, Z. Scheduling for single agile satellite, redundant targets problem using complex networks theory. Chaos Solitons Fractals 2016, 83, 125–132. [Google Scholar]
  9. Liang, X.; Wang, H.; Xu, R.; Chen, H. Priority-based constructive algorithms for scheduling agile earth observation satellites with total priority maximization. Expert Syst. Appl. 2016, 51, 195–206. [Google Scholar]
  10. Yuan, Z.; Chen, Y.; He, R. Agile earth observing satellites mission planning using genetic algorithm based on high quality initial solutions. In Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC), Beijing, China, 6–11 July 2014; pp. 603–609. [Google Scholar]
  11. Li, Y.; Xu, M.; Wang, R. Scheduling observations of agile satellites with combined genetic algorithm. In Proceedings of the Third International Conference on Natural Computation, Haikou, China, 24–27 August 2007. [Google Scholar]
  12. Lopez, P.; Tangpattanakul, P.; Jozefowiez, N. Multi-objective optimization for selecting and scheduling observations by agile earth observing satellites. In Proceedings of the Third International Conference on Natural Computation, Espoo, Finland, 18–22 October 2012. [Google Scholar]
  13. Chen, M.; Wen, J.; Song, Y.J.; Xing, L.N.; Chen, Y.W. A population perturbation and elimination strategy based genetic algorithm for multi-satellite tt&c scheduling problem. Swarm Evol. Comput. 2021, 65, 100912. [Google Scholar]
  14. Li, J.; Jing, N.; Emmerich, M.; Li, L.; Chen, H. Preference-based evolutionary many-objective optimization for agile satellite mission planning. IEEE Access 2018, 6, 40963–40978. [Google Scholar] [CrossRef]
  15. Li, X.; Li, Z.; Chen, H. A multi-objective binary-encoding differential evolution algorithm for proactive scheduling of agile earth observation satellites. Adv. Spa Res. 2019, 63, 3258–3269. [Google Scholar] [CrossRef]
  16. He, L.; Weerdt, M.D.; Yorke-Smith, N. Time/sequence-dependent scheduling: The design and evaluation of a general purpose tabu-based adaptive large neighbourhood search algorithm. J. Intell. Manuf. 2020, 31, 1051–1078. [Google Scholar] [CrossRef]
  17. Hopfield, J.J.; Tank, D.W. “Neural” computation of decisions in optimization problems. Biol. Cybern. 1985, 52, 141–152. [Google Scholar] [CrossRef] [PubMed]
  18. Hu, H.; Zhang, X.; Yan, X.; Wang, L.; Xu, Y. Solving a new 3d bin packing problem with deep reinforcement learning method. arXiv 2017, arXiv:1708.05930. [Google Scholar]
  19. Chen, M.; Chen, Y.; Chen, Y.; Qi, W. Deep reinforcement learning for agile satellite scheduling problem. In Proceedings of the 2019 IEEE Symposium Series on Computational Intelligence (SSCI), Xiamen, China, 6–9 December 2019. [Google Scholar]
  20. Chen, M.; Chen, Y.; Du, Y.; Wei, L.; Chen, Y. Heuristic algorithms based on deep reinforcement learning for quadratic unconstrained binary optimization. Knowl.-Based Syst. 2020, 207, 106366. [Google Scholar] [CrossRef]
  21. Kool, W.; Van Hoof, H.; Welling, M. Attention, learn to solve routing problems! arXiv 2018, arXiv:1803.08475. [Google Scholar]
  22. Vinyals, O.; Fortunato, M.; Jaitly, N. Pointer networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  23. Bello, I.; Pham, H.; Le, Q.V.; Norouzi, M.; Bengio, S. Neural combinatorial optimization with reinforcement learning. arXiv 2016, arXiv:1611.09940. [Google Scholar]
  24. Nazari, M.; Oroojlooy, A.; Snyder, L.V.; Takáč, M. Reinforcement learning for solving the vehicle routing problem. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  25. Peng, G.; Song, G.; He, Y.; Yu, J.; Vansteenwegen, P. Solving the agile earth observation satellite scheduling problem with time-dependent transition times. IEEE Trans. Syst. Man Cybern. Syst. 2020, 52, 1614–1625. [Google Scholar] [CrossRef]
  26. Cho, K.; Van Merrienboer, B.; Gulcehre, C.; Ba Hdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  27. Wang, Y.; Lue, Z.; Glover, F.; Hao, J.K. Probabilistic grasp-tabu search algorithms for the ubqp problem. Comput. Oper. Res. 2013, 40, 3100–3107. [Google Scholar] [CrossRef]
  28. Glover, F.; Kochenberger, G.; Alidaee, B.; Amini, M.M. Tabu search with critical event memory: An enhanced application for binary quadratic programs. In Meta-Heuristics; Springer: Boston, MA, USA, 1999; pp. 93–109. [Google Scholar]
  29. Glover, F.; Ye, T.; Punnen, A.P.; Kochenberger, G. Integrating tabu search and vlsn search to develop enhanced algorithms: A case study using bipartite boolean quadratic programs. Eur. J. Oper. Res. 2015, 241, 697–707. [Google Scholar] [CrossRef]
  30. Palubeckis, G. Multistart tabu search strategies for the unconstrained binary quadratic optimization problem. Ann. Oper. Res. 2004, 131, 259–282. [Google Scholar] [CrossRef]
  31. Glover, F.; Lü, Z.; Hao, J.K. Diversification-driven tabu search for unconstrained binary quadratic problems. 4OR 2010, 8, 239–253. [Google Scholar] [CrossRef]
  32. Ying, Z. A decomposition-based multi-objective tabu search algorithm for tri-objective unconstrained binary quadratic programming problem. In Proceedings of the IEEE International Conference on Computational Science & Engineering, Guangzhou, China, 21–24 July 2017. [Google Scholar]
  33. Alkhamis, T.M.; Hasan, M.; Ahmed, M.A. Simulated annealing for the unconstrained quadratic pseudo-boolean function. Eur. J. Oper. Res. 1998, 108, 641–652. [Google Scholar] [CrossRef]
  34. Katayama, K.; Narihisa, H. Performance of simulated annealing-based heuristic for the unconstrained binary quadratic programming problem. Eur. J. Oper. Res. 2001, 134, 103–119. [Google Scholar] [CrossRef]
  35. Trottier, L. Off-policy actor-critic. arXiv 2012, arXiv:1205.4839. [Google Scholar]
Figure 1. Diagram of agile satellite performing observation tasks.
Figure 1. Diagram of agile satellite performing observation tasks.
Mathematics 10 03498 g001
Figure 2. Traditional heuristic algorithm flow diagram of static parameter input.
Figure 2. Traditional heuristic algorithm flow diagram of static parameter input.
Mathematics 10 03498 g002
Figure 3. DRL-HA operation process diagram.
Figure 3. DRL-HA operation process diagram.
Mathematics 10 03498 g003
Figure 4. Structure of multi-head attention.
Figure 4. Structure of multi-head attention.
Mathematics 10 03498 g004
Figure 5. The overall structure of the HCNN model.
Figure 5. The overall structure of the HCNN model.
Mathematics 10 03498 g005
Figure 6. Task observation arrangement process.
Figure 6. Task observation arrangement process.
Mathematics 10 03498 g006
Figure 7. Observation arrangement process for tasks deleted due to constraint violation.
Figure 7. Observation arrangement process for tasks deleted due to constraint violation.
Mathematics 10 03498 g007
Figure 8. Performance of HCNN at different sizes. (a) Average profits for HCNN at different sizes. (b) Total time for HCNN at different sizes.
Figure 8. Performance of HCNN at different sizes. (a) Average profits for HCNN at different sizes. (b) Total time for HCNN at different sizes.
Mathematics 10 03498 g008
Figure 9. Performance of the algorithms for scenarios with different scales. (a) Average profits of the algorithms for scenarios with different scales. (b) Total time of the algorithms for scenarios with different scales.
Figure 9. Performance of the algorithms for scenarios with different scales. (a) Average profits of the algorithms for scenarios with different scales. (b) Total time of the algorithms for scenarios with different scales.
Mathematics 10 03498 g009
Figure 10. Performance of HCNN at different sizes. (a) Average profits for the single-orbit scenarios with 50 tasks; (b) Average profits for the single-orbit scenarios with 100 tasks.
Figure 10. Performance of HCNN at different sizes. (a) Average profits for the single-orbit scenarios with 50 tasks; (b) Average profits for the single-orbit scenarios with 100 tasks.
Mathematics 10 03498 g010
Figure 11. The average profit of total tasks of the comparison algorithms under different task scales.
Figure 11. The average profit of total tasks of the comparison algorithms under different task scales.
Mathematics 10 03498 g011
Figure 12. Interactions of GRILS at different sizes. (a) Interactions of GRILS at size 50. (b) Interactions of GRILS at size 100. (c) Interactions of GRILS at size 150. (d) Interactions of GRILS at size 200.
Figure 12. Interactions of GRILS at different sizes. (a) Interactions of GRILS at size 50. (b) Interactions of GRILS at size 100. (c) Interactions of GRILS at size 150. (d) Interactions of GRILS at size 200.
Mathematics 10 03498 g012
Table 1. Attitude transition and task parameters.
Table 1. Attitude transition and task parameters.
ParametersValues
Attitude Transition Speeds v 1 = 1.5, v 2 = 2, v 3 = 2.5, v 4 = 3
Attitude Stabilization Time λ 0 = 11.66, λ 1 = 5, λ 2 = 10, λ 3 = 16, λ 4 = 22
Thresholds of Attitude Transition Angles θ 1 = 10, θ 2 = 30, θ 3 = 60, θ 4 = 90
Duration Time Required for Imagingl∼N (15, 30)
Table 2. Satellite orbit parameters.
Table 2. Satellite orbit parameters.
Satelliteaei ω Ω m
AS-017141701.7 km0.00062798.5964 deg95.5069 deg342.307 deg125.2658 deg
Table 3. Competitors.
Table 3. Competitors.
RandomWFPFPDFPWF
TS-TS-WTS-P--
SA-SA-WSA-P--
GRILS-G-WG-P-
Table 4. Main parameters of the actor network.
Table 4. Main parameters of the actor network.
Actor-HCNN
Global: Q,K,V-dim = 128, Head = 8, Layers = 3, Feed forward = 512, Dropout = 0.1
Encoder: Conv-1D( D i n p u t s i z e , Filter = 128, kernel size = 1, stride = 1)
Conv-1D( D i n p u t s i z e = 1 , Filter = 128, kernel size = 1, stride = 1)
Decoder: GRU(hidden size = 128, layers = 1, dropout = 0.2)
Table 5. Main parameters of the critic network.
Table 5. Main parameters of the critic network.
Critic-HCNN
Encoder: Conv-1D( D i n p u t s i z e , Filter = 128, kernel size = 1, stride = 1)
Feed forward layer 1 (input features = 256; output features = 20)
Feed forward layer 2 (input features = 20; output features = 20)
Output layer (input features = 20; output features = 1)
Table 6. Experimental results of scenarios for scale from 25 to 100.
Table 6. Experimental results of scenarios for scale from 25 to 100.
255075100
PT(s)PT(s)PT(s)PT(s)
H-25116.377.22160.9713.04186.8522.16201.5933.51
H-50112.827.67163.7312.8188.7822.62206.0833.74
H-75112.247.69162.8112.84187.7723.55205.3934.19
H-100111.996.5161.1412.98187.0323.67204.7434.65
H-125111.285.25161.0412.99187.8822.66205.7434.25
H-150111.265.29160.7212.69186.5322.34203.6633.96
Random85.353.68106.339.48115.6419.19121.4331.97
WF109.772.49127.566.76136.0713.67140.3522.18
PF103.653.69145.359.44169.5519.72181.5132.11
PDF101.613.74143.289.71165.6420.26180.1133.57
PWF104.393.61148.199.12171.7318.69186.0130.5
Table 7. Experimental results of scenarios with different scales for scale from 125 to 200.
Table 7. Experimental results of scenarios with different scales for scale from 125 to 200.
125150175200
PT(s)PT(s)PT(s)PT(s)
H-25212.249.23220.3467.84229.4185.19236.23100.9
H-50215.3648.82223.5667.6231.8878.81238.37100.74
H-75214.7949.78225.9468.05233.0778.97241.1998.74
H-100215.1450.83224.2869.75232.8780.27238.42104.43
H-125216.6150.44227.2268.45236.8578.4324397.41
H-150216.8149.47227.0566.98237.7977.4244.36102.11
Random122.9850.49129.8773.42131.2195.08132.99119.95
WF145.6234.75149.1250.72151.6764.14155.0180.98
PF192.8950.34200.6473.97207.8496.27213.48120.7
PDF191.1553.24198.977.75205.73102.33212.24126.45
PWF199.7747.33208.0167.68218.4686.34223.36107.52
Table 8. Experimental results for the single-orbit scenarios with 25 tasks.
Table 8. Experimental results for the single-orbit scenarios with 25 tasks.
Orbits O-1O-2O-3O-4O-5O-6O-7O-8
HCNNP112.82120.55108.54110.9797.97108.65112.497.39
T(s)5.175.195.095.175.175.265.265.15
RandomP84.0597.776.7582.6863.1284.7886.2568.78
T(s)2.82.852.782.812.362.782.832.68
WFP105.1115.51100.01103.7372.72102.35105.4986.57
T(s)1.931.342.021.951.081.441.941.95
PFP102.2112.0296.5999.3882.15101.74102.989.84
T(s)2.842.892.752.782.392.762.842.69
PDFP101.3311294.399.3283.09100.2101.3487.85
T(s)2.892.952.792.822.422.842.882.75
PWFP103.36113.0698.59100.884.02102.53103.7490.96
T(s)2.732.772.632.72.242.672.732.64
Table 9. Experimental results for the single-orbit scenarios with 100 tasks.
Table 9. Experimental results for the single-orbit scenarios with 100 tasks.
Orbits O-1O-2O-3O-4O-5O-6O-7O-8
HCNNP206.08226.75185.15195.79146.8197.33196.7172.29
T(s)32.1432.930.1733.7229.9329.6332.1332.18
RandomP119.93137.38101.25112.3987.03121.53115.5997.92
T(s)32.3635.0831.3532.5825.7232.4732.3434.33
WFP135.87173.02123.26131.1101.94161.61136.02124.81
T(s)21.8416.6320.2221.848.7114.6721.5723.84
PFP183.76211.46156.44174.77132.43186.93181.06148.41
T(s)32.1434.831.4632.7326.3932.5632.234.31
PDFP181.26209.77154.25172.28135.04183.3177.97145.37
T(s)33.4636.4732.9834.2127.7434.5733.4936.08
PWFP188.92217.46161.33179.34139.06194.55186.94152.41
T(s)30.0533.1229.6531.0516.4630.2230.5131.98
Table 10. The average total task profit results of the comparison algorithms under different task scales.
Table 10. The average total task profit results of the comparison algorithms under different task scales.
50100150200
Pstd.T(s)Pstd.T(s)Pstd.T(s)Pstd.T(s)
WF1280.000.21620.000.51780.000.81630.001.2
PF1530.000.31780.000.71870.001.12080.001.5
HCNN1660.000.82020.001.72280.002.42470.003.2
SA-W159.87.861819.11188.612.183615.162057.825423.85188.312.425977.93
SA-P163.65.121644.76197.65.594413.37214.18.235580.30220.87.705530.17
SA-HCNN178.42.201215.25205.53.692665.23237.76.483687.63251.73.164646.30
TS-W175.94.182385.94221.86.016065.45236.17.416171.87224.87.576264.22
TS-P174.35.481940.55215.15.456054.29231.18.176206.29240.96.746134.78
TS-HCNN185.52.201711.24225.61.816061.48259.32.976162.82262.43.806341.28
G-W1953.77563.68235.15.032327.79250.66.443109.30256.95.564926.20
G-P197.63.88528.63234.35.181899.61251.97.032784.32257.55.874346.50
G-HCNN199.61.85512.67237.54.651583.61253.73.112574.07262.34.493946.41
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, J.; Chen, M.; Wen, J.; He, L.; Liu, X. A Heuristic Construction Neural Network Method for the Time-Dependent Agile Earth Observation Satellite Scheduling Problem. Mathematics 2022, 10, 3498. https://0-doi-org.brum.beds.ac.uk/10.3390/math10193498

AMA Style

Chen J, Chen M, Wen J, He L, Liu X. A Heuristic Construction Neural Network Method for the Time-Dependent Agile Earth Observation Satellite Scheduling Problem. Mathematics. 2022; 10(19):3498. https://0-doi-org.brum.beds.ac.uk/10.3390/math10193498

Chicago/Turabian Style

Chen, Jiawei, Ming Chen, Jun Wen, Lei He, and Xiaolu Liu. 2022. "A Heuristic Construction Neural Network Method for the Time-Dependent Agile Earth Observation Satellite Scheduling Problem" Mathematics 10, no. 19: 3498. https://0-doi-org.brum.beds.ac.uk/10.3390/math10193498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop