Figure 1.
LSTM generic unit (
left) together with an LSTM unit with teacher forcing (
right). This figure originates from [
40].
Figure 1.
LSTM generic unit (
left) together with an LSTM unit with teacher forcing (
right). This figure originates from [
40].
Figure 2.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the sequence length in the range [1:500]. The highlighted thick lines represent the training loss for a sequence length of 1.
Figure 2.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the sequence length in the range [1:500]. The highlighted thick lines represent the training loss for a sequence length of 1.
Figure 3.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the sequence length in the range [1:500]. The highlighted thick lines represent the training loss for a sequence length of 1.
Figure 3.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the sequence length in the range [1:500]. The highlighted thick lines represent the training loss for a sequence length of 1.
Figure 4.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 4.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 5.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 5.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 6.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with mini-batch size in the [2:128] range. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 6.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with mini-batch size in the [2:128] range. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 7.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the training mini-batch sizes in the [2:128] range. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 7.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. For each architecture, the lines with the same colors represent the loss for networks trained with the training mini-batch sizes in the [2:128] range. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 8.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 8.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 9.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 9.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 10.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 10.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 11.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 11.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 12.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 12.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 13.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 13.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 14.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 14.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 15.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 15.
Illustration of the training loss for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode. The highlighted thick lines represent the training loss for a mini-batch size of 2.
Figure 16.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 16.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 17.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 17.
Illustration of the testing mean absolute error value for Experiment 1, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 18.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 18.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 19.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 19.
Illustration of the training loss for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 20.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 20.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 21.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 21.
Illustration of the testing mean absolute error value for Experiment 2, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 22.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 22.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-many prediction mode.
Figure 23.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 23.
Illustration of the mean loss value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 24.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 24.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left), VLSTM (middle), and LSTMTFC (right), using a many-to-many prediction mode.
Figure 25.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 25.
Illustration of the testing mean absolute error value for Experiment 3, LSTMTF (left) and VLSTM (right), using a many-to-one prediction mode.
Figure 26.
Illustration of the prediction performance for experiment 4. Subfigures (a,b,d–f) showcase the observed/predicted values on an entire cycle of 500 observations. Subfigure (c) showcases the observed/predicted values for LSTMTF using an M2O prediction mode for 5000 observations.
Figure 26.
Illustration of the prediction performance for experiment 4. Subfigures (a,b,d–f) showcase the observed/predicted values on an entire cycle of 500 observations. Subfigure (c) showcases the observed/predicted values for LSTMTF using an M2O prediction mode for 5000 observations.
Table 1.
The list of selected inputs and outputs for the MISO and MIMO configurations. A detailed description of the variables can be found in the original TEP paper [
42,
43].
Table 1.
The list of selected inputs and outputs for the MISO and MIMO configurations. A detailed description of the variables can be found in the original TEP paper [
42,
43].
Configuration | Output Variables | Input Variables |
---|
MISO | Product Analysis F (Stream 11) | A Feed (stream 1) |
A and C Feed (Stream 4) |
Product Separator Pressure |
Stripper Pressure |
Stripper Temperature |
Stripper Steam Flow |
Reactor Cooling Water Outlet Temperature |
Reactor feed Analysis B |
Reactor feed Analysis E |
A feed flow (Stream 1) |
Reactor Cooling Water Flow |
MIMO | Product Analysis F (Stream 11) Purge gas analysis (Stream 9) | All Continuous Process Measurements |
All Process Manipulated Variables |
Table 2.
The complete list of fixed and varying hyperparameter values for Experiments 1–3 for all the tested neural network architectures.
Table 2.
The complete list of fixed and varying hyperparameter values for Experiments 1–3 for all the tested neural network architectures.
Experiment | Hidden Layers | Sequence Input Length | Hidden Units | Learning Rate | Mini-Batch Size | Epochs | Lags |
---|
1 | 1 | [10, 500] | 16 | 0.01 | 32 | 100 | [1, 50] |
2 | 1 | 40 | 16 | 0.01 | [1, 128] | 100 | 1 |
3 | 1 | 40 | [1, 128] | [0.00001, 0.1] | 32 | 100 | 1 |
Table 3.
The training convergence time (epochs) for LSTMTF, computed with respect to the number of lags, for = 1% and P = 5.
Table 3.
The training convergence time (epochs) for LSTMTF, computed with respect to the number of lags, for = 1% and P = 5.
| Many-to-Many | Many-to-One |
---|
| Min | Avg | Max | Min | Avg | Max |
---|
1 Lag | 41 | 86 | 100 | 16 | 65 | 100 |
10 Lags | 26 | 91 | 100 | 16 | 72 | 100 |
20 Lags | 46 | 95 | 100 | 21 | 73 | 100 |
30 Lags | 46 | 95 | 100 | 21 | 73 | 100 |
40 Lags | 46 | 96 | 100 | 26 | 71 | 100 |
50 Lags | 46 | 97 | 100 | 41 | 87 | 100 |
Table 4.
The average training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the input sequence length, for = 1%, P = 5, and 1 lag for LSTMTF.
Table 4.
The average training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the input sequence length, for = 1%, P = 5, and 1 lag for LSTMTF.
| LSTMTF | VLSTM |
---|
ISL | Many-to-Many | Many-to-One | Many-to-Many | Many-to-One |
---|
[10, 100] | 70 | 83 | 32 | 38 |
[100, 200] | 73 | 69 | 36 | 51 |
[200, 300] | 86 | 82 | 44 | 57 |
[300, 400] | 93 | 82 | 66 | 74 |
[400, 500] | 90 | 86 | 80 | 79 |
Table 5.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the mini-batch size, for = 1% and P = 5.
Table 5.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the mini-batch size, for = 1% and P = 5.
| LSTMTF | VLSTM |
---|
MBS | Many-to-Many | Many-to-One | Many-to-Many | Many-to-One |
---|
2 | 100 | 46 | 12 | 66 |
8 | 51 | 31 | 16 | 41 |
16 | 86 | 41 | 41 | 32 |
32 | 100 | 51 | 46 | 56 |
64 | 86 | 56 | 51 | 56 |
128 | 100 | 56 | 51 | 57 |
Table 6.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the number of hidden units, for = 1% and P = 5.
Table 6.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the number of hidden units, for = 1% and P = 5.
| LSTMTF | VLSTM |
---|
HU | Many-to-Many | Many-to-One | Many-to-Many | Many-to-One |
---|
2 | 81 | 38 | 30 | 38 |
8 | 73 | 40 | 25 | 44 |
16 | 76 | 43 | 24 | 43 |
32 | 79 | 47 | 22 | 44 |
64 | 77 | 44 | 22 | 42 |
128 | 58 | 47 | 23 | 47 |
Table 7.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the learning rate, for = 1% and P = 5.
Table 7.
The training convergence time (epochs) for LSTMTF and VLSTM, computed with respect to the learning rate, for = 1% and P = 5.
| LSTMTF | VLSTM |
---|
LR | Many-to-Many | Many-to-One | Many-to-Many | Many-to-One |
---|
[0.0001, 0.02] | 71 | 46 | 23 | 45 |
[0.02, 0.04] | 71 | 44 | 23 | 44 |
[0.04, 0.06] | 71 | 44 | 23 | 44 |
[0.06, 0.08] | 71 | 45 | 23 | 45 |
[0.08, 0.1] | 71 | 44 | 23 | 44 |
Table 8.
The actual training and testing times for 50,000 data points, measured in milliseconds, for LSTMTF and VLSTM.
Table 8.
The actual training and testing times for 50,000 data points, measured in milliseconds, for LSTMTF and VLSTM.
| LSTMTF | VLSTM |
---|
Time [ms] | Many-to-Many | Many-to-One | Many-to-Many | Many-to-One |
---|
MISO Training | 1820 | 1831 | 1810 | 1870 |
MISO Testing | 46.55 | 49.70 | 47.13 | 45.72 |
MIMO Training | 2209 | 2294 | 2208 | 2236 |
MIMO Testing | 57.27 | 63.08 | 60.74 | 61.71 |