Next Article in Journal
Attention-Based STL-BiLSTM Network to Forecast Tourist Arrival
Next Article in Special Issue
Biomass Combustion Control in Small and Medium-Scale Boilers Based on Low Cost Sensing the Trend of Carbon Monoxide Emissions
Previous Article in Journal
Estimation of Gene Regulatory Networks from Cancer Transcriptomics Data
Previous Article in Special Issue
A Comparison of Ethanol, Methanol, and Butanol Blending with Gasoline and Its Effect on Engine Performance and Emissions Using Engine Simulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Selection and Uncertainty Analysis for Bubbling Fluidized Bed Oxy-Fuel Combustion Data

by
Katerina Marzova
1,† and
Ivo Bukovsky
1,2,*,†
1
Department of Mechanics, Biomechanics and Mechatronics, Czech Technical University in Prague, 166 07 Prague, Czech Republic
2
Research Centre for Low-Carbon Energy Technologies, Czech Technical University in Prague, 166 07 Prague, Czech Republic
*
Author to whom correspondence should be addressed.
Current address: Faculty of Mechanical Engineering, Czech Technical University in Prague, Technicka 4, 166 07 Prague, Czech Republic.
Submission received: 26 August 2021 / Revised: 20 September 2021 / Accepted: 21 September 2021 / Published: 30 September 2021

Abstract

:
This paper presents a novel feature extraction and validation technique for data-driven prediction of oxy-fuel combustion emissions in a bubbling fluidized bed experimental facility. The experimental data were analyzed and preprocessed to minimize the size of the data set while preserving patterns and variance and to find an optimal configuration of the feature vector. The Boruta Feature Selection Algorithm (BFSA) finds feature vector’s configuration and the Multiscale False Neighbours Analysis (MSFNA) is newly extended and proposed to validate the BFSA’s design for emission prediction to assure minimal uncertainty in mapping between feature vectors and corresponding outputs. The finding is that the standalone BFSA does not reflect various sampling period setups that appeared significantly influencing the false neighborhood in the design of feature vectors for possible emission prediction, and MSFNA resolves that.

1. Introduction

In general, oxy-fuel combustion is one of the ‘carbon capture’ technologies that aim to reduce CO x emissions (mainly) from coal power plants [1] and also within other fuel combustion systems such as biomass [2] that is still prospective for utilization and research. The fuel is oxidized in almost pure oxygen, unlike the air in the traditional air combustion, and therefore the flue gas consists primarily of carbon dioxide and water vapor [3,4].
Combustion processes are complex, nonlinear, and their measurements generate complex data that requires proper analysis for further design of prediction and control systems.
Computational intelligence approaches with machine learning have been attractive to study for combustion processes in the last decades, e.g., [5], where multilayer feed-forward neural network with error back-propagation learning was used for approximation of measured CO/lambda biomass combustion dependence and that shows a significant variance in data that can be seen as a kind of uncertainty, and the neural network is used to extract the prevailing dependence in data; further example of filtration with membership function design can be found in [6]. Further computational intelligence techniques based on immune systems and applied to biomass combustion and that highlights the nonlinearity and complexity of the process can be found in [7]. Predictions of NO x emissions from biomass-fired combustion process using experimentally established dataset of flame radical images and deep learning was presented in [8], where also Morphological Component Analysis and region-of-interest extraction were used to reduce the dataset size and improve the clarity and unambiguity of the samples.
As equivalent inputs, digital flame images were used in [9], where the successful operation condition recognition was achieved by building a combined Principal Component Analysis and Random Weight Network model and estimating optimal model parameters using cross-validation (PCA-RWN).
One of the traditional approaches to CO x emission reduction is presented in [10], where the emissions, caused by temporary fluctuations in fuel feed of the wood pellets, were compensated by precise control of primary and secondary airflow (air staging).
Apart from emission prediction or process control itself, feature selection techniques represent a significant part of research with a similar topic. A feature selection technique evaluating the features by comparing the MAE (Mean Absolute Error) between the real and predicted value of different subsets of primary and secondary variables from the original dataset can be found in [11]. Selected features were used to build a model of a brown coal-fired boiler to predict fresh steam properties for a suitable combination of input parameters. An analogous approach was also used for short-term forecasting of CO 2 emission intensity [12], where a forward selection algorithm was applied after removing highly correlated variables based on Pearson’s correlation coefficient and LASSO regression. Apart from Pearson’s correlation coefficient, a statistic popular, especially for large data, is MIC (Maximal Information Coefficient [13]) that can capture both functional and nonfunctional relationships and gives similar scores to equally noisy relationships. MIC was used by [14] for NO x emission prediction with Long Short-Term Memory (LSTM).
With prospects for future research and design of a machine-learning-based prediction system of emissions from the oxyfuel system, we propose a technique that designs a feature vector based on random forests in combination with further validation via uncertainty analysis by Multiscale False Neighbour analysis (MSFNA) [15]. In this study, a feature vector is proposed by Boruta Feature Selection Algorithm (BFSA) [16], which combines two powerful tools for feature selection: decision trees and z-score.
Mathematical notation of the variables is the following: small letters, such as “x”, are for scalars, bold “ x ” for vectors and bold capital “ X ” for matrices. Lower indexes, such as “ x i ” usually indicate the position of the element in the vector x (except for u and y), i in “ x i ” denotes ith row of a matrix, u stands for a process (control) input variable, y denotes process output variables, and u i and y j denotes ith process input variable and jth process variable respectively (see Table 1). Further in text, “ x ” denotes feature vector; notice terms “feature vector”, “state vector”, and ”input vector” have the same meaning and are used interchangeably in this paper. Capital “N” denotes the length of data (the count of all processed data samples), and ”n” is generally used to denote other quantities such as “ n x ” for the length of the vector x , or n c for the number of applied (principal) components. Exclusively, r x and r y denote radii vectors in state space of state vectors x and output variable y, respectively. At last, u i [ k n u i : k 1 ] denotes the vector of last measured step delays of variable u i (the vector length is n u i ).

2. Methods

2.1. Feature Selection

Feature selection is the very fundamental process for classification and prediction. With the decreasing price of data acquisition systems, the amount of measured variables is increasing and so is the need for feature selection and data reduction. Feature selection methods are traditionally divided into three groups as follows [17,18]:
(1)
Filter methods use various techniques (correlation, Mutual Information, chi-squared test, etc.) to score the variables and compare the scores against the threshold to filter out features.
(2)
Wrapper methods apply different machine learning algorithms to get predictor performance as an evaluation of variables. The main disadvantages are the high chances of overfitting and longer computation time.
(3)
Embedded methods aim to reduce computation time by integrating feature selection as part of a training process. Typical representatives are LASSO and Elastic Net.
In studies with similar topics regarding emission prediction, there are three most common approaches:
  • using Principal Component Analysis to reduce the size of an input data set [9,19],
  • searching for statistical dependencies between features [20,21],
  • filtering out the outliers and irrelevant data points [11,22,23].
The current state of art of feature selection algorithms is focused on static mapping functions [24]. In this paper, we apply the BFSA algorithm to reduce the unnecessary length of a state vector of a discrete-time dynamic system with constant sampling, so it also identifies the optimal configuration of step-delayed variables. Furthermore, the feature selection, performed by the BFSA, is proposed to verify the results via the MSFNA, where MSFNA appears useful for the sampling period validation. The use of BFSA and the MSFNA is recalled in the following subsections.

2.1.1. Boruta Feature Selection Algorithm (BFSA)

BFSA is a wrapper feature selection method presented in [16]. The algorithm searches all important features in the existing data set using a random forest algorithm to identify the importance of features.
The main idea behind the BFSA algorithm is as follows: if there exists a feature, that scores lower than the best of the shadow features, the feature is not relevant for the system (unsuccessful performance).
First, BFSA creates shadow features from the existing ones by randomly shuffling the original values of a feature matrix X along the column, as follows
X = x i = 1 x i = 2 x i = N ,
X s h a d o w = x j = 1 x j = 2 x j = N where x i x j for some of i = j where i , j { 1 , 2 , . . . , N } ,
where x is a state (feature, input) vector composed of actual input variables, their step-delayed values, and step-delayed output variables whose configuration is to be found by BFSA and validated by MSFNA. A random forest classifier is used as a predictor that creates a model extended by shadow features, i.e., by additional and irrelevant ones. The random forest model is thereupon used for feature performance evaluation. For each feature, z-scores are computed as the average accuracy loss of the prediction divided by its standard deviation [16]. Feature importance is estimated by comparing its z-score against the smallest z-score value of the shadow features. This value is used as a threshold because the shadow feature importance is nonzero only due to random fluctuations and is not beneficial for the system at all.
The process is repeated for a predefined number of trials and the number of successful and unsuccessful performances of each feature is computed.
The total number of successful or unsuccessful performances, that each feature must reach, is driven by the binomial distribution as follows
f ( s , ϵ , p ) = ϵ s p s ( 1 p ) ϵ s ,
where ϵ is the number of trials, p 0 , 1 is the success probability for each trial and s is the total number of successful trials, so it defines confidence intervals as in the example in Figure 1. The number of successful performances of a feature, that is to be labeled as important for the data, is given only by the number of trials used. Successful features then create a new data set.

2.1.2. Multiscale False Neighbours Analysis (MSFNA)

The MSFNA is a technique used for the uncertainty evaluation of the state vector’s configuration as it evaluates the determinism for general input-output mapping f ( x ) = y , i.e., the mapping f ( ) between feature vectors x and corresponding outputs y [15]. Unlike common false neighbor analysis, which requires precise knowledge about the neighborhood radius, see [25,26], MSFNA uses a vector of several radii to scale the neighborhood condition. If x is the state vector of the system, then system output y is unambiguously determined by the input (state vector) x if the same input x results in different outputs y, the state vector is not complete and other features must be included, these states are then referred to as false neighbors (FN). Two states x 1 , x 2 are FN if they meet the condition
i f x i x j < r x AND y i y j > r y F N ( i , j ) = 1 ; else F N ( i , j ) = 0 ,
where r x , r y are absolute radii that define the condition of state similarity. Because optimal r x and r y is practically unknown, MSFNA utilizes the whole range of radii via heuristically designed vectors r x = [ r x 1 , r x 2 , , r x α ] , r y = [ r y 1 , r y 2 , , r y β ] assuming that the optimal radii lies within the ranges of radii of vectors r x , r y . The Areal Cumulative False Neighbors A C F N is computed according to
A C F N = i = 1 N j = i + 1 N F N ( i , j ) ,
where A C F N is the sum of all false neighbors in the dataset that evaluates the uniqueness of the state vector configuration for each data point. The higher A C F N gets, the less determinism, i.e., more false neighbors, is in the mapping of the designed feature vector configuration and prediction outputs.
To compare datasets with a different number of samples, we introduce the Relative ACFN ( R A C F N ) that eliminates the dependence of A C F N on the number of samples in the dataset by normalization via the total count of all neighbor pairs as follows
R A C F N = i = 1 N j = i + 1 N F N ( i , j ) 2 N 2 2 N 0 , 1 .
Apart from the number of samples, a dimension of the input matrix (state vector) is also a source of deceptive information arising from comparing euclidean distances of variously long state vectors against the same radii vectors r x and r y .
While (6) resolves the issue of various lenght of data, it still suffers from various lengths of feature vectors n x that BFSA can principally find for various datasets. Thus, we propose to resolve the non-unique feature-vector-length problem by compressing (column-wise) the matrix of state vectors X as in (1) using Principal Component Analysis (PCA) into a customized number of components n c , i.e., the matrix X is compressed to a constant number of columns (new features), and the MSFNA is then evaluated for state vectors of the same length n c . The PCA compression into n c features is applied as follows
X P C A = v 1 v 2 v n c X ,
where v 1 , v 2 , , v n c , are selected eigenvectors of covariance matrix C X = X X T .
It can be noted, that the PCA compression prior R A C F N not only standardizes its results by unifying the feature vector length for various feature setups, but it also naturally increases the robustness of the MSFNA results as the PCA compression further removes unique features because n c of the most significant components are used for compression.

2.2. Data and Preprocessing

The experimental device is a 30 kW laboratory-scale oxy-fuel BFB combustor, the scheme is in the Figure 2. Sunflower pellets were used as biomass fuel.
The variables used for the analysis are listed in Table 1. As input variables u i are denoted features, that can be changed or indirectly influenced by the facility set-up. Output variables y i consist of the volumetric fractions of the flue gas components.
The experiment was realized by measuring four states of the oxy-fuel mode defined by secondary and primary oxygen flow ratios V O 2 , s e c V O 2 , p r i m . In each mode, there were 900 up to 1100 samples measured, as shown in Table 2. All oxy-fuel modes were measured at the same environmental conditions in one day.
Collected data were first checked for missing values and relevant data points were removed and replaced with the previous value. Data were resampled with four different sampling constants to compare the results: Δ T 1 = 5 s, Δ T 2 = 10 s, Δ T 3 = 15 s and Δ T 4 = 25 s. The new point is calculated as a mean value of previous points, which behaves as a simple smoothing filter. In the next step, the data were normalized using the Min-Max scaler as shown in (8) and (9).
X M i n M a x ( i ) = X S ( i ) ( F R m a x F R m i n ) + F R m i n
X S ( i ) = X ( i ) min ( X ) max ( X ) min ( X )
F R m a x and F R m i n is the upper, respectively lower interval limit of the desired feature range; we used 0 , 1 . The last step is to design the state vector and matrix of states. The matrix of all state vectors X is defined as follows
X = x T ( k = 1 ) x T ( k = 2 ) x T ( k = N ) ,
where the full X before applying BFSNA for selecting optimal features is defined as follows.
x ( k ) = [ u 1 [ k n u 1 : k 1 ] , u 2 [ k n u 2 : k 1 ] , y 1 [ k n y 1 : k 1 ] ) , y 2 [ k n y 2 : k 1 ] ) , ] ,
where the maximum number of step delays n u i and n y j are initialized heuristically, the meaning of input and output process variables is defined in Table 1. The snapshot of measured and min-max scored process variables u i and y j is shown in Figure 3.
For further details of the real device and the experimental setup, please kindly see [3,4,27].

2.3. Algorithm Complexity

The computational complexity of the BFSA is O ( n x · N ) , where n x is the number of features and N is the number of samples [16]. The computational complexity of the Multiscale False neighbor analysis is O ( n r · N ) , where n r = α is the length of MSFNA absolute radii vector, see the text bellow Equation (4).

3. Experimental Analysis and Results

In this section, experimental analysis on artificial data and then on real data is shown for the algorithms described in Section 2.

3.1. Artificial Data

The first goal was to test the proposed algorithms on an artificial dataset of a simple MISO (multiple-input, single-output) system given by equations
X [ k , : ] = [ 2 k + s i n ( k ) , ( 0.5 · k ) 2 + k 2 1 , k 3 + k 2 ] , y ( k ) = k = 1 3 X [ k , i ] , where k = 1 , 2 , . . . N = 250 .
Two additional variables X [ : , 3 ] and X [ : , 4 ] were added to test the BFSA, these columns are a combination of a sine, resp. cosine curve and white noise. The artificial data from (12) are plotted in Figure 4, where X [ : , 3 ] and X [ : , 4 ] are added to demonstrate that BFSA eliminates them and MSFNA detects more false neighbors via higher R A C F N .
BFSA was applied on artificial data and features X [ : , 0 ] , X [ : , 1 ] and X [ : , 2 ] were correctly selected. In the next step, 15 combinations of possible input features were prepared as a combinations of 3 and 4 out of five possible features; 5 3 + 5 4 = 15 , compressed with PCA and R A C F N s were calculated according to (6). The results are in Figure 5. The lowest value of R A C F N was determined for the input state vector proposed by BFSA R A C F N B F S A = 0.0178 , for the rest of the input vectors the R A C F N values lie in the interval 0.0309 , 0.036 . The input combination proposed by BFSA was proven to have the lowest count of R A C F N following the expectations. The result is in the Figure 5.

3.2. Experimental Data

For each output variable, a state vector x was designed. The maximum number of history samples of each feature was set to 15, as in Equation (11). Results for sampling periods Δ T = [ 5 , 10 , 15 , 25 ] s for each output variable are shown in Table 3, Table 4, Table 5 and Table 6; one can notice, that variables u 1 = ψ f g , u 2 = t r u n and u 3 = t d were not selected by BFSA as input features for any output variable y i . Primary air flow u 4 = V a i r , p r i m was used rather sparsely. The total number of state vector features is listed in the last row of Table 3, Table 4, Table 5 and Table 6.
The Multiscale False Neighbours Analysis was performed with the same configuration for all samplings Δ T = [ 5 , 10 , 15 , 25 ] s. The step delays n u i and n y j were set as [ 5 , 10 , 15 ] , resulting to 3 state vector configurations with 3 different numbers of features; n x = [ 65 , 130 , 195 ] . The fourth examined state vector is proposed by BFSA. Results are shown in Figure 6, Figure 7, Figure 8 and Figure 9. The R A C F N was chosen as a comparative quantity due to the ability to compare the examined state vectors’ data variance and graphically express the results.

4. Discussion

The major outcome of the feature selection by BFSA and its newly proposed validation of MSFNA is demonstrated via Figure 6, Figure 7, Figure 8 and Figure 9, where the different sampling periods ( Δ T = [ 5 , 10 , 15 , 25 ] s) results in various levels of uncertainty ( R A C F N ) between by-BFSA-designed feature vectors and process outputs. In Figure 6 and Figure 7, where the sampling is chosen faster (than in Figure 8 and Figure 9), the BFSA proposes such feature vector configuration that displays more uncertainty via R A C F N than other merely heuristic manual feature selections (Figure 6 and Figure 7). For longer sampling period (Figure 8 and Figure 9), the by-BSFNA-designed feature vectors already displays lowest uncertainty via R A C F N as well. Recall, that all R A C F N for all various feature vector configurations were evaluated by PCA compression of feature vectors to the equal length n c and that even naturally increases the robustness of the method as already discussed in Section 2.1.
Primary airflow V a i r , p r i m , recirculation flow V r e c , and secondary air velocity v a i r , s e c were selected by BFSA less often and differently for each sampling frequency that indicates their lower importance for emission prediction.
BFSA results, further validated by MSFNA, supports the assumption that the flue gas composition is strongly dependent on previous values of process output variables y i , and this corresponds to the high complexity of the combustion dynamics where the measured values provide us only with a minimum necessary information about the actual process, so the emission prediction is still a challenge.
With a specific sampling frequency, a shape similarity can be observed across all output variables. It is the result of the strong sampling frequency dependence of the dynamic systems.
A minimum R A C F N was reached for the state vector designed by BFSA for longer sampling periods Δ T = 15 s and Δ T = 25 s. This finding corresponds to the assumptions of slower dynamical behavior of important phenomena in the real process for which we intend to predict the emissions (but in the following research as it would exceed the scope of this paper). It can be observed, that the uncertainty of the by-BFSA-designed features became improved for longer sampling of Δ T = 15 s (Figure 8), and it did not improve for longer samplings further (Figure 8 vs. Figure 9). Thus, the combination of BFSA and MFSNA not only selects features but also suggests a proper shortest sampling (for which R A C F N became the smallest out of the other non-BFSA feature designs).

5. Conclusions

The feature selection algorithm (BFSA) was studied for mapping between available measurements (features) and emissions (outputs) of bubbling fluidized bed oxy-fuel combustion data. Newly, the original uncertainty analysis (MFSNA) was extended (PCA & RACFN) and proposed to validate the BFSA in terms of uncertainty via false neighbors. The proposed technique is promising for the automatic configuration of feature vectors from measured data (for future prediction systems with machine learning). Furthermore, the combination of conventional BFSA with MSFNA appeared capable to clearly validate and propose a proper sampling period that is computationally difficult with standalone BFSA otherwise.

Author Contributions

Conceptualization, I.B.; methodology, I.B. and K.M.; software, K.M.; validation, K.M.; formal analysis, I.B. and K.M.; investigation, K.M.; writing—original draft preparation, K.M.; writing—editing and finalization, I.B. and K.M.; visualization, K.M.; supervision, I.B. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the Ministry of Education, Youth and Sports of the Czech Republic under OP RDE grant number CZ.02.1.01/0.0/0.0/16_019/0000753 “Research Centre for Low-Carbon Energy Technologies”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available.

Acknowledgments

Authors would like to thank Ondrej Budik for technical support, Jan Hrdlicka, Matej Vodicka, Pavel Skopec for valuable consultations about the real experiment facility, and special thanks to Honza for MDPI grand appraisal.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buhre, B.; Elliott, L.; Sheng, C.; Gupta, R.; Wall, T. Oxy-fuel combustion technology for coal-fired power generation. Prog. Energy Combust. Sci. 2005, 31, 283–307. [Google Scholar] [CrossRef]
  2. Toftegaard, M.B.; Brix, J.; Jensen, P.A.; Glarborg, P.; Jensen, A.D. Oxy-fuel combustion of solid fuels. Prog. Energy Combust. Sci. 2010, 36, 581–625. [Google Scholar] [CrossRef]
  3. Vodička, M.; Haugen, N.E.; Gruber, A.; Hrdlička, J. NOX formation in oxy-fuel combustion of lignite in a bubbling fluidized bed—Modelling and experimental verification. Int. J. Greenh. Gas Control 2018, 76, 208–214. [Google Scholar] [CrossRef]
  4. Vodička, M.; Hrdlička, J.; Skopec, P. Experimental study of the NOX reduction through the staged oxygen supply in the oxy-fuel combustion in a 30 kWth bubbling fluidized bed. Fuel 2021, 286, 119343. [Google Scholar] [CrossRef]
  5. Pitel, J.; Mižák, J. Computational intelligence and low cost sensors in biomass combustion process. In Proceedings of the 2013 IEEE Symposium on Computational Intelligence in Control and Automation (CICA), Singapore, 16–19 April 2013; pp. 181–184. [Google Scholar]
  6. Mižáková, J.; Pitel’, J.; Hošovský, A.; Kolarčík, M.; Ratnayake, M. Using Special Filter with Membership Function in Biomass Combustion Process Control. Appl. Sci. 2018, 8, 1279. [Google Scholar] [CrossRef] [Green Version]
  7. Tóthová, M.; Dubják, J. Using computational intelligence in biomass combustion control in medium-scale boilers. In Proceedings of the 2016 IEEE 14th International Symposium on Applied Machine Intelligence and Informatics (SAMI), Herl’any, Slovakia, 21–23 January 2016; pp. 81–85. [Google Scholar] [CrossRef]
  8. Li, N.; Lu, G.; Li, X.; Yan, Y. Prediction of NOx Emissions from a Biomass Fired Combustion Process Based on Flame Radical Imaging and Deep Learning Techniques. Combust. Sci. Technol. 2016, 188, 233–246. [Google Scholar] [CrossRef] [Green Version]
  9. Bai, X.; Lu, G.; Hossain, M.M.; Szuhánszki, J.; Daood, S.S.; Nimmo, W.; Yan, Y.; Pourkashanian, M. Multi-mode combustion process monitoring on a pulverised fuel combustion test facility based on flame imaging and random weight network techniques. Fuel 2017, 202, 656–664. [Google Scholar] [CrossRef] [Green Version]
  10. Korpela, T.M.; Björkqvist, T.K.; Lautala, P.A. Control Strategy for Small-Scale Wood Chip Combustion. IFAC Proc. Vol. 2009, 42, 119–124. [Google Scholar] [CrossRef]
  11. Smrekar, J.; Assadi, M.; Fast, M.; Kuštrin, I.; De, S. Development of artificial neural network model for a coal-fired boiler using real plant data. Energy 2009, 34, 144–152. [Google Scholar] [CrossRef]
  12. Leerbeck, K.; Bacher, P.; Junker, R.G.; Goranović, G.; Corradi, O.; Ebrahimy, R.; Tveit, A.; Madsen, H. Short-term forecasting of CO2 emission intensity in power grids by machine learning. Appl. Energy 2020, 277, 115527. [Google Scholar] [CrossRef]
  13. Reshef, D.N.; Reshef, Y.A.; Finucane, H.K.; Grossman, S.R.; McVean, G.; Turnbaugh, P.J.; Lander, E.S.; Mitzenmacher, M.; Sabeti, P.C. Detecting novel associations in large data sets. Science 2011, 334, 1518–1524. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Xie, P.; Gao, M.; Zhang, H.; Niu, Y.; Wang, X. Dynamic modeling for NOx emission sequence prediction of SCR system outlet based on sequence to sequence long short-term memory network. Energy 2020, 190, 116482. [Google Scholar] [CrossRef]
  15. Bukovsky, I.; Kinsner, W.; Maly, V.; Krehlik, K. Multiscale Analysis of False Neighbors for state space reconstruction of complicated systems. In Proceedings of the 2011 IEEE Workshop on Merging Fields of Computational Intelligence and Sensor Technology, Paris, France, 11–15 April 2011; pp. 65–72. [Google Scholar] [CrossRef]
  16. Kursa, M.B.; Rudnicki, W.R. Feature selection with the Boruta package. J. Stat. Softw. 2010, 36, 1–13. [Google Scholar] [CrossRef] [Green Version]
  17. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  18. Kumar, V.; Minz, S. Feature selection: A literature review. SmartCR 2014, 4, 211–229. [Google Scholar] [CrossRef]
  19. Yang, G.; Wang, Y.; Li, X. Prediction of the NOx emissions from thermal power plant using long-short term memory neural network. Energy 2020, 192, 116597. [Google Scholar] [CrossRef]
  20. Cuccu, G.; Danafar, S.; Cudré-Mauroux, P.; Gassner, M.; Bernero, S.; Kryszczuk, K. A data-driven approach to predict NOx-emissions of gas turbines. In Proceedings of the 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 11–14 December 2017; pp. 1283–1288. [Google Scholar] [CrossRef]
  21. Adams, D.; Oh, D.H.; Kim, D.W.; Lee, C.H.; Oh, M. Prediction of SOx–NOx emission from a coal-fired CFB power plant with machine learning: Plant data learned by deep neural network and least square support vector machine. J. Clean. Prod. 2020, 270, 122310. [Google Scholar] [CrossRef]
  22. Shi, Y.; Zhong, W.; Chen, X.; Yu, A.; Li, J. Combustion optimization of ultra supercritical boiler based on artificial intelligence. Energy 2019, 170, 804–817. [Google Scholar] [CrossRef]
  23. Tuttle, J.F.; Blackburn, L.D.; Powell, K.M. On-line classification of coal combustion quality using nonlinear SVM for improved neural network NOx emission rate prediction. Comput. Chem. Eng. 2020, 141, 106990. [Google Scholar] [CrossRef]
  24. Kumar, S.S.; Shaikh, T. Empirical Evaluation of the Performance of Feature Selection Approaches on Random Forest. In Proceedings of the 2017 International Conference on Computer and Applications (ICCA), Doha, United Arab Emirates, 6–7 September 2017; pp. 227–231. [Google Scholar] [CrossRef]
  25. Abarbanel, H.D.I.; Kennel, M.B. Local false nearest neighbors and dynamical dimensions from observed chaotic data. Phys. Rev. E 1993, 47, 3057–3068. [Google Scholar] [CrossRef] [PubMed]
  26. Kennel, M.B.; Abarbanel, H.D.I. False neighbors and false strands: A reliable minimum embedding dimension algorithm. Phys. Rev. E 2002, 66, 026209. [Google Scholar] [CrossRef]
  27. Hrdlicka, J.; Skopec, P.; Opatril, J.; Dlouhy, T. Oxyfuel Combustion in a Bubbling Fluidized Bed Combustor. Energy Procedia 2016, 86, 116–123. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Example of a decision-making process for 20 trials according to Equation (3). Using 20 trials, the feature can be labeled as important if it scores higher than shadow features at least 16 times.
Figure 1. Example of a decision-making process for 20 trials according to Equation (3). Using 20 trials, the feature can be labeled as important if it scores higher than shadow features at least 16 times.
Processes 09 01757 g001
Figure 2. Scheme of the 30 kWth BFB experimental, (1) fluidized bed region, (2) distributor of the fluidizing gas, (3) gas burner mount, (4) fluidized bed spillway, (5) fuel feeder, (6) cyclone separator, (7) flue gas fan, (8) flue gas vent, (9) and (10) water coolers, (11) condensate drain, (12) primary fan, (13) air-suck pipe, (14) vessels with oxygen, (sketch based on [3]).
Figure 2. Scheme of the 30 kWth BFB experimental, (1) fluidized bed region, (2) distributor of the fluidizing gas, (3) gas burner mount, (4) fluidized bed spillway, (5) fuel feeder, (6) cyclone separator, (7) flue gas fan, (8) flue gas vent, (9) and (10) water coolers, (11) condensate drain, (12) primary fan, (13) air-suck pipe, (14) vessels with oxygen, (sketch based on [3]).
Processes 09 01757 g002
Figure 3. Snapshot of normalized, preprocessed data from one of the real experiments, u i are process input variables and y i are process output variables described in Table 1.
Figure 3. Snapshot of normalized, preprocessed data from one of the real experiments, u i are process input variables and y i are process output variables described in Table 1.
Processes 09 01757 g003
Figure 4. Artificial data for testing, x 0 , x 1 and x 2 together create an output y, x 3 and x 4 are added to the system in order to test the feature selection and wrong feature elimination via BFSA and its validation via MSFNA.
Figure 4. Artificial data for testing, x 0 , x 1 and x 2 together create an output y, x 3 and x 4 are added to the system in order to test the feature selection and wrong feature elimination via BFSA and its validation via MSFNA.
Processes 09 01757 g004
Figure 5. MSFNA on artificial data, the orange column corresponds with the input vector designed by BFSA, the rest of input combinations (blue) have a significantly higher number of FN, the results are consistent with the hypothesis.
Figure 5. MSFNA on artificial data, the orange column corresponds with the input vector designed by BFSA, the rest of input combinations (blue) have a significantly higher number of FN, the results are consistent with the hypothesis.
Processes 09 01757 g005
Figure 6. Relative Areal Cumulative False Neighbors for sampling periods Δ T = 5 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Figure 6. Relative Areal Cumulative False Neighbors for sampling periods Δ T = 5 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Processes 09 01757 g006
Figure 7. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 10 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Figure 7. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 10 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Processes 09 01757 g007
Figure 8. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 15 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Figure 8. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 15 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Processes 09 01757 g008
Figure 9. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 25 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Figure 9. Relative Areal Cumulative False Neighbors R A C F N for sampling periods Δ T = 25 s. Orange columns are state vectors proposed by BFSA and blue columns are state vectors created according to (11) with n u i = n y j = [ 5 , 10 , 15 ] . Axis x represents the number of features n x for specified y i .
Processes 09 01757 g009
Table 1. Input and output variables used in the article were measured on the laboratory device, see Section 2.2. Notice, the previous values of the output variables y i ( i = 1 5 ) can also be in the state vector.
Table 1. Input and output variables used in the article were measured on the laboratory device, see Section 2.2. Notice, the previous values of the output variables y i ( i = 1 5 ) can also be in the state vector.
Input Variables u i iDescriptionOutput Variables y i iDescription
ψ f g [ % ] 1volumetric fraction of flue gas ψ CO [ppm]1volumetric fracion of CO
t r u n [s]2conveyor run ψ SO 2 [ppm]2volumetric fracion of SO 2
t d [s]3conveyor delay ψ NO x [ppm]3volumetric fracion of NO x
V a i r , p r i m [Nm 3 h 1 ]4primary air flow ψ O 2 [ % ] 4volumetric fracion of O 2
V r e c [Nm 3 h 1 ]5recirculation flow ψ CO 2 [ % ] 5volumetric fracion of CO 2
v a i r , s e c [ms 1 ]6secondary air velocity
V O 2 , p r i m [Nm 3 h 1 ]7primary oxygen flow
V O 2 , s e c [Nm 3 h 1 ]8secondary oxygen flow
Table 2. Oxy-fuel modes measured during the experiment are defined by secondary and primary oxygen flow ratios.
Table 2. Oxy-fuel modes measured during the experiment are defined by secondary and primary oxygen flow ratios.
mode 11109 samples with sampling period Δ T s = 2  s V O 2 , s e c V O 2 , p r i m 0
mode 2930 samples with sampling period Δ T s = 2  s V O 2 , s e c V O 2 , p r i m 0.35
mode 3900 samples with sampling period Δ T s = 2  s V O 2 , s e c V O 2 , p r i m 0.57
mode 4900 samples with sampling period Δ T s = 2  s V O 2 , s e c V O 2 , p r i m 0.91
Table 3. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 5  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 8 is not chosen for y 5 .
Table 3. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 5  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 8 is not chosen for y 5 .
Variables y 1 = ψ C O y 2 = ψ S O 2 y 3 = ψ NO x y 4 = ψ O 2 y 5 = ψ C O 2
Variables used for the input vector u 1 = ψ f g -----
u 2 = t r u n -----
u 3 = t d -----
u 4 = V a i r , p r i m 47, 58, 2, 112, 10, 9, 8,
6, 5, 4
6
u 5 = V r e c 15, 14, …, 115, 12, 11, 9,
8, 5, 1
15, 14, …, 115, 14, 13, 11,
10, …, 1
15, 14, 12, 9,
7, 5, 3
u 6 = v a i r , s e c 15, 14, …, 115, 13, 9,
8, 7, 3
15, 14, …, 115, 14, …, 115, 12, 9,
8, 5, 4
u 7 = V O 2 , p r i m 15, 14, 13, 11,
9, 8, …, 1
15, 7, 6, 5,
3, 2, 1
15, 14, …, 115, 14, …, 115, 4, 3, 1
u 8 = V O 2 , s e c 15, 13, 12, 11,
9, 8, 5,
3, 2, 1
10, 615, 14, …, 115, 14, …, 1-
y 1 = ψ CO 15, 14, 13,
10, 9, 7,
5, …, 1
15, 14, 12, …, 115, 14, …, 115, 14, …, 115, 12, 10,
9, 7, 1
y 2 = ψ SO 2 15, 14, …, 115, 14, …, 115, 14, …, 115, 14, …, 114, 1
y 3 = ψ NO x 15, 14, …, 115, 14, …, 115, 14, …, 115, 14, …, 19, 8, 3, 2, 1
y 4 = ψ O 2 15, 14, …, 18, 5, 4,
3, 2, 1
15, 14, …, 115, 14, …, 115, 13, 11,
8, 7, …, 1
y 5 = ψ CO 2 15, 13, 12, 11,
7, 6, 5, …, 1
15, 13, 12, 10,
9, 8, 7, 5,
3, 2, 1
15, 14, …, 115, 14, …, 115, 14, 12, 11,
9, 7, 6, …, 1
n x 1218413814154
Table 4. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 10  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not chosen for y 5 and u 7 is not included in state vector for y 2 .
Table 4. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 10  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not chosen for y 5 and u 7 is not included in state vector for y 2 .
Variables y 1 = ψ C O y 2 = ψ S O 2 y 3 = ψ NO x y 4 = ψ O 2 y 5 = ψ C O 2
Variables used for the input vector u 1 = ψ f g -----
u 2 = t r u n -----
u 3 = t d -----
u 4 = V a i r , p r i m 3311, 2, 112, 4-
u 5 = V r e c 14, 13, …, 10,
5, 4, …, 1
15, 12, 11, 9,
8, 5, 1
15, 14, …, 115, 14, …113, 11, 10, 8,
4, 2, 1
u 6 = v a i r , s e c 15, 14, …, 19, 215, 14, …, 115, 14, …, 114, 13, …7,
3, 2
u 7 = V O 2 , p r i m 15, 14, …, 1-15, 14, …, 115, 14, …, 115, 12, 11, 9,
8, 6, 5, 3
u 8 = V O 2 , s e c 15, 14, 10,
9, 6
14, 1315, 14, …, 115, 14, …, 113, 10, 9, 6, 1
y 1 = ψ CO 15, 14, …, 115, 5, 4, …, 115, 14, …, 115, 14, …, 115, 13, 12, …, 8,
5, 4, 2, 1
y 2 = ψ SO 2 15, 14, …, 115, 14, 8,
7, …, 5, 2, 1
15, 14, …, 115, 14, …, 114, 12, 11, 10,
5, 3, 2, 1
y 3 = ψ NO x 15, 14, …, 113, 8, 7,
6, …, 3, 1
15, 14, …, 115, 14, …, 115, 14, 12, …, 7,
5, 3, 1
y 4 = ψ O 2 15, 13, 12,
11, …, 5,
3, 2, 1
5, 4, 3, 215, 14, …, 115, 14, …, 115, …, 12,
10, …, 8,
5, …1
y 5 = ψ CO 2 15, 14, …, 18, 7, …, 4,
2, 1
15, 14, …, 115, 14, …, 114, 13, 11, 10,
9, 8, 6, 4,
3, 2, 1
n x 1174513813683
Table 5. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 15  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not chosen for y 3 .
Table 5. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 15  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not chosen for y 3 .
Variables y 1 = ψ C O y 2 = ψ S O 2 y 3 = ψ NO x y 4 = ψ O 2 y 5 = ψ C O 2
Variables used for the input vector u 1 = ψ f g -----
u 2 = t r u n -----
u 3 = t d -----
u 4 = V a i r , p r i m 15, 14,
10, 2
1-15, 7, 28, 7, 6
u 5 = V r e c 15, 14, …, 8,
6, 5, 1
14, 13, 11,
9, 8, …, 5,
3, 1
15, 14, 13, 9,
8, 6, 3, 2, 1
15, 14, …115, 14, …, 1
u 6 = v a i r , s e c 15, 14, …, 110, 8, 515, 14, 12, …, 115, 14, …, 115, 14, …, 1
u 7 = V O 2 , p r i m 15, 14, …, 115, 7, 5, 4, 215, 14, …, 115, 14, …, 115, 14, …, 1
u 8 = V O 2 , s e c 15, 13, 10, 6,
4, 3, 1
14, 13, 12, 615, 14, …, 115, 14, …, 115, 14, …, 1
y 1 = ψ CO 15, 14, …, 115, 13, 11, 10,
4, 2, 1
15, 14, …, 115, 14, …, 115, 14, …, 1
y 2 = ψ SO 2 15, 14, …, 115, 14, …, 10,
8, …, 3, 1
15, 14, …, 115, 14, …, 115, 14, …, 1
y 3 = ψ NO x 15, 14, 12,
11, …, 11
15, 14, 13, 5,
3, 2, 1
15, 14, …, 115, 14, …, 115, 14, …, 1
y 4 = ψ O 2 15, 14, 17, 6, …, 115, 14, …, 115, 14, …, 115, 14, …, 1
y 5 = ψ CO 2 15, 14, …, 115, 12, 11,
6, 5, …1
15, 14, …, 115, 14, …, 115, 14, …, 1
n x 12667127138138
Table 6. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 25  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not included in state vector for y 5 .
Table 6. Table of state vectors x i for outputs y i (listed in columns) proposed by Boruta Feature Selection Algorithm for sampling period Δ T s = 25  s. Features u 1 , u 2 , u 3 were not picked by BFSA at all, u 4 is not included in state vector for y 5 .
Variables y 1 = ψ C O y 2 = ψ S O 2 y 3 = ψ NO x y 4 = ψ O 2 y 5 = ψ C O 2
Variables used for the input vector u 1 = ψ f g -----
u 2 = t r u n -----
u 3 = t d -----
u 4 = V a i r , p r i m 14, 10, 9, 29, 5, 110, 7, 214-
u 5 = V r e c 15, 14, 12, 10,
9, 7, 6,
5, …, 1
14, 8, 6, 4,
3, 2, 1
15, 13, …, 4,
3, 1
15, 14, …11,
9, 8, …, 1
15, 13, …, 10,
8, …, 4, 2, 1
u 6 = v a i r , s e c 15, 14, …, 115, 10, 915, 14, …, 115, 14, …, 115, 14, …, 1
u 7 = V O 2 , p r i m 15, 14, …, 112, 11, …, 715, 14, 13, 12,
10, 9, …, 1
15, 14, …, 115, 14, …, 1
u 8 = V O 2 , s e c 15, 14, …, 7,
8, 3
15, 14, 13, 11,
10, 9, 6
15, 14, 12, 9,
8, 7, …, 1
15, 14, …, 113, 10, 8, 7,
6, 4, 3, 2, 1
y 1 = ψ CO 15, 14, …, 114, 13, …, 9,
6, 5, 1
15, 13, …, 10,
8, 7, …, 1
15, 14, …, 115, 14, …, 1
y 2 = ψ SO 2 15, 14, …, 115, 14, 12,
11, …, 1
15, 14, …, 115, 14, …, 115, 14, …, 9,
7, 6, 4, …, 1
y 3 = ψ NO x 15, 14, …, 112, 11, 2, 115, 14, …, 115, 14, …, 115, 14, …, 1
y 4 = ψ O 2 15, 14, …, 113, 10, 3,
2, 1
15, 14, …, 115, 14, …, 115, 14, …, 1
y 5 = ψ CO 2 15, 14, …, 111, 10, 9,
3, 1
15, 14, …, 115, 14, …, 115, 14, …, 1
n x 13063129135124
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Marzova, K.; Bukovsky, I. Feature Selection and Uncertainty Analysis for Bubbling Fluidized Bed Oxy-Fuel Combustion Data. Processes 2021, 9, 1757. https://0-doi-org.brum.beds.ac.uk/10.3390/pr9101757

AMA Style

Marzova K, Bukovsky I. Feature Selection and Uncertainty Analysis for Bubbling Fluidized Bed Oxy-Fuel Combustion Data. Processes. 2021; 9(10):1757. https://0-doi-org.brum.beds.ac.uk/10.3390/pr9101757

Chicago/Turabian Style

Marzova, Katerina, and Ivo Bukovsky. 2021. "Feature Selection and Uncertainty Analysis for Bubbling Fluidized Bed Oxy-Fuel Combustion Data" Processes 9, no. 10: 1757. https://0-doi-org.brum.beds.ac.uk/10.3390/pr9101757

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop