Acessibilidade / Reportar erro

Forecasting Chemical Characteristics of Aircraft Fuel Using Artificial Neural Networks

ABSTRACT

Aircraft fuels, called jet propulsion, are used in several areas of activity within aeronautics. There are jet fuels based on kerosene, that is, those obtained commercially, and there are synthetics produced in the laboratory. All of these fuels are included within the so-called propellants. In this article, Jet propulsion-8 (JP 8) fuel was used as the basis for data analysis, and thus two temperature ranges were analyzed. The first range, from 300 to 2500 K, was analyzed for specific heat, enthalpy and entropy. Based on theoretical and experimental data, artificial neural networks (ANNs) were developed to identify these properties in other working conditions, that is, at other temperatures.

Keywords
Fuel; Temperature; Enthalpy; Entropy; Heat

INTRODUCTION

Fuel is a substance that, when reacts with oxygen, releases energy. Therefore, there are several fuels, and among the best-known are those derived from petroleum, such as gasoline and diesel, which are normally used in land motor vehicles. Another type of fuel used in motor vehicles is nonpolluting fuels, such as electric cars.

In this context, there are also fuels that are used in aircraft. Most are based on kerosene, which is also a derivative of oil, and they have uses and characteristics that can be very different. Such characteristics are linked to the fact that these fuels are classified as propellants, that is, they are substances that, alone or in combination with others, are burned in a large mass of gaseous products at high temperature, which, escaping through the chamber discharge piston combustion, provide by reaction, the engine impulse.

These propellants may have commercial and/or military uses. Those for commercial use are employed in commercial or transport aircrafts and are based on kerosene, derived from petroleum, which has the name QAV-1. Those for military use have other names, such as jet propulsion-8 (JP-8). The jet propulsion-4 (JP-4) was used by the United States Air Force until 1995 and, after that, replaced by JP-8, as it is less flammable and less dangerous (Center 2008Center USDES (2008) Turbine Fuel, Aviation, Kerosene Type, JP-8 (NATO F-34), NATO F-35, and JP-8+100 (NATO F-37). [S.l.]. [accessed 9 10 2020].).

The JP-8 has 81.20% isoparaffins, 0.30% oils, 4.90% naphthalene and 13.00% aromatic compounds (Pires et al 2018Pires APP, Han Y, Kramlich J, Garcia-Perez M (2018) Chemical composition and fuel properties of alternative jet fuels. BioResources 13(2):2632-2657. https://doi.org/10.15376/biores.13.2.2632-2657
https://doi.org/10.15376/biores.13.2.263...
). Among them, 16.90% is C9H12, 62.55% is C11H24 and 21.05% is C10H20 (Xu et al. 2015Xu R, Wang H, Colket M, Edwards M (2015) Thermochemical properties of jet fuels [Internet]. Stanford: Stanford University. [accessed Sep 18 2020]. https://web.stanford.edu/group/haiwanglab/HyChem/approach/Report_Jet_Fuel_Thermochemical_Properties_v6.pdf
https://web.stanford.edu/group/haiwangla...
).

State-of-the-art

The study of propellants is something that is being done a lot nowadays, because more and more are looking for cheap and easily manufactured fuels. As they are flammable fuels, it is not possible to work at high temperatures in the laboratory. Therefore, the study at high temperatures simulates certain fuels at very high combustion rates, where in a laboratory it would not be able to reach.

Therefore, it is possible to study a variety of temperatures and select within them the one that is the most interesting. The use of neural networks helps in this process, where it is possible to train a network to analyze and give results close to the experimental ones and also at temperatures that it is not possible to work with.

So, with that the range of temperatures to be studied for future fuels is huge.

THEORY

Artificial neural network

The work on artificial neural networks, usually called “neural networks”, has been motivated since the beginning by the recognition that the human brain processes information in an entirely different way from the convectional digital computer (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.). They allow the projection of values on a laboratory scale and on pilot scale in the various areas studied worldwide. Thus, the human nervous system can be seen as a three-stage system, as show in Fig. 1 (Arbib 1987Arbib MA (1987) Brains, machines, and mathematics. New York: Springer.).

Figure 1
Block diagram representation of nervous system.

The center of the system is the brain, represented by the neural network, which continuously receives information, perceives and makes appropriated decisions. Two sets of arrows are shown in Fig. 1. Those from left to right indicate the forward transmission of the information-carrying signal through the system. Arrows pointing from right to left indicate the presence of feedback in the system. The receptors convert stimuli from the human body or the external environment into electrical impulses that transmit information to the neural network. The actuators convert electrical impulses generated by the neural network into discernible responses as outputs from the system (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.).

In practice, however, neural networks cannot provide a solution working individually. Instead, they need to be integrated into a consistent system engineering approach (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.). Thus, a complex problem is decomposed into a number of relatively simple tasks, and a subset of tasks is assigned to neural networks that coincide with their ability to solve such problem.

The use of neural networks offers the following useful properties and capabilities:

  • Nonlinearity;

  • Input-output mapping;

  • Adaptability;

  • Responses to evidence;

  • Contextual information;

  • Fault tolerance;

  • Very large-scale integration;

  • Uniformity of analysis and design;

  • Neurobiological analogy.

Activation function

There are several activation functions that can be used and the choice depends on the purpose and complexity of the neural network. For this, it is necessary to analyze the problem involved in order to choose the best activation function. Table 1 shows the most frequently used.

Table 1
Activation functions.

Learning/training algorithms

Learning and training algorithms serve to “teach” the neural network how to work with the data. Thus, it is able to predict values and make approximations closest to the real value. There are two types of learning algorithms, the so-called supervised and the unsupervised.

Unsupervised learning is used when the desired answer is not known, that is, there is no output value. It is possible to map the input data and bring it close to the expected output value. The most cited algorithms to date in this category are:

  • Additive Grossberg (AG) (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.);

  • Adaptive resonance theory (ART) (Carpenter and Grossberg 2010Carpenter GA, Grossberg S (2010) Adaptive resonance theory. In: Sammut C, Webb GI, editors. Encyclopedia of Machine Learning and Data Mining. Boston: Springer. https://doi.org/10.1007/978-1-4899-7502-7_6-1
    https://doi.org/10.1007/978-1-4899-7502-...
    );

  • Symmetric and asymmetric Hopfield (Sousa 2014Sousa FB (2014) Análise de modelo de Hopfield com topologia de rede complexa (Master thesis). São Carlos: Universidade de São Paulo. In Portuguese.);

  • Bidirectional associative memory (BAM) (Sousa 2014Sousa FB (2014) Análise de modelo de Hopfield com topologia de rede complexa (Master thesis). São Carlos: Universidade de São Paulo. In Portuguese.);

  • Temporal associative memory (TAM) (Vieira and Araujo 1997Vieira M, Araújo AFR (1997) Memória Associativa Multidirecional Temporal para Planejar Trajetória de um Manipulador. Paper presented 3° Simpósio Brasileiro de Automação Inteligente. SBAI; Vitória, Espírito Santo, Brazil.);

  • Competitive learning – SOM by Kohonen (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.).

There is also supervised learning, where the output values are known. This type of learning is widely used when there is a theoretical or experimental data, so it is possible to compare and even calculate errors between those values. It also allows to map the input data to get close to or equal to the expected output data. The most cited algorithms in this category are:

  • Boltzmann machine (BM) (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.);

  • Recurring real-time learning (Oliveira and Siqueira 2013Oliveira MA, Siqueira JO (2003) Aplicação do algoritimo de rede neural de aprendizagem recorrente em tempo real (RTRL) para previsão da série do Ibovespa. FEA. São Paulo: EAD/FEA/USP.);

  • Adaptive logic network (ALN);

  • Back propagation (BP) (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.);

  • Cauchy machine (CM);

  • Kalman filter (EKF) (Aguirre 2015Aguirre LA (2015) Introdução à Identificação de Sistemas: Técnicas Lineares e não Lineares Aplicadas a Sistemas. Minas Gerais: UFMG.).

Multilayer artificial neurons (multilayer perceptron)

Typically, the network consists of a set of sensory units (source nodes) that constitute the input layer, one or more hidden layers of computational nodes and an output layer of computational nodes. The input signal propagates forward through the network layer-by-layer (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.). Figure 2 shows the representation of a multilayer perceptron (Coutinho et al. 2016Coutinho ER, Silva RM; Delgado ARS (2016) Utilização de técnicas de inteligência computacional na predição de dados meteorológicos. Rev Bras Meteorol 31(1):24-36. https://doi.org/10.1590/0102-778620140115
https://doi.org/10.1590/0102-77862014011...
).

Figure 2
Representation of a multilayer perceptron.

Multilayer perceptron has been successfully applied to solve several complex problems through their supervised training with a very popular algorithm, known as error propagation algorithm. This algorithm is based on the error correction-learning rule (Haykin 2001Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.). One of the mostly used algorithms nowadays (and that will be used in this article) is the multilayer perceptron network trained by the retro propagation algorithm. The learning by retro propagation of error consists of two steps.

The descending gradient

The descending gradient is a technique used in learning algorithms that involves calculating derivatives of the objective function to find the point with the greatest possible slope in the direction of the minimum error. Such variables, to be optimized, are shifted in a negative direction, that is, this will reduce the value of the objective function and give the best possible optimization. This method can have two points, the gradient with and without the moment factor (Rodríguez 2011Rodríguez RAM (2011) Teoria e aplicação do classificador baseado em segmentos de reta em problemas de multiclassificação (Master thesis). São Paulo: Universidade de São Paulo. In Portuguese.).

MATLAB

MATLAB is a high-performance interactive software focused on numerical calculation. It also integrates numerical analysis, matrix calculation, signal processing and graph construction in an easy-to-use environment, where problems and solutions are expressed only as they are written mathematically, unlike traditional programming.

EXPERIMENTAL

The experimental part was based on the article by Xu et al. (2015)Xu R, Wang H, Colket M, Edwards M (2015) Thermochemical properties of jet fuels [Internet]. Stanford: Stanford University. [accessed Sep 18 2020]. https://web.stanford.edu/group/haiwanglab/HyChem/approach/Report_Jet_Fuel_Thermochemical_Properties_v6.pdf
https://web.stanford.edu/group/haiwangla...
. In the mentioned work, the authors obtained, experimentally, physical-chemical properties, such as enthalpy, entropy, specific heat, variation of enthalpy, heat of formation and Gibbs free energy for several types of aircraft fuels, such as jet fuel-A (JET-A), jet propulsion-5 (JP-5) and JP-8, and also alternative fuels from some companies, such as Gevo and Virent. The percentage of components that each fuel has is also determined. The experimental data are shown in Table 2.

Table 2
Jet propulsion-8 fuel experimental data. Retrieved from Xu R, Wang H, Colket M, Edwards M (2015)Xu R, Wang H, Colket M, Edwards M (2015) Thermochemical properties of jet fuels [Internet]. Stanford: Stanford University. [accessed Sep 18 2020]. https://web.stanford.edu/group/haiwanglab/HyChem/approach/Report_Jet_Fuel_Thermochemical_Properties_v6.pdf
https://web.stanford.edu/group/haiwangla...
Thermochemical properties of jet fuels [Internet]. Stanford: Stanford University. p. 6.

RESULTS AND DISCUSSION

Jet propulsion-8 was tested at three different temperatures for data prediction. The temperatures were chosen at random; the results were interpolated according to the values provided (Table 2), and the information for these tests is shown in Table 3.

Table 3
Neural network test data.

In the tests for neural networks, four prediction analyses were made, where the learning functions and the activation function of the first layer were analyzed, thus, 20 analyses were performed based on the number of neurons. These analyses followed a criterion, starting with 25 neurons, then 24 neurons, 23 neurons and so on until reaching 6 neurons, where the stop criterion for the analyses was made, since the minimum number of neurons must be equal to the sum of the output data plus one. In this case, the output data are: the specific heat, the entropy, the enthalpy variation, the formation enthalpy and the Gibbs free energy.

These data were obtained through predetermined factors and their performance was based on the calculation of the percentage of local and global errors, always based on the best global error with the fewest possible neurons as a stopping criterion.

For all tests, the same training and error function and the same number of layers were used, in this case, two layers. The training function used was trainlm, which refers to the Levenberg–Marquardt algorithm, considered the standard function in the MATLAB toolbox and evaluated by as the fastest method of training feedforward neural networks and with excellent performance (Friderichs et al. 2013Friderichs A, Cunha GA, Weber DP, Castro T, Hoffmann R, Salau NPG (2013) Uso de redes neurais na previsão do desempenho de um reator UASB para tratamento de vinhaça e geração de biogás. Paper presented XV Congresso Brasileiro de Energia (CBE). CBE; Rio de Janeiro, Rio de Janeiro, Brazil. [accessed Sep 18 2020]. http://coral.ufsm.br/cenergia/images/Uso_de_redes_neurais_na_previs%C3%A3o_do_desempenho_de_um_reat__or_UASB_para_tratamento_de_vinha%C3%A7a_e_gera%C3%A7%C3%A3o_de_biog%C3%A1s_1.pdf
http://coral.ufsm.br/cenergia/images/Uso...
).

To calculate the global error, the mean squared error (MSE) function was used, which corresponds to the mean square error. Local errors were calculated according to Eq. 1 (calculation of local error), where the interpolated values were adopted as experimental values and the theoretical value obtained by the MATLAB program.

L o c a l   error = A B S   ( T h e o r e u t i c a l   V a l u e - E c p e r i m e n t a l   V a l u e ) E x p e r i m e n t a l   Value (1)

The other parameters were changed, such as the learngdm and learngd learning function, which correspond to the descending gradient function with and without moment factor and activation function in the first layer. In the second layer, the purelin function was maintained, which corresponds to a first-degree function, where these functions aim to guide and approximate the experimental result to the real result.

For the first prediction, the learngdm learning function and the activation functions tansig and purelin (Table 4) were used and the results obtained for the first test are shown in Table 5. Table 6 shows the percentage of local error.

Table 4
First prediction of neural networks.
Table 5
Results of the first forecast.
Table 6
Local errors of the first forecast.

For the second forecast, the learngdm learning function was maintained, but the activation function of the first layer was changed to logsig (Table 7), where the objective was to compare results in the exchange of a tangential function for a logarithmic function (Table 8) and the local errors in Table 9.

Table 7
Second forecast of neural networks.
Table 8
Results according to forecast.
Table 9
Local errors according to forecast.

When comparing the first forecast with the second forecast, it is noted that the local errors did not have a very large difference in values and, for the first forecast, the global error was 3.72 × 10–5 or, in percentage, 0.00372%, and the number of neurons used was seven. For the second forecast, the global error was 2.04 × 10–5 or 0.00204%, but the number of neurons needed for the best global error was eight. So, it can be said that when changing the activation function of the first layer, there was a decrease in the global error, but one more neuron was necessary for this.

For the third forecast, the learning function was changed to learngd and changes in the activation functions of the first layer were maintained, as in this case the results were observed in the change in the learning function and in the change in the activation function of the first layer (Table 10). The results are shown in Table 11 and the local errors in Table 12.

Table 10
Third forecast neural networks.
Table 11.
Third forecast results.
Table 12
Local errors third forecast.

For the fourth and last forecast, the activation function was changed back to logsig (Table 13) and the results and local errors are shown in Table 14 and 15, respectively.

Table 13
Fourth forecast neural networks.
Table 14
Fourth forecast results.
Table 15
Local errors fourth forecast.

Comparing the results of the third with the fourth forecast, similarities in the results of local errors are noted again. For the third forecast, the global error was 3.43 × 10–5 or 0.00343% with six neurons used. For the fourth forecast, the global error was 1.80 × 10–5 or 0.00180% with the same six neurons used. It can be seen then that the error was greater when the function of the first layer was changed from tansig to logsig.

CONCLUSION

The results obtained were satisfactory because they meet the stopping criteria with the best overall error in the 10–5 range and the fewest possible neurons. The number of neurons is very important due to the propagation of error that can exist, since the greater the number of neurons used, the greater the propagation of error; however, the neurons used do not exceed eight. Therefore, comparing the four simulations, an increase in local and global error was observed due to the change in the learning function, therefore, the descending gradient function without the moment factor raises the value of errors. It was also observed that the change in the learning functions of the first layer, for all tests, raised the value of the errors too, so changing a function of a tangential character to a function of a logarithmic character also interferes with the increase of the error. The best result between the simulations, taking into account the best global error with the fewest possible neurons, was the first prediction.

ACKNOWLEDGEMENTS

Not applicable.

  • FUNDING

    Not applicable.
  • DATA AVAILABILITY STATEMENT

    All data sets were generated or analyzed in the current study.
  • Peer Review History: Single Blind Peer Review.

REFERENCES

  • Aguirre LA (2015) Introdução à Identificação de Sistemas: Técnicas Lineares e não Lineares Aplicadas a Sistemas. Minas Gerais: UFMG.
  • Arbib MA (1987) Brains, machines, and mathematics. New York: Springer.
  • Carpenter GA, Grossberg S (2010) Adaptive resonance theory. In: Sammut C, Webb GI, editors. Encyclopedia of Machine Learning and Data Mining. Boston: Springer. https://doi.org/10.1007/978-1-4899-7502-7_6-1
    » https://doi.org/10.1007/978-1-4899-7502-7_6-1
  • Center USDES (2008) Turbine Fuel, Aviation, Kerosene Type, JP-8 (NATO F-34), NATO F-35, and JP-8+100 (NATO F-37). [S.l.]. [accessed 9 10 2020].
  • Coutinho ER, Silva RM; Delgado ARS (2016) Utilização de técnicas de inteligência computacional na predição de dados meteorológicos. Rev Bras Meteorol 31(1):24-36. https://doi.org/10.1590/0102-778620140115
    » https://doi.org/10.1590/0102-778620140115
  • Friderichs A, Cunha GA, Weber DP, Castro T, Hoffmann R, Salau NPG (2013) Uso de redes neurais na previsão do desempenho de um reator UASB para tratamento de vinhaça e geração de biogás. Paper presented XV Congresso Brasileiro de Energia (CBE). CBE; Rio de Janeiro, Rio de Janeiro, Brazil. [accessed Sep 18 2020]. http://coral.ufsm.br/cenergia/images/Uso_de_redes_neurais_na_previs%C3%A3o_do_desempenho_de_um_reat__or_UASB_para_tratamento_de_vinha%C3%A7a_e_gera%C3%A7%C3%A3o_de_biog%C3%A1s_1.pdf
    » http://coral.ufsm.br/cenergia/images/Uso_de_redes_neurais_na_previs%C3%A3o_do_desempenho_de_um_reat__or_UASB_para_tratamento_de_vinha%C3%A7a_e_gera%C3%A7%C3%A3o_de_biog%C3%A1s_1.pdf
  • Haykin S (2001) Redes neurais: princípios e prática. Brussels: Bookman.
  • Oliveira MA, Siqueira JO (2003) Aplicação do algoritimo de rede neural de aprendizagem recorrente em tempo real (RTRL) para previsão da série do Ibovespa. FEA. São Paulo: EAD/FEA/USP.
  • Pires APP, Han Y, Kramlich J, Garcia-Perez M (2018) Chemical composition and fuel properties of alternative jet fuels. BioResources 13(2):2632-2657. https://doi.org/10.15376/biores.13.2.2632-2657
    » https://doi.org/10.15376/biores.13.2.2632-2657
  • Rodríguez RAM (2011) Teoria e aplicação do classificador baseado em segmentos de reta em problemas de multiclassificação (Master thesis). São Paulo: Universidade de São Paulo. In Portuguese.
  • Sousa FB (2014) Análise de modelo de Hopfield com topologia de rede complexa (Master thesis). São Carlos: Universidade de São Paulo. In Portuguese.
  • Vieira M, Araújo AFR (1997) Memória Associativa Multidirecional Temporal para Planejar Trajetória de um Manipulador. Paper presented 3° Simpósio Brasileiro de Automação Inteligente. SBAI; Vitória, Espírito Santo, Brazil.
  • Xu R, Wang H, Colket M, Edwards M (2015) Thermochemical properties of jet fuels [Internet]. Stanford: Stanford University. [accessed Sep 18 2020]. https://web.stanford.edu/group/haiwanglab/HyChem/approach/Report_Jet_Fuel_Thermochemical_Properties_v6.pdf
    » https://web.stanford.edu/group/haiwanglab/HyChem/approach/Report_Jet_Fuel_Thermochemical_Properties_v6.pdf

Edited by

Section Editor: T John Tharakan

Publication Dates

  • Publication in this collection
    21 May 2021
  • Date of issue
    2021

History

  • Received
    09 Oct 2020
  • Accepted
    16 Mar 2021
Departamento de Ciência e Tecnologia Aeroespacial Instituto de Aeronáutica e Espaço. Praça Marechal do Ar Eduardo Gomes, 50. Vila das Acácias, CEP: 12 228-901, tel (55) 12 99162 5609 - São José dos Campos - SP - Brazil
E-mail: submission.jatm@gmail.com