Acessibilidade / Reportar erro

A neural network for prediction of risk of nosocomial infection at intensive care units: a didactic preliminary model

ABSTRACT

Objective:

To propose a preliminary artificial intelligence model, based on artificial neural networks, for predicting the risk of nosocomial infection at intensive care units.

Methods:

An artificial neural network is designed that employs supervised learning. The generation of the datasets was based on data derived from the Japanese Nosocomial Infection Surveillance system. It is studied how the Java Neural Network Simulator learns to categorize these patients to predict their risk of nosocomial infection. The simulations are performed with several backpropagation learning algorithms and with several groups of parameters, comparing their results through the sum of the squared errors and mean errors per pattern.

Results:

The backpropagation with momentum algorithm showed better performance than the backpropagation algorithm. The performance improved with the xor. README file parameter values compared to the default parameters. There were no failures in the categorization of the patients into their risk of nosocomial infection.

Conclusion:

While this model is still based on a synthetic dataset, the excellent performance observed with a small number of patterns suggests that using higher numbers of variables and network layers to analyze larger volumes of data can create powerful artificial neural networks, potentially capable of precisely anticipating nosocomial infection at intensive care units. Using a real database during the simulations has the potential to realize the predictive ability of this model.

Keywords:
Artificial neural network; Nosocomial infection; Intensive care units; APACHE; Artificial intelligence

RESUMO

Objetivo:

Propor um modelo preliminar de inteligência artificial, baseado em redes neurais artificiais, para previsão do risco de infecção hospitalar em unidades de cuidado intensivo.

Métodos:

Foi usada uma rede neural artificial, que utiliza aprendizagem supervisionada. A geração dos conjuntos de dados baseia-se em dados derivados do sistema Japanese Nosocomial Infection Surveillance . Estudamos como o Java Neural Network Simulator aprende a categorizar esses pacientes para prever o respectivo risco de infecção hospitalar. As simulações são realizadas com diferentes algoritmos de aprendizagem por retropropagação e diversos grupos de parâmetros, comparando-se os resultados com base na soma dos erros quadráticos e erros médios por padrão.

Resultados:

O algoritmo de retropropagação com momentum mostrou desempenho superior ao do algoritmo de retropropagação. O desempenho foi melhor com os valores de parâmetros do arquivo xor. README em comparação aos parâmetros default . Não houve falhas na categorização de pacientes quanto ao respectivo risco de infecção hospitalar.

Conclusão:

Embora esse modelo se baseie em um conjunto de dados sintéticos, o excelente desempenho observado com um pequeno número de padrões sugere que o uso de números maiores de variáveis e camadas de rede para analisar volumes maiores de dados pode criar redes neurais artificiais poderosas, possivelmente capazes de prever com precisão o risco de infecção hospitalar em unidades de cuidado intensivo. O uso de um banco de dados real durante as simulações torna possível a realização da capacidade preditiva desse modelo.

Descritores:
Rede neural artificial; Infecção hospitalar; Unidades de terapia intensiva; APACHE; Inteligência artificial

INTRODUCTION

An important cause of morbidity and mortality are nosocomial infections, implying a significant burden on patients and hospitals.(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) These complications are highly associated with stay at intensive care units (ICU).(22. Stiller A, Schröder C, Gropmann A, Schwab F, Behnke M, Geffers C, et al. ICU ward design and nosocomial infection rates: a cross-sectional study in Germany. J Hosp Infect. 2017;95(1):71-5.44. Merzougui L, Barhoumi T, Guizani T, Barhoumi H, Hannachi H, Turki E, et al. [Nosocomial infections in the Intensive Care Unit: annual incidence rate and clinical aspects]. Pan Afr Med J. 2018;30:143. French.) Studies have shown that ICU-acquired nosocomial infections have an important effect on mortality.(55. Januel JM, Harbarth S, Allard R, Voirin N, Lepape A, Allaouchiche B, et al. Estimating attributable mortality due to nosocomial infections acquired in intensive care units. Infect Control Hosp Epidemiol. 2010;31(4):388-94.77. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.)

Identification of many nosocomial infections is currently conducted via specimen culture, which can take a few days. Consequently, an overuse of broad-spectrum antibiotics is the current treatment, thus treating potential cases who may not suffer from these complications and resulting in the potential risk of evolving antibiotic-resistant strains.(77. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.)

Surveillance of nosocomial infections, together with appropriate prevention measures, can decrease infection rates and improve patient safety. It is crucial for infection control professionals to predict and measure the risk of infection at their hospitals, contributing to planning and assessment of infection control programs.(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) There are national or regional programs to support hospitals in reducing the risk of nosocomial infections, e.g. , the Japanese Nosocomial Infection Surveillance (JANIS) system in Japan, and the Healthcare-Associated Infections Surveillance Network in Europe (HAI-Net). The World Health Organization (WHO) recommends that risk prevention of nosocomial infections must also be supported by hospital programs.

If predictive tools could be evolved to reveal those most likely to contract these nosocomial infections, then a reduced risk of resistant strains would ensue, leading to significant savings regarding mortality and treatment expenses.

Prediction of nosocomial infections is generally subjective, dependent on interpretation, in which clinicians with widely varying experiences assess patient conditions. This dependency restricts the ability to obtain accurate correlations between risk factors and outcome predictions. Thus, it is a matter of experience the value applied to any given situation.(88. Pearl A, Caspi R, Bar-Or D. Artificial neural network versus subjective scoring in predicting mortality in trauma patients. Stud Health Technol Inform. 2006;124:1019-24.) However, protocols derived from the national or regional programs help ensuring standardization of definitions orienting diagnosis and treatment in ICU, in addition to institutional protocol guidance.

The application of artificial neural networks (ANN) in outcome prediction has become increasingly prevalent in physiological modeling(88. Pearl A, Caspi R, Bar-Or D. Artificial neural network versus subjective scoring in predicting mortality in trauma patients. Stud Health Technol Inform. 2006;124:1019-24.,99. Nimgaonkar A, Karnad DR, Sudarshan S, Ohno-Machado L, Kohane I. Prediction of mortality in an Indian intensive care unit. Comparison between APACHE II and artificial neural networks. Intensive Care Med. 2004;30(2):248-53.) and several medical fields,(77. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.,1010. Pearl A, Bar-Or D. Decision support in trauma management: predicting potential cases of ventilator associated pneumonia. Stud Health Technol Inform. 2012;180:305-9.1717. Hsieh MH, Hsieh MJ, Chen CM, Hsieh CC, Chao CM, Lai CC. An Artificial neural network model for predicting successful extubation in intensive care units. J Clin Med. 2018;7(9):240.) due to the ability of the ANN to learn and improve.

Artificial neural networks are artificial intelligence models built based on biological neural systems. These ANN are adaptive systems which are increasingly used for prediction and are especially adequate for predictions of classification events.

Literature comparing different alternatives

The study by Burke et al.,(1818. Burke HB, Rosen DB, Goodman PH. Comparing artificial neural networks to other statistical methods for medical outcome prediction. Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94). 1994; p. 2213-16.) compared ANN to other statistical methods for medical outcome prediction, particularly survival prediction. They found that using just the tumor-node-metastasis (TNM) variables, both the backpropagation neural network and the probabilistic neural network were significantly more accurate than the pathological TNM stage system.(1818. Burke HB, Rosen DB, Goodman PH. Comparing artificial neural networks to other statistical methods for medical outcome prediction. Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94). 1994; p. 2213-16.)

Eftekhar et al.,(1919. Eftekhar B, Mohammad K, Ardebili HE, Ghodsi M, Ketabchi E. Comparison of artificial neural network and logistic regression models for prediction of mortality in head trauma based on initial clinical data. BMC Med Inform Decis Mak. 2005;5:3.) contrasted the performance of ANN and multivariable logistic regression (LR) models in prediction of outcomes in head trauma. They found that ANN significantly outperformed LR in the fields of discrimination and calibration.(1919. Eftekhar B, Mohammad K, Ardebili HE, Ghodsi M, Ketabchi E. Comparison of artificial neural network and logistic regression models for prediction of mortality in head trauma based on initial clinical data. BMC Med Inform Decis Mak. 2005;5:3.) DiRusso et al.,(2020. DiRusso SM, Chahine AA, Sullivan T, Risucci D, Nealon P, Cuff S, et al. Development of a model for prediction of survival in pediatric trauma patients: comparison of artificial neural networks and logistic regression. J Pediatr Surg. 2002;37(7):1098-104.) developed an ANN to predict pediatric trauma death and compared it with LR. The ANN model yielded excellent discrimination and calibration exceeding that of LR.(2020. DiRusso SM, Chahine AA, Sullivan T, Risucci D, Nealon P, Cuff S, et al. Development of a model for prediction of survival in pediatric trauma patients: comparison of artificial neural networks and logistic regression. J Pediatr Surg. 2002;37(7):1098-104.) Erguzel et al.,(2121. Erguzel TT, Noyan CO, Eryilmaz G, Ünsalver BÖ, Cebi M, Tas C, et al. Binomial logistic regression and artificial neural network methods to classify opioid-dependent subjects and control group using quantitative EEG power measures. Clin EEG Neurosci. 2019;50(5):303-10.) used the two mentioned methods to create a model dichotomizing opioid-dependent patients and control subjects. The ANN classifier outperformed the LR.(2121. Erguzel TT, Noyan CO, Eryilmaz G, Ünsalver BÖ, Cebi M, Tas C, et al. Binomial logistic regression and artificial neural network methods to classify opioid-dependent subjects and control group using quantitative EEG power measures. Clin EEG Neurosci. 2019;50(5):303-10.) Liew et al.,(2222. Liew PL, Lee YC, Lin YC, Lee TS, Lee WJ, Wang W, et al. Comparison of artificial neural networks with logistic regression in prediction of gallbladder disease among obese patients. Dig Liver Dis. 2007;39(4):356-62.) retrospectively analyzed the prevalence and risk factors of gallbladder disease using LR and ANN among obese patients. Artificial neural networks, constructed with backpropagation algorithm, were trained to predict the risk of gallbladder disease. Artificial neural networks demonstrated better average classification rate, and lower type II errors than LR.(2222. Liew PL, Lee YC, Lin YC, Lee TS, Lee WJ, Wang W, et al. Comparison of artificial neural networks with logistic regression in prediction of gallbladder disease among obese patients. Dig Liver Dis. 2007;39(4):356-62.)

The use of ANN in this particular setting has not been widely studied. A few related studies have been carried out successfully on trauma patients,(77. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.,1010. Pearl A, Bar-Or D. Decision support in trauma management: predicting potential cases of ventilator associated pneumonia. Stud Health Technol Inform. 2012;180:305-9.) whereas other recent studies have concentrated on mortality prediction.(77. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.99. Nimgaonkar A, Karnad DR, Sudarshan S, Ohno-Machado L, Kohane I. Prediction of mortality in an Indian intensive care unit. Comparison between APACHE II and artificial neural networks. Intensive Care Med. 2004;30(2):248-53.,2323. Ge W, Huh JW, Park YR, Lee JH, Kim YH, Turchin A. An interpretable ICU mortality prediction model based on logistic regression and recurrent neural networks with LSTM units. AMIA Annu Symp Proc. 2018;2018:460-9.)

Supervised learning in artificial neural networks

The formal model underlying neural computation in ANN is a parallel and directed graph in which the nodes are associated to models of local calculi, and the links represent the interconnections between these local calculi.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
)

The output of each node can generally be expressed as the weighted sum of the coordinates of the input array (x) multiplied by the coordinates corresponding to the array of weights (w), as Equation 1 .

(Equation 1) y j * ( t ) = Σ i = 1 to M [ w j i ( t ) * x i ( t ) ] for j = 1 , 2 , N 0

The result of this sum afterwards passes through a decision function, that produces the final outputs yj(t) ( Equation 2 ).

(Equation 2) y j ( t ) = u [ y j * ( t ) ]

By its derivable character (to calculate the new weights), the sigmoid function is usually used as the decision function, which provides an activation value in the range [0.1]. The local function is adjusted through learning processes. By using an algorithm methodology known as supervised learning, the algorithm varies the value of the weights connecting neuronal units according to deviations from a training set of data, which specifies the activation values of the output units corresponding to a set of input observations (labeled data).(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) The problems solved with supervised learning are based on adaptive numerical classifiers.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
)

Backpropagation in multilayer networks

This function is especially adequate for propagating the observed error to hidden layers, which have not available the desired outputs for those hidden neurons. This ANN is trained with supervised learning type of backpropagation.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) In this way, the errors are propagated backward.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
)

OBJECTIVE

The purpose of this artificial neural network is to calculate the risk of nosocomial infection for the intensive care unit patients evaluated. This is from the perspective of a simulated model, based on synthetic datasets derived from an actual database for the purpose of demonstration. It is intended to present a preliminary artificial neural network in this medical context, and show its construction and step-by-step analyses, so that adjustments and improvements can be easily incorporated to meet the needs of specific users. It is explained how to propose and solve this model through an artificial neural network utilizing supervised learning algorithms. The differences in behavior of the algorithms are analyzed in terms of the parameter variations, trying to find an optimal design.

METHODS

Selection of factors associated with the development of nosocomial infection

The selection of the risk factors for the elaboration of the synthetic datasets was based on the factors revealed by Suka et al.(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) This study was grounded on data acquired from the ICU component of the JANIS system to elucidate factors associated with the development of nosocomial infections and to determine infection rates for benchmarking. These factors were selected based on a multivariate analysis. They determined hazard ratios (HR) with 95% confidence intervals for the various factors selected.

The generation of the synthetic datasets was based on the revealed HR of the factors associated to the development of nosocomial infections.(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) Data were constructed derived from the information given by Barraclough et al.,(2525. Barraclough H, Simms L, Govindan R. Biostatistics primer: what a clinician ought to know: hazard ratios. J Thorac Oncol. 2011;6(6):978-82. Erratum in: J Thorac Oncol. 2011;6(8):1454.) with each binary value of each variable obtained from a probability distribution according to its HR, depending on the classification of that patient. The training and validation datasets follow a realistic distribution of the values involved.

Spreadsheet composition

All patients share several attributes that can be quantified. The columns in the spreadsheet represent these attributes, which distinguish the patients. Binary variables are used to represent if each patient possesses each attribute. These binary variables are used to describe the activity of the input and output neuronal units.

In the spreadsheets, each row, or pattern, stands for a different ICU patient. These data are imported into the Java Neural Network Simulator (JavaNNS)(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) to execute the simulations. This work analyzes how the JavaNNS(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) learns to represent and categorize these patients, based on their selected attributes, into the categories of present and absent risk of nosocomial infections. That is, a patient who is at risk of suffering or not from nosocomial infections.

Several major patient-level risk factors associated with the development of nosocomial infections were selected, as described above, for characterization of the patients, according to table 1 .

Table 1
Major patient-level risk factors selected associated with the development of nosocomial infections

Since these factors were selected based on the study of Suka et al.,(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) who utilized the JANIS database, this fictitious ICU would have the same characteristics: patients aged 16 years or more, who had remained at the ICU for 48 to 1,000 hours, who had not been moved to another ICU, and had not been infected within 2 days after ICU admission.(11. Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.) The categorization provided by the ANN is into the following output unit: risk of nosocomial infections – yes=1; no=0.

Each row represents a set of input and output units. Two different spreadsheets were created with these characteristics. One has 15 patients or patterns that was used for training the network.(2727. Nistal-Nuño B. Replication data for: a neural network for prediction of risk of nosocomial infection at intensive care units, V1 [Internet]. Harvard Dataverse; 2020 [cited Oct 15]. Available from: https://doi.org/10.7910/DVN/BET7VI
https://doi.org/10.7910/DVN/BET7VI...
) This training set was chosen to be of a size of 15 times the number of classes to obtain.(2828. Raudys SJ, Jain AK. Small sample size effects in statistical pattern recognition: recommendations for practitioners. IEEE Trans Pattern Anal Mach Intell. 1991;13(3):252-64.) The other spreadsheet has six patients or patterns that was used for validation of the network.(2727. Nistal-Nuño B. Replication data for: a neural network for prediction of risk of nosocomial infection at intensive care units, V1 [Internet]. Harvard Dataverse; 2020 [cited Oct 15]. Available from: https://doi.org/10.7910/DVN/BET7VI
https://doi.org/10.7910/DVN/BET7VI...
) This validation set was chosen to be of a size of six times the number of classes to obtain.(2828. Raudys SJ, Jain AK. Small sample size effects in statistical pattern recognition: recommendations for practitioners. IEEE Trans Pattern Anal Mach Intell. 1991;13(3):252-64.)

Elaboration of the pattern files of Java Neural Network Simulator

To create the JavaNNS pattern files, the MATLAB® program was used, version 7.5.0.342 (R2007b). MATLAB® has this functionality, which is used by typing the command “xls2nns([])”. A new file is then created with the same name as the spreadsheet but with a .pat extension. The two pattern files created, ICU trainingdata_NI.pat(2727. Nistal-Nuño B. Replication data for: a neural network for prediction of risk of nosocomial infection at intensive care units, V1 [Internet]. Harvard Dataverse; 2020 [cited Oct 15]. Available from: https://doi.org/10.7910/DVN/BET7VI
https://doi.org/10.7910/DVN/BET7VI...
) and ICU validationdata_NI.pat,(2727. Nistal-Nuño B. Replication data for: a neural network for prediction of risk of nosocomial infection at intensive care units, V1 [Internet]. Harvard Dataverse; 2020 [cited Oct 15]. Available from: https://doi.org/10.7910/DVN/BET7VI
https://doi.org/10.7910/DVN/BET7VI...
) are used for training and validation of the network respectively. Validation is used to determine the performance of the ANN on patterns that are not trained during learning.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
)

Neural network simulations

A multilayer perceptron ANN, a feed-forward network without shortcut connections is used. In JavaNNS, going to Tools/Create/Layers, the width is changed to be 12, the number of neuronal input units. Again, in Tools/Create/Layers, the width is changed to be 1, the number of output units. Afterwards, the network is connected feed forward.(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) The resulting network is shown in figure 1 .

Figure 1
The utilized neuronal network architecture designed specifically for this problem in Java Neural Network Simulator. The state of the network here shows a trained artificial neural network with the output prediction of the network for one of the validation patterns

The initializing function is set to random weights and the Init button is pressed, setting the weights to random values.(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
)

The utilized learning functions and learning-parameters

Among the supervised learning algorithms, the backpropagation type was used. There are several backpropagation algorithms provided with JavaNNS. The backpropagation and backpropagation with momentum algorithms(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) are employed to train this network.

The values of the parameters used in the simulations are shown in table 2 . These are obtained from the ones given in the xor. README file of the JavaNNS examples, and the ones given by default in JavaNNS.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
,2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) The parameters in the xor. README file are some parameters that Fischer et al.,(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) used to train successfully the XOR-problem on a neural network. This XOR-problem ANN has some similarities to the one constructed in this research. Fischer et al.,(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) suggested the number of cycles indicated in table 2 as the ones they needed to train the XOR-problem network successfully. These parameters were not obtained with extensive studies of statistical significance, they are given in JavaNNS as hints to start training sessions.(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
)

Table 2
Parameters values used for all models of combinations of learning functions and parameter sets

Simulations

Since there is a random component in the experiment, ten simulations are performed for each of the learning algorithms and set of parameters, comparing their results through the sum of the squared errors (SSE) of the Log and Error graph windows.(2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) The results are compared with previous parameters and between the different trials.

Using the validation set, in the Error graph window appears simultaneously an additional line of pink color that corresponds to the error of the validation set. It is utilized the SSE, the sum of the quadratic differences between the target output and the real output for all output units in all training or validation patterns. The Log window shows during training the SSE values in the validation and training sets.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) The SSE from the Log window are afterwards divided manually by the number of patterns in each set ( Table 3 ).

Table 3
Mean squared error per pattern for each one of the learning functions and sets of parameters, shown for each one of the sets of patterns. These are obtained from the sum of the squared errors shown in the Log window corresponding to the last trial and dividing by the number of patterns in each set

Once finished the training, in the Updating window, the training and validation patterns are passed one by one. Each pattern is visualized by colors and numerically, in all neurons of the network, to observe if the network succeeded or failed with each input pattern. The trained ANN is evaluated also by a confusion matrix.

The update mode topological order is used.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) Pressing Learn all in the learning tab cycles through all PATTERNS, for the number of times (cycles) specified. For the ten trials, the buttons Init and Learn all are pressed once for each trial. The graphs accumulate in the Error graph window in different colors ( Figures 2 and 3 ). A flowchart summarizing the methods is displayed in figure 4 .

Figure 2
Ten simulations performed in Java Neural Network Simulator with the backpropagation learning function, comparing their results through the sum of the squared errors in the Error graph window. The pink line corresponds to the sum of the squared errors of the validation set during the last trial. Bottom: with the xor. README file parameter values. Top: with the default parameter values
Figure 3
Ten simulations performed in Java Neural Network Simulator with the backpropagation with momentum learning function, comparing their results through the sum of the squared errors in the Error graph window. The pink line corresponds to the sum of the squared errors of the validation set during the last trial. On the right: with the xor. README file parameter values. On the left: with the default parameter values
Figure 4
Flowchart for supervised learning in Java Neural Network Simulator

RESULTS

Backpropagation learning function

The backpropagation algorithm had the following parameters: learning parameter η, indicating the step width of the gradient descent, and dmax , corresponding to the maximum difference between a target output and an actual output which is tolerated.(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
)

Backpropagation with momentum learning function

The backpropagation with momentum algorithm had the following parameters:(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) learning parameter η, as mentioned; μ, momentum term, indicating the amount of the previous weight adjustment (relative to 1) which was added to the current change;(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) c, flat spot elimination value, a constant added to the derivative of the activation function so the network could pass flat spots of the error surface,(2424. Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
http://www.ra.cs.uni-tuebingen.de/downlo...
) and dmax, as mentioned.

Errors table and comments of the results

In figure 2 , bottom, the SSE from the training patterns disappeared in all trials before 40 cycles, well before the maximum of 2,000 cycles. The MSE of the validation set was 0.02760293, which was a very small error. At 2,000 cycles this value stayed the same, at this MSE the network performed best and was when the training should be stopped.

In figure 2 , top, the SSE decreased progressively until reaching 0.0 in all trials before 450 cycles. The pink line reached its minimum MSE, also a very small error. In comparison to figure 2 , bottom, the default values showed here a worse performance.

In figure 3 , right, the SSE was 0.0 at before ten cycles for all trials. The pink line MSE was 0.0. This backpropagation with momentum algorithm showed improvement in comparison to the backpropagation algorithm.

In figure 3 , left, the SSE decreased progressively until becoming 0.0 at 100 cycles. The pink line MSE was 0.011535088, a very small value. The performance with the default values was worse than with the xor. README file values for this backpropagation with momentum algorithm. For the default parameter values, this backpropagation with momentum algorithm showed better performance than the backpropagation algorithm.

In table 4 , it is presented the classification obtained after training for the 15 training and six validation patterns showing the successes and failures. A pattern was classified correctly into present risk of nosocomial infections if the output unit activation was ≥0.6. A pattern was classified correctly into absent risk of nosocomial infections if the output unit activation was ≤0.4.

Table 4
Classification for the 15 training and six validation patterns, obtained in the Updating tab after training with the training patterns

There were no failures in the categorization of the patients into their corresponding risk of nosocomial infections. The backpropagation with momentum with the xor. README file values showed an almost nil error. The network performed best with the backpropagation with momentum algorithm with the xor. README file values.

Evaluation by a confusion matrix

The most likely prediction state for each pattern was then compared with the true value of the output unit activation for the particular case. The confusion matrix supplies the total number of cases in 21 situations. If the network was performing well, then the entries along the main diagonal would be large, compared to those off of it. This is observed in table 5 .

Table 5
Confusion matrix of predicted and actually occurred outcome of states for risk of nosocomial infection in 21 cases of the training and validation datasets

DISCUSSION

The results revealed relative variations in performance for the different combinations of learning algorithms and parameter sets. The function of backpropagation with momentum showed the best results. The results were better with the xor. README file parameter values. However, as stated by Fischer et al ., (2626. Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
http://www.ra.cs.uni-tuebingen.de/softwa...
) these parameters should not be cited as optimal or used in comparisons of network simulators.

The ability of this ANN to predict those likely to contract nosocomial infections is 100% accurate on the basic features given. If several attributes are included for characterization as input variables and/or numerous classes are included as output variables, it would be necessary to include a big amount of patterns. If after testing the network for its successes and failures with each of the training and validation patterns, a high percentage of successes is not obtained, then it could be necessary to add one or several layers of hidden neurons. Additionally, other kind of backpropagation learning functions could be used in JavaNNS, such as the Quickprop or the Resilient propagation algorithms.

The distribution of risk factors for nosocomial infections differs broadly according to hospital and time. Because of these variations, data from different regions may be of a diverse nature with differences in the incidence of nosocomial infections. Failure to adjust adequately for the particular region and time could lead to false conclusions. It is recommended to use multivariate analyses retrospectively to reveal the risk factors associated with the outcome under study. Float variables can be used also with JavaNNS. In fact, using continuous input data would reduce the number of input neurons and would simplify network architecture. It is recommended to use as few variables as possible.(88. Pearl A, Caspi R, Bar-Or D. Artificial neural network versus subjective scoring in predicting mortality in trauma patients. Stud Health Technol Inform. 2006;124:1019-24.)

In this study, nosocomial infections were not classified by infection site. This research could inform researchers to be able to expand this project, and the spreadsheets could be prepared by infection site. Once the model is set within a context of real data, it should be evaluated with, at least, a validation set of patterns of real cases. This will need to be validated and tested with a sufficient number of actual patient patterns, relying on large volumes of data. In this medical context, the datasets should have hundreds or thousands of pattern samples. In the case of a very reduced dataset, a cross-validation scheme should be used for network validation. Moreover, other types of predictive methods could be tested to compare the results to the ANN. The only study found in the literature applied to a similar clinical setting, addressing some of these limitations, is the study by Chang et al.(2929. Chang YJ, Yeh ML, Li YC, Hsu CY, Lin CC, Hsu MS, et al. Predicting hospital-acquired infections by scoring system with simple parameters. PLoS One. 2011;6(8):e23137.)

CONCLUSION

This article proposes a preliminary method for predicting risk of nosocomial infection at intensive care units, adopting artificial intelligence as artificial neural networks to achieve improvements. This model is used to show its design and step-by-step analyses, using a synthetic dataset derived from the Japanese Nosocomial Infection Surveillance system. While this model is still based on a synthetic dataset, the excellent performance observed with a small number of patterns suggests using higher numbers of variables and network layers, to analyze larger volumes of data, can create powerful artificial neural networks, potentially capable of precisely anticipating nosocomial infection at intensive care units.

REFERENCES

  • 1
    Suka M, Yoshida K, Takezawa J. A practical tool to assess the incidence of nosocomial infection in Japanese intensive care units: the Japanese Nosocomial Infection Surveillance System. J Hosp Infect. 2006;63(2):179-84.
  • 2
    Stiller A, Schröder C, Gropmann A, Schwab F, Behnke M, Geffers C, et al. ICU ward design and nosocomial infection rates: a cross-sectional study in Germany. J Hosp Infect. 2017;95(1):71-5.
  • 3
    Zhu S, Kang Y, Wang W, Cai L, Sun X, Zong Z. The clinical impacts and risk factors for non-central line-associated bloodstream infection in 5046 intensive care unit patients: an observational study based on electronic medical records. Crit Care. 2019;23(1):52.
  • 4
    Merzougui L, Barhoumi T, Guizani T, Barhoumi H, Hannachi H, Turki E, et al. [Nosocomial infections in the Intensive Care Unit: annual incidence rate and clinical aspects]. Pan Afr Med J. 2018;30:143. French.
  • 5
    Januel JM, Harbarth S, Allard R, Voirin N, Lepape A, Allaouchiche B, et al. Estimating attributable mortality due to nosocomial infections acquired in intensive care units. Infect Control Hosp Epidemiol. 2010;31(4):388-94.
  • 6
    Bryan CS, Reynolds KL. Hospital acquired bacteremic urinary tract infection: epidemiology and outcome. J Urol. 1984;132(3):494-8.
  • 7
    Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients' hospitalization period. Stud Health Technol Inform. 2009;150:610-4.
  • 8
    Pearl A, Caspi R, Bar-Or D. Artificial neural network versus subjective scoring in predicting mortality in trauma patients. Stud Health Technol Inform. 2006;124:1019-24.
  • 9
    Nimgaonkar A, Karnad DR, Sudarshan S, Ohno-Machado L, Kohane I. Prediction of mortality in an Indian intensive care unit. Comparison between APACHE II and artificial neural networks. Intensive Care Med. 2004;30(2):248-53.
  • 10
    Pearl A, Bar-Or D. Decision support in trauma management: predicting potential cases of ventilator associated pneumonia. Stud Health Technol Inform. 2012;180:305-9.
  • 11
    Hasan MK, Haque MM, Adib R, Tumpa JF, Begum A, Love RR, et al. SmartHeLP: Smartphone-based hemoglobin level prediction using an artificial neural network. AMIA Annu Symp Proc. 2018;2018:535-44.
  • 12
    Taylor D, Harrison A, Powers D. An artificial neural network system to identify alleles in reference electropherograms. Forensic Sci Int Genet. 2017;30:114-26.
  • 13
    He H, Yang X, Xu Z, Deng N, Shang Y, Liu G, et al. Implementing artificial neural networks through bionic construction. PLoS One. 2019;14(2):e0212368.
  • 14
    Li P, Kong D, Tang T, Su D, Yang P, Wang H, et al. Orthodontic treatment planning based on artificial neural networks. Sci Rep. 2019;9(1):2037.
  • 15
    Zhou YP, Li S, Zhang XX, Zhang ZD, Gao YX, Ding L, et al. [High definition MRI rectal lymph node aided diagnostic system based on deep neural network]. Zhonghua Wai Ke Za Zhi. 2019;57(2):108-13. Chinese.
  • 16
    LaFaro RJ, Pothula S, Kubal KP, Inchiosa ME, Pothula VM, Yuan SC, et al. Neural network prediction of ICU length of stay following cardiac surgery based on pre-incision variables. PLoS One. 2015;10(12):e0145395.
  • 17
    Hsieh MH, Hsieh MJ, Chen CM, Hsieh CC, Chao CM, Lai CC. An Artificial neural network model for predicting successful extubation in intensive care units. J Clin Med. 2018;7(9):240.
  • 18
    Burke HB, Rosen DB, Goodman PH. Comparing artificial neural networks to other statistical methods for medical outcome prediction. Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94). 1994; p. 2213-16.
  • 19
    Eftekhar B, Mohammad K, Ardebili HE, Ghodsi M, Ketabchi E. Comparison of artificial neural network and logistic regression models for prediction of mortality in head trauma based on initial clinical data. BMC Med Inform Decis Mak. 2005;5:3.
  • 20
    DiRusso SM, Chahine AA, Sullivan T, Risucci D, Nealon P, Cuff S, et al. Development of a model for prediction of survival in pediatric trauma patients: comparison of artificial neural networks and logistic regression. J Pediatr Surg. 2002;37(7):1098-104.
  • 21
    Erguzel TT, Noyan CO, Eryilmaz G, Ünsalver BÖ, Cebi M, Tas C, et al. Binomial logistic regression and artificial neural network methods to classify opioid-dependent subjects and control group using quantitative EEG power measures. Clin EEG Neurosci. 2019;50(5):303-10.
  • 22
    Liew PL, Lee YC, Lin YC, Lee TS, Lee WJ, Wang W, et al. Comparison of artificial neural networks with logistic regression in prediction of gallbladder disease among obese patients. Dig Liver Dis. 2007;39(4):356-62.
  • 23
    Ge W, Huh JW, Park YR, Lee JH, Kim YH, Turchin A. An interpretable ICU mortality prediction model based on logistic regression and recurrent neural networks with LSTM units. AMIA Annu Symp Proc. 2018;2018:460-9.
  • 24
    Stuttgart Neural Network Simulator (SNNS). User Manual, Version 4.2 [Internet]. University of Stuttgart; 1995 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
    » http://www.ra.cs.uni-tuebingen.de/downloads/SNNS/SNNSv4.2.Manual.pdf
  • 25
    Barraclough H, Simms L, Govindan R. Biostatistics primer: what a clinician ought to know: hazard ratios. J Thorac Oncol. 2011;6(6):978-82. Erratum in: J Thorac Oncol. 2011;6(8):1454.
  • 26
    Fischer I, Hennecke F, Bannes C, Zell A. JavaNNS: Java Neural Network Simulator. User Manual, Version 1.1 [Internet]. University of Tübingen; 2002 [cited 2020 Jan 20]. Available from: http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
    » http://www.ra.cs.uni-tuebingen.de/software/JavaNNS/manual/JavaNNS-manual.pdf
  • 27
    Nistal-Nuño B. Replication data for: a neural network for prediction of risk of nosocomial infection at intensive care units, V1 [Internet]. Harvard Dataverse; 2020 [cited Oct 15]. Available from: https://doi.org/10.7910/DVN/BET7VI
    » https://doi.org/10.7910/DVN/BET7VI
  • 28
    Raudys SJ, Jain AK. Small sample size effects in statistical pattern recognition: recommendations for practitioners. IEEE Trans Pattern Anal Mach Intell. 1991;13(3):252-64.
  • 29
    Chang YJ, Yeh ML, Li YC, Hsu CY, Lin CC, Hsu MS, et al. Predicting hospital-acquired infections by scoring system with simple parameters. PLoS One. 2011;6(8):e23137.

Publication Dates

  • Publication in this collection
    20 Nov 2020
  • Date of issue
    2020

History

  • Received
    01 Nov 2019
  • Accepted
    24 June 2020
Instituto Israelita de Ensino e Pesquisa Albert Einstein Avenida Albert Einstein, 627/701 , 05651-901 São Paulo - SP, Tel.: (55 11) 2151 0904 - São Paulo - SP - Brazil
E-mail: revista@einstein.br