 Research
 Open Access
 Published:
Artificial neural network  an effective tool for predicting the lupus nephritis outcome
BMC Nephrology volume 23, Article number: 381 (2022)
Abstract
Background
Lupus nephropathy (LN) occurs in approximately 50% of patients with systemic lupus erythematosus (SLE), and 20% of them will eventually progress into endstage renal disease (ESRD). A clinical tool predicting remission of proteinuria might be of utmost importance. In our work, we focused on predicting the chance of complete remission achievement in LN patients, using artificial intelligence models, especially an artificial neural network, called the multilayer perceptron.
Methods
It was a single centre retrospective study, including 58 individuals, with diagnosed systemic lupus erythematous and biopsy proven lupus nephritis. Patients were assigned into the study cohort, between 1st January 2010 and 31st December 2020, and eventually randomly allocated either to the training set (N = 46) or testing set (N = 12). The end point was remission achievement. We have selected an array of variables, subsequently reduced to the optimal minimum set, providing the best performance.
Results
We have obtained satisfactory results creating predictive models allowing to assess, with accuracy of 91.67%, a chance of achieving a complete remission, with a high discriminant ability (AUROC 0.9375).
Conclusion
Our solution allows an accurate assessment of complete remission achievement and monitoring of patients from the group with a lower probability of complete remission. The obtained models are scalable and can be improved by introducing new patient records.
Background
Among patients suffering from systemic lupus erythematous (SLE), almost all of them have, to some extent, a renal affection during the disease course, and between 40% and 70% will develop clinically diagnosed renal involvement named lupus nephritis (LN) [1]. It is a major risk factor of morbidity and mortality in SLE, and 10% of patients with LN will eventually develop endstage renal disease (ESRD), within 5 years of disease onset [2].
Renal biopsy is the gold standard for LN diagnosis. Based on kidney biopsy assessment, a patient can be classified into any of six histological categories, according to the International Society of Nephrology/Renal Pathology Society classification, of which classes III–VI are associated with the highest risk of longterm damage[3]. Class VI reflects the most advanced stage, where patients require any type of renal replacement therapy, including kidney dialysis or transplantation [3]. Subsequential treatment decisions are based on glomerular involvement. Unfortunately, current standards for diagnosis and treatment of LN are unsatisfactory and it is neither possible to accurately predict a response to therapy nor the longterm outcome for individual patients [4]. Therefore, there is a need for establishing of predictive models allowing estimation of longterm results. Currently available studies provide several both clinical and histopathological factors, related to unsatisfactory results. Among them, the most crucial predictors of poor outcome are male gender, younger age, hypertension, increased serum creatinine, African American race, proliferative disease, high activity and chronicity index, glomerulosclerosis and crescents, interstitial inflammation, tubular injury, and an extent of interstitial fibrosis [5]. Achievement of a proteinuria < 0.7 g/day at month 12, best predicts good outcome at 7 years and inclusion of haematuria at month 12 undermines the sensitivity of early proteinuria decrease for the prediction of good outcome [6].
Based on the clinical data derived from patients with diagnosed LN and using artificial intelligence techniques, and artificial neural networks, we have built a machine learning model allowing prediction of complete remission in a patient with LN.
Methods
Data collection
It was a single centre trial, including retrospective data of 58 patients with diagnosed systemic lupus erythematosus and biopsyproven LN. The SLE diagnosis was based on EULAR/ACR classification criteria [7]. The following clinical parameters were included: age, gender, serum creatinine concentration, estimated glomerular filtration rate (eGFR) calculated by MDRD equation, C3 and C4 concentrations, serum albumin, extent of proteinuria measured as urine protein to creatinine ratio (UPCR), erythrocytes sedimentation rate (ERS), Creactive protein (CRP) concentration, erythrocyturia assessed as number of red blood cells (RBC) on highpower field (HPF),
All parameters were collected at the time of kidney biopsy. Only patients with significant proteinuria (assessed as UPCR > 1.0 mg/mg) were included into study group. After 6 months of followup, a complete remission (CR) of LN was defined as UPCR < 0.5 and stable renal function, according KDIGO guidelines [8]. All patients were treated according to EUROLUPUS regimen, using 6 intravenous pulses cyclophosphamide (500 mg each), followed by oral mycophenolate mofetil, unless contraindicated [9].
Statistical scoring
The performance of the artificial neural network models was assessed with the following statistical indicators: area under the receiveroperator curve (AUROC), Accuracy, Precision, Recall and F1Score. AUROC was used to assess the discriminant power of the artificial neural network.
Artificial neural network
The entire project was created and run in the python 3.6.8 environment. Incomplete rows, containing blank cells, were removed from the original database, allowing reduction of the amount of available data, but got 100% complete dataset. In our previous work we analysed mostly random forest classifiers, due to their better performance against neural networks [10]. An artificial neural network is a complex structure consisting of several basic units, called artificial neurons. In its simplest form, there are perceptrons containing several inputs, with assigned weights and one output. Functions responsible for building a multilayer perceptron came from the scikitlearn library. It is, to some extent, analogous to a biological neuron with many dendrites but only one axon. The interior of the perceptron is an activating function, superimposed on the sum of the products of the neuron’s inputs and the corresponding weights. The bias vector affects performance and results in better fitting to the data. Neurons are arranged in layers that are interconnected. In a multilayer perceptron, these layers are organized in the input layer, hidden layers, and output neurons. Depending on the number of neurons and layers, different complexity may be obtained. Naturally, the greater the complication, there more of the possibilities of such network, but at the same time, the more time cost needed to train it.
The activation function is analogous to the excitability threshold of a biological neuron. In MLP, this is a ReLu function that returns zero for all nonpositive values and takes the input value for positive values.
The activation function for the output in MLP is the logistic function, given by the following formula:
The complexity of the MLP neural network is related to the number of samples in the training set, the number of input features, predicted classes, and neurons in the respective layers. In mathematical notation it is written as O (n·m·o·h1·h2), where “n” is the number of samples in the training set, “m” is the number of input features, and “o” is the number of predicted classes. The sizes of the hidden layers are h1 and h2, respectively, and they denote the number of iterations leading to the best model.
The completed database has been recursively split into subsets per column. For example, the subsets contained data for all patients, but only for selected columns. The selection of input parameters was based on recursive searching of the subset space, individual evaluation of each statement, selection of hyperparameters and evaluation on the test set. Initially, we thought about applying heuristics to optimize models, but with a cutoff size of 1 to 45 neurons in the hidden layer, we did not experience an appreciable loss of resources, using brute force search. Naturally, we are aware that heuristics in model optimization are necessary in more advanced models and for larger input data. The search for optimization solutions for modelling in medicine can be an interesting subject of research and bring enormous progress in the field of personalized medicine. The main hyperparameters of the neural network are the number of neurons in the individual hidden layers. Due to the speed of calculations and their parallelism, we used a forloop nested in the forloop and limited the maximum number of neurons to 150 in a single layer. We are aware that the complexity was high, but in practice we were able to trace how the performance of the network changes depending on its structure, which, however, is not the subject of this work, but is discussed in another of our work [10]. The performance measured by AUROC, and Accuracy has been saved and finally the best configurations was chosen, allowing the most accurate prediction of total remission.
Results
Study population baseline characteristics
Retrospective data of 58 patients with biopsy proven LN, aged 18–72 years (36.05 ± 13.98), 48 women and 10 men, were included. All evaluated parameters and variables are presented in Table 1.
The input database was randomly divided into training and testing cohorts. The characteristics of the divided groups are described in Table 2.
A multilayer perceptron with 40 neurons in the first hidden layer and 45 neurons in the second hidden layer, appeared to be the model with the best performance with AUROC of 0.9375 (0.94), Accuracy of 91.67%, Positive Predictive Value (precision) of 0.9333 and Sensitivity (recall) of 0.9167 (Fig. 1). A similar result was achieved by 2 models built with 8 in the first and 22 in the second layer, and 30 in the first and 41 in the second hidden layer, respectively, but this model turned out to have a lower AUROC of 0.9067.
The best model of artificial neural network achieved 100% precision, for predicting the occurrence of complete remission, in LN from the input variables. Sensitivity 0.88 for a class with complete remission. For the group without complete remission, it achieved 100% sensitivity and 80% positive predictive ability.
The search for the best solution required construction of several models. We made the original assumption about the maximum size of the neural network up to 45 neurons in each of the two layers. In case of failure or unsatisfactory results, we would consider increasing this limit. The obtained result is within the initially assumed limits, i.e., has a relatively low complexity and a superior performance, so it has been considered as an optimal solution combining costs with efficiency.
Figure 2 shows the Accuracy distribution, depending on the number of neurons in the first and second hidden layers. The number of neurons in the first hidden layer is marked on the horizontal axis, whereas the number of neurons in the second hidden layer on the vertical axis. The colour corresponds to an Accuracy value, in the range from 0.3333 to 0.9167, from the worst to the best model constructed. The observation allows to indicate the area where the models were useless and, in the future, it may be possible to construct a metaheuristic, avoiding ineffective solutions and shorten the time of model exploration. The optimal result is a model combining all the parameters as high as possible, considering the costs of its construction and practical application.
Figure 3 shows the distribution of AUROC, depending on the number of neurons in the first and second hidden layers, with the axes labelled like at Fig. 2. The colour scale starts from 0.500, which is a typical value for a random classifier. The graphic shows an edge area where one layer of the neural network has several neurons and is unable to achieve satisfactory performance regardless of calibrating other hyperparameters or modifying the input variables. Some of the models had AUROC 1.0000, while they had accuracy lower than 0.9. The optimal solution should have both great accuracy and very discriminant power.
Figure 4 shows the precision distribution, depending on the number of neurons, in the corresponding hidden layers. Big data analysis, in combination with a recursive algorithm, allowed to generate various models and select those with higher sensitivity, in relation to the selected weighted average sensitivity target.
Figure 5 shows the Recall distribution, depending on the number of neurons in the corresponding hidden layers. The simplest models, located at the edge of the chart, do not have the worst recall. Due to the slight unbalance of the data set, the average results are recalled around 0.65. The worst outcomes overall and the best ones are scattered inside the graph, showing the complex structure of neural network models.
Graphing a neural network, with significant numerical values, may be difficult due to the complexity of the model. Figure 6. shows the matrices, with the values of individual connections between the relevant neurons in specific layers. Our network has the following layers: an input layer with 8 neurons corresponding to specific variables. The first hidden layer consists of 40 neurons. Each of them is connected to the input layer neurons, and the weights of these connections are shown in the upper 8 × 40 matrix in Fig. 6. The second hidden layer consists of 45 neurons, each connected to each of the 40 first hidden layer neurons. The weights of these connections are illustrated by the largest matrix of size 40 × 45 in Fig. 6. The output neuron is connected to each of the 45 neurons of the second hidden layer, and the weights of these connections are shown in the matrix in the right part of the graphic with the size 1 × 45. Due to the transparency of the graphics, we omitted the representation of the socalled vector bias, which are an important element of the network, improving its performance, but we focused on conveying the basic principle of MLP neural network operation.
Discussion
The input parameters of all neural networks included ERS, CRP, concentrations of serum albumin and triglycerides, complement C3 and C4 levels, presence of ANA, UPCR and data derived from the histopathological examination. Their significance in the assessment of LN progression stay in accordance with the results of studies carried out with implementation of classical statistical analysis. The variables, selected by the computer program, correspond with the conclusions of the research regarding the relationship of individual variables with the severity of the disease.
Simple designs may also achieve a great performance. Liu et al. [11] presented the model, based on UPCR, reaching AUC 0.778, and established with implementation of serum albumin with AUC 0.773. The differences in UPCR and serum albumin were assessed after 3 months followup. The cutoff points for change of UPCR and serum albumin concentration were for UPCR ≥ 59%, and for serum albumin ≥ 32.9 g/l, respectively and allowed to predict remission of LN, at sixth month followup. The level of C3 complement component, at the time of followup, allowed the prediction of LN remission, with an AUC of 0.701. Similar parameters were demonstrated in our study as reliable markers in prediction of LN remission. Chen et al. [12] obtained a design with AUC 0.819, in the validation cohort, using 59 input variables. Most of them were assessed at the point of remission. The simplified Cox risk score model implemented 6 variables, derived from initial features set, and subsequently employed to assess the risk of renal flare with AUC of 0.746. Tang et al. [13] investigated clinical indices with respect to machine learning techniques and achieved an accuracy of 40.1–56.2% in depending on the predicted LN class.
Adamichou et al. developed a more complex model, capable of recognizing LN with accuracy of 97.9% [14]. In our work, we tried to avoid too obvious variables, directly leading to a given result, so we avoided differentiating the healthy versus sick ones as a trivial issue. A comparative solution, with a list of several machine learning techniques, was presented by Helget et al. [15], with results of AUC 0.800 for Random Forest Classifier, using 4 variables: chronicity score, intestinal inflammation, UPCR and WBC. An Artificial neural network design based on activity score, chronicity score, intestinal fibrosis, intestinal inflammation and UPCR, achieved AUC of 0.775.
Regarding renal histopathology, as a crucial factor for the clinical management and outcome of patients with LN, it is worth to mention that deep learningbased AI procedure was also tested for automatic assessment of glomerular pathological findings in LN [16]. The main motive for the development of such an arrangement was an unsatisfactory interpathologist agreement. Deep convolutional neural networkbased system detected and classified glomerular pathological findings in LN (dataset of 349 renal biopsy wholeslide images). Authors suggested that deep learning is a feasible assistive tool for the objective and automatic assessment of pathological LN lesions: at the perpatient kidney level, the model achieved a high agreement with nephropathologist (linear weighted kappa: 0.855, 95% CI; quadratic weighted kappa: 0.906, 95% CI).
One of the most serious limitations of our study was the small size of the examined population. This was a was singlecentre study, conducted on an ethnically homogeneous population. The obtained models are scalable, and, in the future, we hope to test them on a larger group. A particular advantage is the use of neural networks that may be retrained on a smaller group of samples, called partial fitting. Machine learning is not a technique, which may be comparable between centres, as are the classic analysis, based on odds ratio and survival models. Despite the insight into the mechanism of operation, we were not able to draw greater conclusions without an indepth mathematical and computer analysis of the algorithm, requiring knowledge and experience in computer science. The MLP neural network, on the other hand, is a practical tool that may be used in clinical practice after appropriate calibration for the population.
Conclusion
The use of an artificial neural network, learned even on a small patient cohort, allows the construction of a predictive model, with good or very good performance. A huge advantage is the ability to scale models to larger and more diverse populations and overwrite the values stored in the network structure with partial fitting. We emphasize the possibility of using this solution in a pilot program after conducting further observations on a larger research group.
Availability of data and materials
The datasets used and/or analysed during the current study are available from the corresponding author on request.
Abbreviations
 ALB:

serum albumin
 ANA:

antinuclear antibodies,
 C3:

complement component 3
 C4:

complement component 4
 CRP:

Creactive protein
 eGFR:

estimated glomerular filtration rate
 ERS:

erythrocytes sedimentation rate
 HGB:

haemoglobin
 LN:

lupus nephritis
 LYM:

lymphocytes
 MLP:

multilayer perceptron
 NEU:

neutrophils
 NLR:

neutrophiltolymphocyte ratio
 PLR:

platelets to lymphocyte ratio
 PLT:

platelets
 sCr:

serum creatinine concentration.
 SLE:

systemic lupus erythematosus.
 TP:

total protein concentration.
 UPCR:

urine protein to creatinine ratio.
 WBC:

white blood cells.
References
Bastian HM, Roseman JM, McGwin G Jr, Alarcón GS, Friedman AW, Fessler BJ, Baethge BA, Reveille JD. Systemic lupus erythematosus in three ethnic groups. XII. Risk factors for lupus nephritis after diagnosis. Lupus. 2002;11(3):152–60.
Alarcón GS. Multiethnic lupus cohorts: what have they taught us? Reumatol Clin. 2011;7(1):3–6.
Mahajan A, Amelio J, Gairy K, Kaur G, Levy RA, Roth D, Bass D. Systemic lupus erythematosus, lupus nephritis and endstage renal disease: a pragmatic review mapping disease severity and progression. Lupus. 2020;29(9):1011–20.
Davidson A, Aranow C, Mackay M. Lupus nephritis: challenges and progress. Curr Opin Rheumatol. 2019;31(6):682–8.
Mackay M, Dall’Era M, Fishbein J, Kalunian K, Lesser M, SanchezGuerrero J, Levy DM, Silverman E, Petri M, Arriens C, et al. Establishing surrogate kidney end points for lupus nephritis clinical trials: development and validation of a novel approach to predict future kidney outcomes. Arthritis Rheumatol. 2019;71(3):411–9.
Tamirou F, Lauwerys BR, Dall’Era M, Mackay M, Rovin B, Cervera R, Houssiau FA. A proteinuria cutoff level of 0.7† g/day after 12†months of treatment best predicts longterm renal outcome in lupus nephritis: data from the MAINTAIN Nephritis Trial. Lupus Sci Med. 2015;2(1):e000123.
Aringer M, Costenbader K, Daikh D, Brinks R, Mosca M, RamseyGoldman R, Smolen JS, Wofsy D, Boumpas DT, Kamen DL, et al. 2019 European league against rheumatism/american college of rheumatology classification criteria for systemic lupus erythematosus. Arthritis Rheumatol. 2019;71(9):1400–12.
Rovin BH, Adler SG, Barratt J, Bridoux F, Burdge KA, Chan TM, Cook HT, Fervenza FC, Gibson KL, Glassock RJ, et al. Executive summary of the KDIGO 2021 guideline for the management of glomerular diseases. Kidney Int. 2021;100(4):753–79.
Houssiau FA, Vasconcelos C, D’Cruz D, Sebastiani GD, Garrido Ed Ede R, Danieli MG, Abramovicz D, Blockmans D, Mathieu A, Direskeneli H, et al. Immunosuppressive therapy in lupus nephritis: the EuroLupus Nephritis Trial, a randomized trial of lowdose versus highdose intravenous cyclophosphamide. Arthritis Rheum. 2002;46(8):2121–31.
Konieczny A, Stojanowski J, Rydzynska K, Kusztal M, Krajewska M. Artificial intelligencea tool for risk assessment of delayedgraft function in kidney transplant. J Clin Med 2021, 10(22).
Liu G, Wang H, Le J, Lan L, Xu Y, Yang Y, Chen J, Han F. Earlystage predictors for treatment responses in patients with active lupus nephritis. Lupus. 2019;28(3):283–9.
Chen Y, Huang S, Chen T, Liang D, Yang J, Zeng C, Li X, Xie G, Liu Z. Machine learning for prediction and risk stratification of lupus nephritis renal flare. Am J Nephrol. 2021;52(2):152–60.
Tang Y, Zhang W, Zhu M, Zheng L, Xie L, Yao Z, Zhang H, Cao D, Lu B. Lupus nephritis pathology prediction with clinical indices. Sci Rep. 2018;8(1):10231.
Adamichou C, Genitsaridi I, Nikolopoulos D, Nikoloudaki M, Repa A, Bortoluzzi A, Fanouriakis A, Sidiropoulos P, Boumpas DT, Bertsias GK. Lupus or not? SLE risk probability index (SLERPI): a simple, clinicianfriendly machine learningbased model to assist the diagnosis of systemic lupus erythematosus. Ann Rheum Dis. 2021;80(6):758–66.
Helget LN, Dillon DJ, Wolf B, Parks LP, Self SE, Bruner ET, Oates EE, Oates JC. Development of a lupus nephritis suboptimal response prediction tool using renal histopathological and clinical laboratory variables at the time of diagnosis. Lupus Sci Med 2021, 8(1).
Zheng Z, Zhang X, Ding J, Zhang D, Cui J, Fu X, Han J, Zhu P. Deep learningbased artificial intelligence system for automatic assessment of glomerular pathological findings in lupus nephritis. Diagnostics (Basel) 2021, 11(11).
Acknowledgements
Not applicable.
Funding
The study is supported by the Wroclaw Medical University statutory funds (SUB.C160.21.016). It was investigatorinitiated research. The funding body had no role in the study design, data collection, analyses, and interpretation, or in writing the manuscript.
Author information
Authors and Affiliations
Contributions
Conceptualization, A.K., J.S., I.K., A.M., T.G., M.K. and M.Ku. (Mariusz Kusztal); methodology, A.K., J.S. and M.Ku.; software, J.S.; validation, J.S.; formal analysis, J.S., A.K.; resources, A.K., I.K., AM; writing—original draft preparation, A.K., J.S., K.R. and M.Ku.; writing review and editing, A.K., J.S., K.R., T.G., M.Ku. and M.K.; supervision, T.G., M.K., M.Ku.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of Wroclaw Medical University No KB609/2019. An informed consent has been obtained from all participants.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Stojanowski, J., Konieczny, A., Rydzyńska, K. et al. Artificial neural network  an effective tool for predicting the lupus nephritis outcome. BMC Nephrol 23, 381 (2022). https://doi.org/10.1186/s12882022029782
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12882022029782
Keywords
 Artificial intelligence
 Machine learning
 Proteinuria
 Systemic lupus erythematosus
 Lupus nephritis
 Endstage renal disease