Applsci 10 05776 With Cover

Download as pdf or txt
Download as pdf or txt
You are on page 1of 50

2.838 3.

Review

A Review of the Artificial Neural


Network Models for Water Quality
Prediction

Yingyi Chen, Lihua Song, Yeqi Liu, Ling Yang and Daoliang Li

Special Issue
Hydrologic and Water Resources Investigations and Modeling
Edited by
Dr. Nejc Bezak

https://doi.org/10.3390/app10175776
applied
sciences
Review
A Review of the Artificial Neural Network Models for
Water Quality Prediction
Yingyi Chen 1,2,3,4, * , Lihua Song 1,2,3 , Yeqi Liu 1,2,3 , Ling Yang 1,2,3 and Daoliang Li 1,2,3,4
1 Precision Agricultural Technology Integration Research Base (Fishery), Ministry of Agriculture and Rural
Affairs, China Agricultural University, Beijing 100083, China; [email protected] (L.S.);
[email protected] (Y.L.); [email protected] (L.Y.); [email protected] (D.L.)
2 College of Information and Electrical Engineering, China Agricultural University, Beijing 100083, China
3 National Innovation Center for Digital Fishery, China Agricultural University, Beijing 100083, China
4 Beijing Engineering and Technology Research Centre for Internet of Things in Agriculture,
China Agricultural University, Beijing 100083, China
* Correspondence: [email protected]; Tel.: +86-10-6273-8489

Received: 13 July 2020; Accepted: 17 August 2020; Published: 20 August 2020 

Abstract: Water quality prediction plays an important role in environmental monitoring, ecosystem
sustainability, and aquaculture. Traditional prediction methods cannot capture the nonlinear and
non-stationarity of water quality well. In recent years, the rapid development of artificial neural
networks (ANNs) has made them a hotspot in water quality prediction. We have conducted
extensive investigation and analysis on ANN-based water quality prediction from three aspects,
namely feedforward, recurrent, and hybrid architectures. Based on 151 papers published from 2008 to
2019, 23 types of water quality variables were highlighted. The variables were primarily collected by the
sensor, followed by specialist experimental equipment, such as a UV-visible photometer. Five different
output strategies, namely Univariate-Input-Itself-Output, Univariate-Input-Other-Output,
Multivariate-Input-Other(multi)-output, Multivariate-Input-Itself-Other-Output, and Multivariate-
Input-Itself-Other (multi)-Output, are summarized. From results of the review, it can be concluded
that the ANN models are capable of dealing with different modeling problems in rivers, lakes,
reservoirs, wastewater treatment plants (WWTPs), groundwater, ponds, and streams. The results of
many of the review articles are useful to researchers in prediction and similar fields. Several new
architectures presented in the study, such as recurrent and hybrid structures, are able to improve the
modeling quality of future development.

Keywords: ANNs; feedforward; recurrent; hybrid; water quality prediction

1. Introduction
Water quality plays an important role in any aquatic system, e.g., it can influence the growth of
aquatic organisms and reflect the degree of water pollution [1]. Water quality prediction is one of the
purposes of model development and use [2], which aims to achieve appropriate management over
a period of time [3]. Water quality prediction is to forecast the variation trend of water quality at a
certain time in the future [4]. Accurate water quality prediction plays a crucial role in environmental
monitoring, ecosystem sustainability, and human health. Moreover, predicting future changes in
water quality is a prerequisite for early control of intelligence aquaculture in the future [5]. Therefore,
water quality prediction has great practical significance [6].
At present, there are many traditional water quality prediction methods, such as multiple linear
regression (MLR) [7], auto-regressive integrated moving average (ARIMA) [8], etc. MLR is not able to
detect a nonlinear relationship between water quality parameters because of its linear inherence [9].

Appl. Sci. 2020, 10, 5776; doi:10.3390/app10175776 www.mdpi.com/journal/applsci


Appl. Sci. 2020, 10, 5776 2 of 49

The main drawback of ARIMA is the pre-assumption of the linear model [10]. During the model
identification phase, the time series data must be checked to see whether they are stationary or
not, because it is critical in creating the ARIMA model. In fact, traditional methods are not able to
capture the non-linear [11] and non-stationarity [12] of water quality well due to their complex and
sophisticated nature.
With the increase in data scale, traditional techniques cannot meet the demand of researchers.
Owing to the improvement of computing power, artificial neural network (ANN) models,
data-driven models, have been further developed. They can capture functional relationships among
the water quality data from the examples [13]. When the underlying relationships of obtained data are
difficult to describe, ANN models still work. Moreover, ANNs require fewer prior assumptions [14]
and can achieve higher accuracy [15] compared with traditional approaches. In addition, ANNs are
suitable for solving the non-linear and uncertain problems due to their similar characteristics with the
brain nervous system [4], and have become a hotspot in water quality research [16].
ANNs are a family of models inspired by biological neural networks [17] which specifically
refers to the human brain [18], a kind of central nervous system of animals. In general, ANN can be
represented as a system of interconnected “neurons” [19] which form the basis of neural network
operation. Weight parameters and activation functions are part of the neurons [20]. ANNs are
generally divided into three layers of input, hidden and output. When neurons receive information
from different inputs, they obtain nonlinearity through activation functions. ANN models depend
heavily on the quantity of data [21]. Therefore, it is not recommended to use relatively small data
sizes for predictors (inputs). This is because some useful information is lost in short-term data,
which may lead to poor prediction results [3]. In addition, data dividing is a necessary step in the
modeling process. Furthermore, choosing the training algorithm to calibrate the model parameters
(e.g., connection weights) is a vital step so that the network can approximate complicated non-linear
input-output relationship [10]. The Levenberg–Marquardt [22] algorithm and the back-propagation
(BP) algorithm [23] are the most commonly used algorithms.
ANN models architectures determine the number of connection weights and the way information
flows through the network [20]. The most widely used architecture is Multilayer Perceptron (MLPs)
with only three layers in many types of feedforward ANNs. Radial Basis Function neural networks
(RBFNNs) [24], General regression neural networks (GRNNs) [25] and Extreme learning machines
(ELMs) [5] are three typical feedforward ANNs. A Long Short-Term Memory (LSTM) neural network
is an improvement of recurrent neural networks (RNNs), which aims to address the well-known
vanishing gradient problem [26]. The hybrid models in this review are three classes: model-intensive,
technique-intensive, and data-intensive [27]. The emerging frameworks, such as Convolutional Neural
Network (CNN) [28], widely used in the field of the image, are also included in this review.
In this review, ANN models for water quality variables prediction are summarized. Previous
reviews [20,27,29] about ANNs are more concerned about the water quantity (e.g., flow and
rainfall-runoff) prediction, while less attention has been paid to water quality prediction
(e.g., Suspended solids (SS)), and the major scenarios they investigated are river systems. At the
same time, previous reviews care about the development of the model while ignoring the output
strategies between input(s) and output(s) in a given prediction task. To overcome the limitations
above, this review focuses on the use of ANNs methods for water quality prediction, with more water
quality variables investigated than previous reviews, which are mainly divided into three categories,
namely chemical, biological and physical variables [30].
The research scenarios include not only the river system that was the focus of the previous review,
but also reservoirs, lakes, wastewater treatment plant (WWTP), groundwater, etc. It must be pointed
out that the review did not consider drinking water systems. The reason for this is that drinking
water is a system that includes source, treatment, and distribution, and should be considered as an
independent branch or subject for systematic research [30]. In addition to the increased number of
water quality variables reviewed and broader research scenarios, this review also summarizes five
Appl. Sci. 2020, 10, 5776 3 of 49

output strategies. The period of the investigated papers covered was from 2008 to 2019. This period
was chosen as it follows on from the period covered in the review by [27] (i.e., 1999–2007). The review
is organized as follows. Section 2 presents the process of the paper collection. Section 3 describes three
basic model structures in water quality prediction. In Section 4, the applications of artificial neural
networks in water quality are surveyed. Then, Section 5 represents the results of this review. Finally,
the discussions are given in Section 6. All the abbreviations are mentioned in Table 1.

Table 1. The abbreviations in this review.

Abbreviations Full Name Abbreviations Full Name Abbreviations Full Name Abbreviations Full Name
Oxidation
Electrical total chromium
AH air humidity EC ORP reduction TCC
conductivity concentration
potential
August, October, total iron
AODD Evap evaporation Q discharge TIC
December, data concentration
flow travel Pondus total anions and
AP air pressure FTT pH TAC
time Hydrogenii cations
AT air temperature Fe iron Precip precipitation TNs total nutrients
As Arsenic F flow P phosphate TA total alkalinity
relative total
B boron HCO3 bicarbonate RH TP
humidity phosphorus
Biochemical
Hydrogenated Redox
BOD Oxygen HA RP Tur turbidity
Amine potential
Demand;
ionic total dissolved
C carbon ICs RO runoff TDS
concentrations solids
Cl chloride K potassium RF rainfall TN total nitrogen
Cu Copper Lon longitude RainP Rainy period TH total hardness
total organic
Ca calcium Lat latitude SR solar radiation TOC
carbon
2- sunshine time total suspended
CO3 Carbonate LV lake volume Sth TSS
hours solids
month, day,
Coli Coliform MDHM SD transparence VP volatile phenol
hour, minute
Chemical sodium
COD Oxygen Mn manganese; SAR absorption WL Water Level
Demand ratio
permanganate water
COD Mn Mg magnesium SM Soil Moisture WT
index temperature
soil
Chl-a Chlorophyll a Na sodium ST WS wind speed
temperature
dissolved
DO Ns nutrients SO4 sulphate WD wind direction
oxygen
the year
DOY day of year NO2 nitrite S salinity YMDH
numbers

2. Methods
This review focuses on the application of ANNs to water quality variables prediction excluding
drinking water from 2008 to 2019. The papers to be reviewed were selected using the following steps:

1. First, we identified ANN-related papers in influential water-related and environmental-related


journals to ensure that high-quality papers are included in the review. These papers are mainly
from journals whose subjects are environmental science and ecology, water resources, engineering
and application.
2. Thereafter, a keyword search of the ISI Web of Science was then conducted for the period
2008–2019 using the keywords; water quality, river, lake, reservoir, WWTP, groundwater, pond,
prediction, and forecasting, accompanied by the names of ANN methods (one or more), such as
neural network, MLP, RBFNN, GRNN, RNN, to name but a few.
3. Then, through the search process from 1 to 2, 151 articles in English relevant to our focus were
selected. The basic information of the papers, including authors (year), locations, water quality
variables, meteorological factors, other factors, output strategy, data size, time step, data dividing,
methods, and prediction lengths are provided in Appendix A.
Appl. Sci. 2020, 10, 5776 4 of 49

3. Three Basic Model Structures in Water Quality Prediction


In this review, the model architecture refers to the overall structure and manner of how information
flows from one layer to another. The three model architectures include feedforward, recurrent networks,
and hybrid models (see Figure 1) [31]. In addition to categorizing each architecture, Table 2 summarizes
the foundation and advantage(s) of the development model structure.

3.1. Feedforward Architectures


The term ‘feed-forward’ means that a neuron connection only exists from a neuron in the input
layer to other neurons in the hidden layer or from a neuron in the hidden layer to neurons in the output
layer. However, the neurons within a layer are not interconnected [9]. MLPs with only three layers are
the most widely used architectures [59] in many types of feedforward ANNs (see Figure 2), followed by
BPNNs [37] which use the back-propagation algorithms to train networks. Other commonly used
feed-forward network architectures in water quality prediction include TDNNs [36], RBFNNs [60],
GRNNs [61], WNNs [62], ELMs [5], CCNN [63] and MNN [50].

Figure 1. Three main model architectures in the reviewed papers.

Figure 2. The common architectures of MLPs.


Appl. Sci. 2020, 10, 5776 5 of 49

Table 2. The developments and advantages of different ANNs architectures.

Categories Structure(s) Advantage(s) Reference(s)


They are based on an understanding of the
MLPs Solving the nonlinear problems [19,23,30,32–35]
biological nervous system
Using time delay cells to deal with the
TDNNs They are based on the structure of MLPs [36]
dynamic nature of sample data
The structure of RBFNNs is similar to
the MLPs To overcome the local
RBFNNs [5,18,37,38]
The radial basis activation function is in the minimum problems
hidden layer
A modified form of the RBFNNs model
GRNNs There is a pattern and a summation layer Solving the small sample problems [24,39–43]
between the input and output layers
Wavelet function replace the linear sigmoid
WNNs Solving the non-stationary problems [16,44]
activation functions of MLPs
Reducing the computation problems
The structure of ELMs is similar to the MLPs
ELMs because the weights of the input and [31,45–48]
Only need to learn the output weight
hidden layer need not be adjusted
A constructive neural network that
aims to solve the problems of the
Start with input and output layer without a
CCNN determination of potential neurons [49]
hidden layer
which are not relevant to the
output layer
A special feedforward network Choosing the
neural network which have the maximum Solving the problem of low
MNNs [30,50]
similarity between the inputs and centroids prediction accuracy
of the cluster
Solving the problems of long-term
The RNNs are developed with the
RNNs dependence which are not captured [12,31,38,51,52]
development of deep learning
by the feedforward network
Its structure is similar to RNNs Addressing the well-known vanishing
LSTMs [15,26,45,53,54]
Memory cell state is added to hidden layer gradient problem of RNNs
Its structure is similar to MLPs Reducing the influence of the noise
TLRN It has the local recurrent connections in the and owning the advantage of [55]
hidden layer adaptive memory depth
Sub-classes of RNNs
Solving the problems of
NARX Their recurrent connections are from [12]
long-term dependence
the output
A context layer that can store the internal
It is useful in dynamic system
Elman states is added besides the traditional [3]
modeling because of the context layer
three layers
Different from the above recurrent
neural networks To overcome the problems of the local
ESN [3]
The three layers are input, reservoir, and minima and gradient vanishing
readout layer
They are based on the structure of ESN which To overcome the ill-posed problem
RESN [3]
has a large and sparsely connected reservoir existing in the ESN
The combination of conventional or
Hybrid Exploring the advantages of
preprocess methods with ANNs [56]
methods each methods
The internal integration of ANN methods or
Input, convolution, fully connection, and An emerging method to solve the
CNN [57]
output layers dissolved oxygen prediction problem
They are based on the structure of DBN Investigating the problem of
SODBN whose visible and hidden layers are dynamically determining the [58]
stacked sequentially structure of DBN
Appl. Sci. 2020, 10, 5776 6 of 49

TDNNs is a subclass of MLPs that learns temporal behavior from continuous past and present
signals [36]. The major difference between RBFNNs and MLPs is that the hidden layer of RBFNNs is
self-organizing while the latter is not, although the structure of RBFNNs is similar to MLPs. As the
center of RBF, the training weights can be defined by a clustering algorithm. For example, the k-means
algorithm is a commonly used one [24]. GRNNs is a modified form of the RBFNNs model, but it
differs from RBFNNs in structure. Patten and summation layers are located between the input and
output layers [27]. The training between the input and pattern layer of GRNNs is equivalent to the
research on the input and hidden layer of the RBFNNs. WNNs have made some changes based on
the traditional MLPs, in which the non-linear sigmoid activation functions is replaced by the Morlet
wavelet function commonly used in the WNNs hidden layer. Therefore, WNNs are suitable for solving
non-stationary time series problems [64]. The biggest innovation of ELMs is the random selection of
hidden nodes and the use of a least squares method to determine the output layer weight. CCNN is
different from the above feedforward networks because it constructs the neural network without a
hidden layer at first and automatically adds hidden units instead of fixing the network architectures
and then training the weights and thresholds. The first step of MNN, a special feedforward network,
is data clustering using the fuzzy c-means method [65]. The second step is updating the clusters by
adding the new datasets. To achieve better prediction accuracy, a neural network with the maximum
similarity between the inputs and centroids of the cluster is chosen.

3.2. Recurrent Architectures


Compared with feedforward ANNs, RNNs differs in that neurons within a layer are interconnected
and allow feedback [53]. Different types of RNNs are developed so that the neural networks have
better memory ability (see Figure 1). LSTM, an improvement over RNN, adds a processor called
“memory cell state” to its hidden layer to determine whether the information is useful or not [66],
and this is also suitable for SRU (Simple Recurrent Unit) [67]. Furthermore, the forget gate also
determines what information should be discarded from the cell state [66]. TLRN has a similar structure
to MLPs, but has local recurrent connections in the hidden layer (see Figure 3), with the advantages of
low noise sensitivity and adaptive storage depth [55]. NARX networks are also sub-classes of RNNs
and can be utilized to establish a long-term temporal relationship. The recurrent connections of NARX
networks come from the output (see Figure 3) [12]. In addition to the input, hidden, and output
layers, the Elman neural network has a context layer to store the internal states [3]. The Elman neural
network is sensitive to the historical information of inputs because of the self-connections of the context
nodes (see Figure 3). The three layers of ESN are different from the above recurrent neural networks.
The three layers are input, reservoir, and readout layer. The feature of the reservoir layer is randomly
and sparsely connected. The echo state property whose internal states are particularly dependent on
the inputs is the key to the ESN. To overcome the ill-posed problem existing in the ESN, an RESN
method using the ridge regression algorithm instead of linear regression to calculate output weights is
proposed [38].

3.3. Hybrid Architectures


There is a growing tendency to use hybrid ANNs models, which play a huge role in modeling,
for their ability to integrate with other conventional and more advanced modeling techniques [68],
to create flexible and efficient models in recent years (see Figure 1). Hybrid models are divided into three
categories, namely model-intensive, technique-intensive, and data-intensive [27]. The model-intensive
approaches model the sub-components of the whole physical system and aggregate the overall response
of each model. Relevant forms, such as LSTM-RNN [26] or FNN-WNN [69], are model-intensive
methods. The core of the technique intensive methods is to develop a modeling framework that is
able to take advantage of different technologies. Methods that combine ensemble approaches [32]
or time series models that remove trends or periodicities like Autoregressive Integrated Moving
Average-Radial Basis Function neural networks (ARIMA-RBFNNs) [70] or ARIMA-ANN [71] are
Appl. Sci. 2020, 10, 5776 7 of 49

technique-intensive methods. In this review, data-intensive approaches are to combine different


technologies to preprocess the data. Wavelet analysis approaches such as WANN [72] can provide
some useful information about the physical structure of the data. ANNs models the approximation and
details component from the discrete wavelet transformation (see Figure 4). Dimensionality reduction
methods such as PCA can reduce the dimension of the input data space to prevent redundancy [73].
Then, ANNs models some aggregative indices obtained by PCA (see Figure 4). Clustering methods [50]
such as K-means-MLP [43] identify the data belonging to a particular class. Other data-intensive
approaches include decomposition [5] and evolution-related [16] methods. ANN models the Intrinsic
Model Function (IMF) obtained from the decomposition of complicated signals.

Figure 3. Five categories of recurrent model architectures.

Figure 4. The modeling process of three data-intensive approaches.

3.4. Emerging Methods


CNN is a feed-forward neural network, primarily used in the image field. Input, convolution,
pooling, full connection, and output layers are the basic elements of the traditional CNN. In recent
years, CNN has been used as an emerging method in water quality prediction. The operation of
convolution can be implemented more than one time to reveal the relationship between the parameters
hidden in the input matrix [57]. However, since the purpose of the prediction model is to extract
potential factors rather than simply raise the convolutional layer’s results to a higher level, the pooling
layer is removed (see Figure 5). In the meantime, the number of calculations can be reduced.
Appl. Sci. 2020, 10, 5776 8 of 49

Figure 5. The architecture of a Convolutional Neural Network.

Deep belief network (DBN) is a kind of neural network based on deep learning which is similar to
feedforward structure and has been widely used in recent years. The blue virtual box in Figure 6 shows
several visible and hidden layers, stacked in order to make up the DBN [74]. However, the researches
about dynamically determining the structure are seldom investigated. To overcome the limitations
above, a SODBN has been proposed. The structure of the SODBN is not determined by artificial
experience but the automatic growing and pruning algorithm (AGP) [58]. Especially, the hidden layers
and neurons are changed by the AGP at first. Then, the weights of the SODBN are continuously
adjusted in the process of self-organization. Finally, some aspects of network performance, such as
running time and prediction accuracy have been improved.

Figure 6. The architecture of deep belief network.

4. Artificial Neural Networks Models for Water Quality Prediction


From 2008 to 2019, the use of the ANN technique has been very popular in the field of water quality
prediction. Many researchers have utilized ANNs to model and predict water quality. Dogan et al. [75]
adopt ANN to predict the BOD, which is difficult to measure and needs at least five days to get the final
results in WWTP. Results showed that COD was the most effective variables on BOD estimation after
conducting the sensitivity analysis. Elhatip and Kömür [76] revealed that ANN techniques depend
on using more input data to solve the water quality problems, although they did not illustrate the
size of the appropriate datasets. Palani et al. [40] tested MLP and GRNN models with various input
selected by stepwise constructive methods for multistep prediction of S, DO, and Chl-a. They pointed
out that the limited data set was one of the drawbacks of their research and encouraged others
to collect more data to recalibrate and revalidate the model. Wang et al. [19] employed a typical
three-layer of MLP structure [77–89] with the BP algorithm to achieve Chl-a prediction. They divided
the dataset into training (75%) and testing parts (25%). Results indicated that ANNs could establish
a stable and effective model for Chl-a prediction. This result is also suitable for other parameters
Appl. Sci. 2020, 10, 5776 9 of 49

prediction. Yeon et al. [90] evaluated ANN, MNN, and adaptive neuro-fuzzy inference system (ANFIS)
performance in 1-h and 2-h ahead prediction of DO and TOC. They added Q to inputs because rainfall
affected the water quality prediction. It was found that using the Levenberg–Marquart algorithm to
train the MNN could provide the least error and better results. Dogan et al. [91] divided the data
into training (60%), validation (20%), and testing sets (20%). They adopted a sensitivity analysis
method to find out the important water quality parameters and excluded fewer influence variables,
resulting in a compact network. Miao et al. [92] used BPNN to COD and ammonia nitrogen (NH3 -N)
prediction. The whole datasets were normalized at first and then divided into training (80%) and
testing (20%) sets. The sigmoid transfer function that can establish the random nonlinear map between
inputs and outputs were adopted. Oliveira Souza da Costa et al. [93] divided the data into training
(50%), validation (25%), and testing sets (25%). Shen et al. [94] employed a golden section method
to select the hidden layer nodes of BPNN. Singh et al. [95] investigated the partition approach in
evaluating the relative importance of eleven environmental variables to the output layer. They divided
the datasets into training (60%), validation (20%), and testing sets (20%). Results showed that the
predicted values of the ANN model were close to the measured value. Yeon et al. [96] combined Precip
and Q to realize a one-step prediction of Q. Then, the connected system utilized the prediction value
of Q and historical TOC to fulfill the one-step prediction of TOC. Finally, the connected system had
better performance than a single ANN model. Zuo and Yu [97] pointed out that ANN models could
process complex and multivariable problems. Akkoyunlu and Akiner [98] verified the feasibility of
ANN technique, data-driven models, in predicting DO. Results showed that the ANN method was
superior to the nonlinear regression (NLR) technique. Chen et al. [99] scaled the datasets to lie between
0 and 1 [9,16,59,62,100–104] so that it could be compatible with the sigmoid transfer functions used in
the hidden layer and applied the constructive and pruning of stepwise methods that aim to maximize
the model’s performance through a constant adjustment to surface water quality prediction. Markus et
al. [105] purely relied on a trial-and-error approach to determine the model structure and dividing the
data into training (50%) and testing sets (50%). Result found that ANN could improve the forecast
accuracy of NO3 compared with previous studies. Merdun and Çinar [106] preprocessed the data set
by normalization and moving average techniques. They improved the representation of the acquisition
data through a data preprocessing technique. Ranković et al. [107] used a sensitivity analysis method
to determine the influence of input variables on outputs and found out that 15 hidden neurons gave
the best choice. Zhu et al. [108] not only predicted the water quality using ANN models but also
introduced a remote wireless monitoring system. Banerjee et [109] checked that ANN models were
an accurate alternative to the numerical methods. They used quick propagation algorithm to realize
super linear convergence speed. Han et al. [110] demonstrated the effectiveness of a flexible structure
RBFNN which using neuron activity and mutual information (MI) to add or remove hidden neurons to
reduce network complexity and improve computational efficiency. The connected weights are trained
by an online learning algorithm. Zare et al. [10] used a UV-visible photometer to measure the NO3
concentration in the laboratory.
Asadollahfardi et al. [111] utilized Q to forecast TDS when TDS was not available.
Al-Mahallawi [77] revealed that the reason why ANN models could model complex water quality
phenomena was that they provided a non-linear function mapping from input to corresponding network
outputs. Ay and Kisi [112] divided the data into training (50%), validation (25%), and testing sets (25%).
In the three parts of data division, the validation set can be implemented more than once to monitor
whether the model is overfitting or not. Comparison results showed that the RBNN model performed
better than MLP in DO prediction. Baek et al. [50] chose the neural network of MNN, which has
the maximum similarity between the inputs and centroids of the cluster, to solve the problem of low
prediction accuracy. They introduced Gradient descent with momentum and Levenberg–Marquardt
backpropagation (TRAINLM) to train the neural network. Bayram et al. [79] used the one-year Tur
data whose time step is fortnightly to achieve the prediction of SS. Gazzaz et al. [113] scaled the data
into the scope between 0 and 1 and utilized cross-validation to improve the generalization ability
Appl. Sci. 2020, 10, 5776 10 of 49

and limit the overfitting problem. Cross-validation was suitable for the situation where the size of
the training data was small or the number of parameters in the model was large. Overfitting refers
to the situation that when the error on the training set is driven to a very small value, the test data
are presented to the network with a large error. That means the network has memorized the training
examples, but it has not learned to generalize to new situations. Hong [78] took the AT, AP, WD,
and WS variables measured by meteorological station into account. They divided the data samples
into training (70%) and testing (30%) sets. Results indicated that MLP also could deal with large data
samples. Liu and Chen [114] recorded the location information to complete the three-dimensional
DO prediction. Tota-Maharaj and Scholz [22] assessed the influence of bp, Levenberg–Marquardt,
Quasi-Newton, and Bayesian Regularization algorithms on BOD prediction. Results showed that the
combination of bp and ANN had low minimum statistical errors. Kakaei Lafdani et al. [115] firstly used
M-test to obtain several data points through the winGamma software. Then, the genetic algorithm
(GA) method was implemented to make the best combination which extracted from a list of possible
inputs as inputs. Karakaya et al. [116] conducted research, namely temporal partitioning, to divide
the data into diel, diurnal, and nocturnal in order to obtain continuous records, and chose MLP as
a prediction model. Antanasijević et al. [117] utilized Monte Carlo simulation (MCS), a sensitivity
analysis method that involves repeatedly generating a probability distribution of random input
values, to ultimately create an ANN model with fewer inputs. Moreover, other input selection
techniques include correlation analysis and genetic algorithm were tested. Chen and Liu [118]
utilized sigmoid and linear transfer function in the hidden and output layer, respectively. Results
showed that ANFIS and BPNN could predict DO with reasonable accuracy. Han et al. [119] adopted
linear interpolation whose data increment was calculated by the slope of the assumed line to fill the
missing data. Then, hierarchical ELM based on a hierarchical structure was chosen to model the
DO, pH, and SS. The advantage of hierarchical ELM is able to learn sequential information online.
Results demonstrated the effectiveness of the proposed methods. Researchers tended to divide the
training set data into 70% to 90% of the total data [39,42,49,52,72,120–127]. Iglesias et al. [35] divided
the data into training (90%) and testing sets (10%). Then, they applied three typical MLP architectures
to complete the Tur prediction whose inputs were NH3 -N, EC, DO, pH, and WT. Klçaslan et al. [128]
randomly divided the datasets and pointed out that when the data tended to be roughly periodic after
a year, the time length of data acquisition, covering a long period such as a year or more was highly
recommended in order to capture long-term variation. Yang et al. [129] found the most significant
parameters by using analysis of variance (ANOVA) techniques. Result indicated that rainfall records
were the most significant parameters for turbidity forecasting. Khashei-Siuki and Sarbazi [130] took
the normalization step to control the scale of each feature, in the same range in case the difference
of the order of magnitude will lead to the dominance of larger attributes thereby slowing down the
iterative convergence. However, they did not give clear details about normalization. Gholamreza
et al. [36] used time delay cells of TDNNs, designed based on the structure of MLPs, to deal with
the dynamic nature of sample data. Then, they applied factor analysis to select the model inputs.
Results illustrated that TDNN with 2 hidden layers of 15 neurons in each of the layers was the best
architecture. Nourani et al. [9] provided a new solution to EC and TDS prediction. When the predictive
variables were not available, researchers could realize the final predictions through modeling other
relevant variables. They utilized monthly meteorological data RF, RO, and WL to forecast EC and TDS
due to the lack of historical records of outputs. Zounemat-Kermani [82] introduced a Quasi-Newton
method, Broyden–Fletcher–Goldfarb–Shanno (BFGS), to train the parameters of MLP in SS forecasting.
Hameed et al. [60] conducted the sensitivity analysis of the obtained data and scaled it to between 0.1
to 0.9. Results indicated that RBFNN could achieve high-performance accuracy. Heddam and Kisi [47]
utilized open-source data from Eight United States Geological Survey stations (USA) and preprocessed
the data by standardization method. Several ELM models are applied for DO prediction. Yousefi [131]
discussed the Garson method to find the relative importance of each input variable. Results indicated
that including meteorological and hydrologic variables could improve the accuracy of the models
Appl. Sci. 2020, 10, 5776 11 of 49

with fewer influential variables. Elkiran et al. [32] and Najah et al. [132] demonstrated the feasibility
of the ANFIS method in predicting river water quality. This model overcame the shortcomings of
ANN models such as overfitting and local minima, and combined fuzzy logic with ANN to provide
a method to solve uncertain problems. Sinshaw et al. [133] took interrelated and easily measurable
parameters of pH, EC, and Tur, as inputs to realize TN and TP predictions.
Liu et al. [3] pointed out that if more historical data were available [15], ANN models may provide
better predictions than a relatively small data set. Antanasijević et al. [41] tested the performance
of RNN, GRNN, and MLP in small samples prediction. Results indicated that the error of RNN
in test data was less than 10%. Besides, the error of GRNN was lower than MLP. Evrendilek and
Karakaya [55] deleted the missing data directly. Then, discrete wavelet transforms (DWT) with the
orthogonal wavelet families was applied to denoise the data measured by proximal sensors. The result
indicated that the modeling effect of using TLRN to the data after noise reduction was superior
to TLRN, TDNN, and RNN. Chang et al. [12] attempted to use NARX, a dynamic neural network,
to model ten-year seasonal water quality data. Then, 42-fold cross-validation was used to divide the
data. Results demonstrated that the NARX network outperformed BPNN because it could capture the
important dynamic features of TP data. Wang et al. [6] tested the prediction performance of LSTM,
BPNN, Online sequential (OS)-ELM in DO, and TP. The results indicated that LSTM was more accurate
and generalizable than the above feedforward ANNs. Zhao et al. [38] used an improvement of the
ESN, namely RESN, to predict the BOD and TP. This new method used the ridge regression algorithm
to calculate the output weights to solve the ill-posed problem existing in the ESN. Hu et al. [66]
fully preprocessed the acquired water quality data. They firstly imputed, corrected, and denoised
the data by using linear interpolation, smoothing which could attenuate high-frequency signals,
and moving average filtering techniques. Then, correlation analysis, which belongs to analytical
methods, was carried out. The LSTM was adopted for model establishment. Experimental results
showed that the prediction accuracy was high and could reach 98.97% and LSTM was suited for
long-term prediction. J. Liu [67] introduced Back-propagation through time (BPTT) to train the SRU
model. The main difference between SRU and RNN is the “cell state” part added in the hidden layer.
They proposed an Improved mean value method to solve the breakpoint phenomenon of the mean
value method and the linear interpolation method. Results showed that the prediction error was small,
within the range of 1%. Lim et al. [53] converted the irregular data into daily data by using a linear
interpolation method and provided a solution to abnormal data identification. They used a fixed
threshold method to set the upper and lower threshold ranges and proved that linear interpolation had
better robustness than spline interpolation, nearest-neighbor interpolation, and cubic interpolation
according to model results when water quality changed dramatically. Results showed that the removal
of abnormal data beyond the threshold value could preliminarily improve the data convergence.
Partal and Cigizoglu [134] decomposed the measured SS data into wavelet components via
DWT. The DWT-ANN method could more accurately approximate the peak values, which have lesser
distributions compared with non-peak values. Anctil et al. [135] applied MLP to forecast daily SS
and NO3 without considering missing data. They applied a self-organizing map (SOM), a stratified
method, to construct a topological map to visualize the clustered input variables, thereby ensuring that
the statistical properties of the subsets were similar. Levenberg–Marquardt algorithm [24,136–139]
and Bayesian methods were conducted to train the network. Results showed that ANN models could
achieve high accuracy. Sahoo et al. [140] used the SR and AT meteorological data to achieve the WT
prediction. They introduced micro-genetic algorithms (u GA), a creep mutation in small populations,
to update the weights. Wu et al. [141] reported that the GA-BP algorithm whose relative errors were
below 35% was more suitable for TP, TN, and Chl-a prediction than simple multivariate regression
analysis. Kişi [142] utilized neural differential evolution (NDE) models, a combination of neural
networks and differential evolution approaches, to model SS. The result showed that NDE has a low
mean square error. Ömer Faruk [71] investigated the performance of ARIMA-ANN in WT, DO, and B
prediction. Afshar and Kazemi [143] combined PSO and ANN methods in water quality parameter
Appl. Sci. 2020, 10, 5776 12 of 49

prediction. Han et al. [1] used cross-correlation and mutual information to select the input to achieve
the prediction of BOD and DO, respectively. The conjugate gradient algorithm was carried out to train
the model. Areerachakul et al. [144] presented two cluster technique, namely K-means, fuzzy c-means
(FCM) in DO prediction. Results indicated that the performance of hybrid methods was better than
single models. Y. Wang [64] designed a missing–refilling scheme which divided the data into incidental
missing (ID) and structural missing (SD). Then, a temporal exponentially moving average was applied
to fill the missing data. They investigated the time relationship of the DO, NH3 -N univariate time
series using a bootstrapped wavelet neural network (BWNN). Aleksandra and Antanasijevi [42] used
the databases of the European Statistical Office and World Bank to complete the BOD prediction.
Ay and Kisi [43] integrated k-means clustering and MLP in daily COD concentration modeling by
using SS, pH, and WT. Result indicated that this hybrid methods performed better than MLP, RBFNN,
and two different ANFIS approaches (subtractive clustering and grid partition). Ding et al. [120]
collected 23 water quality parameters and considered the problems of data dimensionality. Therefore,
the PCA techniques was used to compress the original data into 15 aggregative indices. Then, the GA
approach was applied to optimize the parameters of BPNN. The result showed that the average
prediction accuracy was up to at least 88%. Gazzaz et al. [145] developed a data mining method,
namely re-sampling, to solve the unbalance problem. Heddam [146] recommended collecting more
than one-year water quality data, because they wanted to include all four seasons in the validation
and testing phases. Liu et al. [147] proposed a hybrid model, namely empirical mode decomposition
(EMD)-BPNN. BPNN predicted each sub-series which are IMFs and the residue decomposed by EMD.
The results demonstrated that a hybrid model could capture the non-stationary characteristics of WT
after EMD. Qiao et al. [44] scaled the datasets between -1 and 1 and then used phase space reconstruction
(PSR) of chaos theory to extract much more information from BOD datasets. Results showed that the
hybrid model, namely chaos theory-PCA-ANN, had high prediction accuracy. Sakizadeh et al. [73]
applied early stopping which is fit for small networks and datasets to determine the model structure.
Yu et al. [148] utilized 5-fold cross-validation to divide the data and applied RBFNN to fuse data
from multiple sensors. The convergence rate and the solution accuracy could be improved through
the variant of PSO (IPSO). The comparison of prediction results validated the effectiveness of the
hybrid model. Zhao et al. [149] converted the signal into an output linear system by the Kalman
filter. The result showed that this hybrid method was a good and effective approach to water quality
prediction. Huang et al. [69] simulated the nonlinearity of data by the combination of the neural
network, fuzzy logic, wavelet transform, and the GA. Results showed that this hybrid model could
handle the problems of data fluctuation. Li et al. [123] adopted the most extreme form of K-fold
cross-validation, namely leave-one-out cross-validation to divide the datasets. Zhang et al., 2017 [16]
divided the dataset into training (98%) and testing sets (2%) and adopted the PSO algorithm to
accelerate the training speed of WNN. Karaboga proved that artificial bee colony (ABC) algorithms
were more precise than GA and PSO [150]. Chen et al. [4] proposed an improved method of ABC
(IABC) which added the optimal and global optimal solution to the updated formulas. The result
indicated that the limitation of the method above was that water quality data needed to obey the
normal distribution appropriately. Li et al. [54] used sparse auto-encoder (SAE) to pre-train the hidden
layer data because SAE contained deep latent features. Qiao et al. [58] determined the structure of DBN
by growing and pruning algorithms instead of artificial experience (SODBN). Results showed that
SODBN could short running time and improve accuracy. Ta and Wei [57] applied Adam optimization
method which could handle sparse gradients on noisy problems to train the parameters of CNN.
Zhou et al. [151] focused on the Improved Grey Relational Analysis (IGRA) method which calculated
the similarity and proximity by relative area change ratio. Fijani et al. [5] used variational mode
decomposition (VMD) algorithm to decompose the highest frequency component produced by a
complete ensemble empirical mode decomposition algorithm with adaptive noise (CEEMDAN).
ELM was applied for modeling. Results indicated that this hybrid model could reduce error whether
in root mean square or mean absolute error. Jin et al. [152] proposed an improvement variant namely
Appl. Sci. 2020, 10, 5776 13 of 49

improved genetic algorithm (IGA) to avoid the situation where excellent individuals are discarded by
the GA. Li et al. [15] introduced evidence theory, that has good data fusion ability, since it is able to
reason with uncertainty to synthesize the evidence from SRU, Gated Recurrent Unit (GRU), LSTM
sources in DO, pH, TP prediction, and eventually reached a certain level of belief. The improved
probability assignment function of the evidence theory, designed based on the softmax function,
could solve the failure of weight allocation problems existing in the traditional probability assignment
function. As a general framework of uncertain reasoning, the application of evidence theory can be
further extended. Tian et al. [153] combined transfer learning (TL) and ANNs approaches which do
not require a large amount of training data because TL has the ability to transfer knowledge from past
tasks to predict Chl-a dynamics. The biggest difference between TL and traditional ANNs methods is
that the former does not need to learn each task from scratch while the latter does. Results indicated
that the hybrid models enhanced the generalization ability compared with the dropout and parameter
norm penalties methods in the long-term application. At the same time, the impact of mutable data
distribution on the models was decreased. Yan et al. [154] utilized mean value method using a median
of k data before and after to correct wrong data and got the missing data by the values of model
prediction of other water quality variables at the missing point. The restricted condition of the model
was that the data were appropriately and normally distributed. Therefore, it is uncertain whether
the above method can be applied to other prediction tasks that do not meet the above conditions.
Yan et al. [68] proposed a hybrid optimized algorithm, namely PSO and GA, to optimize BPNN with
reasonable accuracy. Y. Liu [45] investigated the DO prediction, which considered a temporal and
spatial relationship. Spatial relationship refers to the spatial correlations between external variables
instead of the geographic distributions. The newly proposed attention-RNN model achieved excellent
performance whether in short-term and long-term prediction. Zounemat-Kermani et al. [63] tested the
performance of decomposition approaches, DWT and VMD, in DO prediction. They concluded that
these two methods are an alternative tool for accurate prediction when the input was combination III
and model was MLP.

5. Result
The year of the publication is analyzed at first. Figure 7 plots the number of articles published
from 2008 to 2019 each year. There is a growing number of publications since 2008 that use the ANN
models to predict the water quality, including above 50% of the papers published since 2015, despite the
fact that there are some fluctuations in the quantity of papers—which was in decline in 2010 and
2011. The increasing popularity of ANNs in the field of water resources [155] and environmental
engineering [16] may be explained by the major advantage of the ANNs—that researchers can utilize
them to model nonlinear and complex phenomena even if they do not fully understand the underlying
mechanisms [156]. The popularity of ANNs above is also in agreement with the observations of other
researchers [27,30]. Moreover, the number of papers for different prediction variables is summarized in
Figure 8. The majority of the reviewed papers used chemical water quality variables, such as DO, BOD,
and COD as outputs [30] in the systems of the river, lake, and WWTP. Furthermore, attention was also
directed towards physical variables like pH, WT, and biological variables such as Chl-a.
The number of diverse forecast lengths is shown in Figure 9. The forecast length in this review
refers to the length of time to predict in advance. For example, if researchers used the historical data
of the previous three days to predict the values of the current day, then the forecast length would be
1 [157]. However, 107 papers did not provide details about the forecast length which cast ambiguity
and doubts to researchers in parameter settings [31]. It seems ideal to utilize ANN models to capture
short-term (length = 1) relationships, as the process was carried out 30 times in 44 papers which provide
details about the forecast length, while only 10 papers consider long-term (length > 1) forecasting.
Appl. Sci. 2020, 10, 5776 14 of 49

Figure 7. The distribution of papers between 2008 and 2019.

Figure 8. Number of papers for different prediction variables.

Figure 9. The distribution of prediction lengths.

As mentioned in the Introduction, this review not only includes more water quality parameters but
also more extensive research scenarios compared with the previous reviews. On the whole, there are
23 types of water quality variables examined in this review. They are mainly physical, chemical,
and biological variables. In the field of water quality prediction, relatively mature sensors include
DO, WT, Chl-a, pH, EC, and NH3 -N. There are different application scenarios among the investigated
water quality variables. Table 3 summarizes the main application scenarios of various water quality
variables. Researchers conducted more prediction studies on DO, WT, Chl-a, pH, EC, NH3 -N, Tur,
and S than other water quality variables. It can be seen from Table 3 that there are simple and practical
Appl. Sci. 2020, 10, 5776 15 of 49

sensors that can measure these water quality variables. Therefore, the extensive research of the above
variables may benefit from the wide application of these sensors [148].

Table 3. Basic information of water quality variables.

Water Quality Major


Categories Unit Research Scenarios
Variables Sensors
river, lake, reservoir, WWTP, ponds, coastal waters,
DO chemical mg/L X
creek, drain
BOD chemical mg/L - river, lake, WWTP, mine water experimental system
COD chemical mg/L - river, lake, reservoir, WWTP, groundwater, mine water
WT physical ◦C X river, lake, ponds, catchment, stream, coastal waters
Chl-a biological µg/L X lake, reservoir, surface water, coastal waters
pH physical none X river, lake, WWTP, stream, coastal waters
SS physical mg/L - river, stream, coastal waters, creek, catchment
EC physical us cm−1 X river, lake, reservoir, groundwater, stream
TP physical µg/L - river, lake, WWTP
NH3 -N chemical mg/L X river, lake, reservoir, groundwater experimental system
Tur physical FNU X river, stream
river, groundwater, catchment, wells, aquifer
NO3 chemical mg/L -
experimental system
TDS physical mg/L - river, groundwater, drain
S physical psu X groundwater, coastal waters
TN chemical mg/L - lake, WWTP, coastal waters
B physical mg/L - river
TH physical mg/L - river
TOC chemical mg/L - river
TSS physical mg/L - river
COD Mn chemical mg/L - river
NO2 chemical mg/L - groundwater
P physical mg/L - experimental system
SD physical cm - lake

Table 4 summarizes the data set sizes of feedforward and recurrent neural networks involved in
this review. According to Table 4, the number of samples applied for water quality prediction varies
from 28 [39] to 45,594 [78] which illustrates the fact that ANN models are capable to deal with different
size of the dataset. However, there has been no research studying the optimal amount of data required
for each ANN model. As can be seen from Table 4, the recurrent neural networks [55] generally need
more datasets compared with feedforward neural networks [139]. Research into the water quality
parameter prediction have focused on rivers, WWTP, lake, and reservoir. In contrast, researchers
have done little on artificial facilities, such as stream and pond. In the river system, most researchers
use feed-forward neural networks for modeling, which may be due to the fact that the river system
can be well analyzed using only the feed-forward neural network. This result also applies to WWTP
systems. In the lake system, recurrent neural networks have shown significant results. These two
kinds of neural networks have applications in reservoirs. In contrast, feed-forward neural network can
predict water quality with relatively little data. In addition to being able to perform prediction tasks,
GRNN is also suitable for small data sets (28, 32, 61, 151, 159, 265 samples) compared with other types
of ANNs [24,39–43], so researchers should pay some attention to it.
Appl. Sci. 2020, 10, 5776 16 of 49

Table 4. Datasets of feedforward and recurrent neural networks.

Categories Authors (Year) Methods Scenario (s) Time Step Dataset (Samples)
GRNN, BPNN,
[39] lake weekly 28 (6 months)
RBFNN
[40] ANN(MLP), GRNN coastal waters No details 32 (5 months)
[59] BPNN river No details 39 (3 days)
[158] ANN mine water No details 73
[97] ANN groundwater No details 97
[106] ANN(MLP) surface water No details 110
[159] MLP river No details 110 (8 hours)
[130] ANN(MLP) plain No details 122
[128] ANN groundwater monthly 124 (1 year)
[80] ANN(MLP) stream No details 132 (11 months)
[79] ANN(MLP) basin fortnightly 144 (1 year)
[131] ANN(MLP) river monthly 144 (12 years)
[121] RBFNN river weekly 144
GRNN, ANN(MLP), More than 151
[24] river monthly
RBFNN, MLR samples (6 years)
Open-source
[42] GRNN, MLR No details 159 (9 years)
data
[160] ANN(MLP) river monthly 164 (over 6 years)
[107] ANN reservoir No details 180 (3 years)
[22] ANN system No details 195 (4 years)
[161] ANN(MLP), RBFNN river monthly 200 (17 years)
[139] ANN river No details 200 (16 years)
[93] ANN river No details 232 (3 years)
Feedforward [63] CCNN, MLP river half a month 232 (12 years)
[122] ANN river No details 252 (21 years)
[113] ANN(MLP) river No details 255 (7 months)
ANN(MLP), RBFNN,
[43] WWTP daily 265 (3 years)
GRNN
[119] ELM WWTP daily 360
[75] ANN WWTP daily 364 (1 year)
[118] BPNN reservoir No details 400 (20 years)
[94] BPNN NA No details 500
[95] ANN river monthly 500 (10 years)
[10] ANN groundwater 30 minutes 818 (nearly 17 days)
[162] BPNN river No details 969
[77] MLP, RBF, GRNN Well No details 975 (16 years)
[163] ANN(MLP) stream daily 982 (6 months)
[88] MLP lake No details 1087 (6 years)
[133] ANN lake No details 1217
RBFNN, GRNN, More than 1300
[127] river No details
MLR samples (6 years)
[36] RBFNN, TDNN river monthly 1320 (10 years)
[117] GRNN river No details 1512 (9 years)
[50] MNN WWTP No details 1900
[83] ANN(MLP), RBFNN river daily 2063 (6 years)
[137] ANN river No details 3001
RBFNN, ANN(MLP), upstream and 2063 and 4765
[112] daily
MLR downstream samples (18 years)
more than 3000
[115] ANN river daily
samples (11 years)
[116] ANN lake 15 minutes 6674 (86 days)
[164] ANN river No details 13,800 (5 years)
more than 32,000
[25] GRNN river No details
samples
Open-source
[47] ELM, ANN(MLP) hourly 35,064 (4 years)
data
[78] ANN(MLP) power station 10 minutes 45,594 (2 years)
Appl. Sci. 2020, 10, 5776 17 of 49

Table 4. Cont.

Categories Authors (Year) Methods Scenario (s) Time Step Dataset (Samples)
Elman, GRNN, monthly or
[41] river 61
BPNN, MLR semi-monthly
[12] NARX, BPNN, MLR river monthly 280 (11 years)
[26] LSTM river 12 hours 460 (14months)
[6] LSTM, BPNN lake monthly 657 (7 years)
Mariculture
[66] LSTM, RNN 5 minutes 710 (21 days)
base
Mariculture
Recurrent [67] SRU No details 710
base
[3] Elman pond No details 816 (34 days)
[153] RNN, LSTM reservoir 5 minutes 1440 (5 days)
[15] RNN, BPNN river No details 1448
[165] LSTM lake No details 1520
[54] LSTM, BPNN pond 10 minutes 2880 (20 days)
[38] RESN WWTP No details 5000
[45] RNN Freshwater 10 minutes 5006 (1 year)
[55] TLRN, RNN, TDNN lake 15 minutes 13,744 (573 days)
[154] LSTM WWTP hourly 23,268 (4 years)

The artificial neural network has been widely used in water quality prediction. If researchers only
look at the modeling process, various studies follow some of the steps of the modeling framework
below (see Figure 10).

Figure 10. General framework for water quality modeling.

5.1. Data Collection


The data collection process is not easy due to the requirement of costly measuring instruments
(e.g., water quality sensors, meteorological stations), laboratory equipment, and good operating
conditions. Water quality variables are primarily collected by the sensors. Meteorological variables,
such as AT, WS, RF, SR, Precip, and AP, often influence water quality. Therefore, some researchers took
the meteorological station to obtain the data. In addition, some parameters, such as BOD, COD, need to
be measured by auxiliary laboratory equipment [44]. Location information is essential when researchers
want to make a three-dimensional prediction of water quality. In the above case, the required data is
obtained through the device (see Figure 10). In some studies [42,47], the researchers conduct studies
based on an open-source dataset.
Based on the obtained data, researchers can perform three modeling types. The first type
of modeling is where the researcher models only historical information about the output variable.
Appl. Sci. 2020, 10, 5776 18 of 49

The second type of modeling is when the output variables are difficult to measure, and the researchers
can use easily measured water quality or meteorological data to complete the prediction. In the first
two modeling types, the researchers utilized univariate historical information. However, for the third
type of modeling, the researchers used multivariable historical information. Overall, the researchers
utilized water quality, atmosphere, and other variables such as location data for the prediction task.
The above three modeling types are analyzed from the perspective of data. If analyzed from the
perspective of studying the temporal and spatial relationship between input and output, the above
modeling types can be further divided.

5.2. Output Strategy


The output strategies can be further divided into five categories based on the three modeling
types (see Figure 10). Temporal relationship refers to the relationship learning in the time
dimension. Spatial relationship [45] refers to the spatial correlations between external variables
(see Figure 11). The black origin describes a variety of input variables. Table 5 summarized
the detailed descriptions of the five output strategies. Simply speaking, external variables
are the other variables (more than one) in Multivariate-Input-Itself-Other(multi)-Output.
Univariate-Input-Itself-Output [64] and Univariate-Input-Other(one)-Output [79] refer to the univariate
case, while Multivariate-Input-Other (multi)-Output [35], Multivariate-Input-Itself-Other-Output [52],
and Multivariate-Input-Itself-Other (multi)-Output are multivariate [45] (see Table 5). The model learns
the temporal relationship from five output strategies, while the spatial relationship is only considered
in Multivariate-Input-Itself-Other (multi)-Output. The distinctions between Univariate-Input-Other
(one)-Output and Multivariate-Input-Other (multi)-Output are not only the number of input
variables, but also the fact that the former’s output strategy focuses on time series data while
the latter contains more. The main difference between Multivariate-Input-Itself-Other-Output and
Multivariate-Input-Other (multi)-Output is that the former uses the historical information of the output
variable, while the latter does not.

Figure 11. Temporal and spatial relationship in Multivariate-Input-Itself-Other (multi)-Output.


Appl. Sci. 2020, 10, 5776 19 of 49

Table 5. Five different output strategies.

Category Type Relationship Description


The output(s) at a specific point
Univariate-Input-Itself-Output
Univariate Temporal relationship are learned from its own
(Category 0)
historical information
The output(s) at a specific point
Univariate-Input-Other(one) are learned the historical
Univariate Temporal relationship
-Output (Category 1) information from other
variables (one)
The output(s) at a specific point
Multivariate- Input-Other are learned the historical
Multivariate Temporal relationship
(multi)-Output (Category 2) information from other variables
(more than one)
The output(s) at a specific point
Multivariate-Input-Itself are learned the historical
Multivariate Temporal relationship
-Other-Output (Category 3) information from both its own
and other variables
The output(s) at a specific point
are learned the historical
Multivariate-Input-Itself-Other Temporal relationship
Multivariate information from both its own
(multi)-Output (Category 4) and spatial relationship
and other variables
(more than one)

5.3. Input Selection


There are two main approaches to select the most significant predictors of ANN models which
are model-free and model-based methods (see Table 6) [166]. The biggest difference between the two
methods is that the former does not consider model performance, while the latter does. In the majority
of the studies, many researchers utilized ad-hoc [27] methods to select the inputs, whether in model-free
or model-based methods. Some researchers used cross-correlation and analytical approaches to explore
the linear and non-linear relationship between input(s) and output(s). Other input selection methods
are summarized in Table 6.

Table 6. Model-free and model-based methods in input selection.

Categories Methods Comments


ad-hoc Based on domain knowledge or casual way
model-free analytic The linear and non-linear relationship between input and output
other IGRA, Garson method
ad-hoc e.g., trial-and-error
stepwise Constructive and pruning methods
model-based
sensitivity analysis e.g., MCS
global optimization e.g., GA

5.4. Data Dividing


Data dividing is an important step in the modeling process (see Table 7). The training set is used for
data samples of model fitting [95]. The validation set, which can adjust the model’s hyperparameters,
is a set of samples set apart during model training. Finally, the testing set is to check the model’s
generalization ability [139] and its error is utilized to compare different model’s predictive performance.
Not all data needs to be divided into three sets, because regularization [55] is an approach that can
divide the datasets into two sets—namely training and validation sets—and has the advantage of
providing more data points for the model training and stopping the models from over-fitting [167].
Data dividing methods can be categorized into supervised and unsupervised methods [31]. There are
Appl. Sci. 2020, 10, 5776 20 of 49

no uniform rules for how to divide the training set, the validation set, and the test set which also
applies to the division of training sets and test sets. Most researchers divided the data either by domain
knowledge or in any arbitrary manner. In the majority of the reviewed papers, the data set was divided
into the training and testing two parts (see the ninth column in Appendix A). The division range of the
training set is from 50% to 98% [16], and the test set varies from 2% to 50% [105]

Table 7. Supervised and unsupervised methods in data dividing.

Categories Methods Comments


trial-and-error Taking the statistical properties of each subset into consideration
supervised temporal partitioning Dividing the data into diel, diurnal, and nocturnal
M-test The number of the data points was obtained through the winGamma software
ad-hoc Based on domain knowledge or a casual way
random Divide the data randomly
unsupervised
cross-validation e.g., K-fold cross-validation, leave-one-out cross-validation
stratified method e.g., SOM

5.5. Data Preprocessing


It should be noted that data preprocessing is carried out after the data dividing. Normalization,
missing values imputation and data correct are three primary preprocessing methods in the field of
water quality modeling (see Table 8). Most reviewed papers took the normalization step, although
they did not give clear details about normalization. As [31] pointed out, this step requires matching
the range of the predictors to the transfer function in the hidden layer. Range scaling [132] and
standardization [113] are two popular categories in normalization. There are three main scopes,
namely [0, 1], [−1, 1] and [0.1, 0.9], under range scaling. Although missing data often occurs in
transmission, only a few investigated papers dealt with this phenomenon. The majority of researchers
deleted the missing data directly. This is not a recommended practice, as the obtained data are precious
and limited. As a whole, researchers pay less attention to data imputation, correct, and identification
of abnormal data. Table 4 presented some data preprocessing techniques.

Table 8. The data preprocessing approaches.

Categories Methods Comments


No details Built-in functions in platforms
Normalization Range scaling The scale of each feature is in the same range
Standardization A new variable with zero mean and unit standard deviation
Only mentioned Not recommend
Deletion Not recommend
Linear interpolation The slope of the assumed line to calculate the data increment
Solve the breakpoint phenomenon of mean value method and linear
Missing data Improved mean value method
interpolation method
imputation Dividing of ID and SD and using Temporal exponentially moving
Missing–refilling scheme
average to fill the missing data
Temporal partitioning as gap-filling in order to get continuous
Gap-filling
records
Filling in the predicted values The missing values of predictors at time T0 are obtained by
of the model prediction values of the model at time T0 by other predictors
Smoothing method The moving average filtering can attenuate high-frequency signals
Data correct
Mean value method Need to be corrected as a median of k data before and after
Data abnormal The fixed threshold method Setting the upper and lower threshold ranges (discard)

5.6. Model Structure Determination


Until recently, a general method for determining the optimal model structure remains unknown [31].
Therefore, different approaches have been adopted to determine the ANN model structure to avoid
the initial difficulty in model building step as much as possible. There are three mainstream
Appl. Sci. 2020, 10, 5776 21 of 49

methods—namely ad-hoc, stepwise trial-and-error, and global methods—for the determination of


an optimal model structure [31] (see Table 9). The neural network structure defines the functional
form of the input–output relationship [59]. The model structure determination, an essential step in
the model development, refers to the number of layers, the number of nodes in each layer and the
way they connect [30], aiming to strike a balance between network complexity and generalization
ability [27]. The model structure determination and model training process are often conducted
together. For example, when the trial-and-error method is implemented, the weights of the MLPs are
optimized at the same time. Categories and comments on the ANN methods in the model structure
determination are given in Table 9. M, N, and O are the number of neurons of the hidden layer,
input layer, and output layer. A is a constant from 1 to 10. Sqrt is a mathematical function [83] used to
calculate the square root of a non-negative real number. Nearly half of the investigated papers did not
provide details on the methods used to determine the ANNs structure. When using fixed network
structures such as GRNNs, this step is not necessary to carry out, although its proportion is relatively
small compared with papers that did not mention this step (see the last two lines in Table 9). In 73 of
the 90 times which provide details of the methods, ad-hoc approaches were utilized to determine the
structure of model. That is to say, most studies still rely on trial-and-error approach to determine the
model structure. This also reveals that researchers have not been very innovative in the methodology
of model structure determination. Seven empirical formulas can help to determine the structure of
the model to a certain extent in the investigated articles. Table 9 also presents the various global
approaches and their improvements in the reviewed articles.

Table 9. Three main model structure determination methods.

Categories Methods Comments Typical Examples


Rule 1: M is less than N minus 1
Rule 2: one range of M is equal to the sqrt of N
plus O and finally plus A [123]
Rule 3: the other range of M is equal to log base
2 logarithm of N
Empirical formula Rule 4: M is equal to 5 multiplied by sqrt of N [102]
Ad-hoc and trial-and-error
Rule 5: M is equal to half of the sum of N and O
approach
plus square root of the number of [102]
training patterns
Rule 6: M is equal to sqrt of N plus one and
[33]
finally plus A
Rule 7: M is equal to sqrt of N multiplied by O [99]
Trial-and-error Purely on a trial-and-error approach [105]
With each modification of the trial, a structure
Stepwise Stepwise
that is neither too complex nor too simple [99]
trial-and-error trial-and-error
is building
Searching the solution space through simulated
GA [166]
natural evolution
Introducing creep mutation in a
u GA [140]
small population
Selecting excellent individuals effectively to
Global IGA [152]
avoid the situation of discarding by GA
methods
Excitation function does not need to be
PSO [143]
differentiable and derivable
The convergence rate and accuracy of the
IPSO [148]
solution are improved
ABC More precise than PSO and GA [4]
IABC Updating formulas just like the PSO algorithm [4]
Not mentioned Not recommend [40]
Others
Not required Fixed structures such as GRNNs [25]
Appl. Sci. 2020, 10, 5776 22 of 49

5.7. Model Training


There are two main training methods, namely deterministic and stochastic methods [31]
(see Table 10). Deterministic methods look for a single parameter vector while the stochastic
methods search for the distribution of the model parameters with the purpose of minimizing
the model error [27]. In a more detailed division, local methods (L) that often work on
gradient information and global optimization approaches are two kinds of deterministic methods.
Gradient methods can be further sub-divided into first-order methods or second-order methods.
Deterministic methods based on gradient information have been widely used in model training
algorithm. The Levenberg–Marquardt algorithm, a second-order method, was most widely used
in deterministic methods. The Levenberg–Marquardt method combines the advantages of BP and
Newton algorithm, and its training speed is obviously faster than BP and momentum algorithm [81].
However, it has the disadvantage of being incompatible with regular terms, and requires a lot of
memory when datasets are large. Sixty-two of the papers did not provide details about the model
training algorithm. Seven categories of local methods (see line 1 to line 7 in Table 10) are summarized.
Relatively speaking, there were few works on network training using Bayesian [27] and the Adam
optimization methods [57].

Table 10. The deterministic and stochastic methods in model training.

Categories Methods Comments


BP algorithm(L) Computing the direction of gradient descent
Newton’s methods(L) The computing tasks are implemented by Hessian matrix
Conjugate gradient The search direction is carried along the conjugate
method(L) direction and does not need to use Hessian matrix
Levenberg–Marquardt A method, combination of BP and Newton algorithm,
Deterministic method(L) use Jacobian matrix to do the computing tasks
The Quasi-Newton It is applied to the situation of that Jacobian matrix or
method(L) Hessian matrix is difficult or even impossible to compute
A Quasi-Newton method implemented by the built-in
BFGS
function in R
A gradient descent with momentum and
TRAINLM
Levenberg–Marquardt backpropagation
Global optimization See Table 9
Bayesian methods Prediction limits can be obtained
Stochastic methods It implemented a reverse gradient update with the value
Adam optimization method
obtained by Mini batch data
Emerging methods Online learning algorithm Quickly adjust the model in real time

6. Discussion

6.1. Data Are the Foundation


Data selection strategy: Data collection is a costly and time-consuming process. This is mainly
due to the expensive equipment, limited experimental time, and conditions. The ANN model is a
data-driven model, so obtaining enough data is the basis of modeling. The need to collect as much data
as possible has been put forward in the existing literature. To address this need, researchers need to
consider two factors. One is whether the historical information of the output variable can be collected.
The second is what strategy researchers need to choose when the historical information of output
variables are difficult to measure. If researchers can collect historical observations of target variables,
they can process the data and model it. If target variables cannot be obtained, researchers can collect
variables such as water quality and meteorology data associated with output variables. Part of the
literatures collect target variables by means of obtaining open-source data. This approach has benefited
from a number of government data collection programs. However, research to obtain open-source
Appl. Sci. 2020, 10, 5776 23 of 49

water quality data is rather limited. Therefore, researchers are encouraged to open up their own data
resources in the future to make contributions to themselves, others, and society.
Data volume demand: According to the results of the existing literature review, there is no
systematic research to investigate how to determine the optimal number of samples required for
each type of ANN model. In general, RNN requires more data than feedforward artificial neural
networks. In addition, GRNN in feedforward artificial neural networks can handle small sample
problems. When researchers use the RNN method to make water quality predictions, they need about
a thousand pieces of data. When the researchers utilized the feedforward artificial neural networks
method to make the prediction, about 500 pieces of data are needed. When researchers used the GRNN
method to make predictions, they need about 100 pieces of data. When researchers want to make
long-term forecasts of water quality data that are periodic after a year, at least one year of data needs to
be collected. This also applies when researchers want to include four seasons in the model validation
and testing phase.

6.2. Data Processing Is Key


Data imbalance problem: Both the peak and the extreme value occupy a relatively small proportion
of the distribution. Only a handful of researchers currently consider data imbalances. In order to obtain
higher prediction accuracy and reduce the error of the peak, some new prediction approaches, such as
wavelet analysis method, can be used for reference. Besides, modelers in the future can develop a
form of extreme value loss for detecting the future occurrence of extreme values (Ding et al., 2019) and
apply it to the water quality prediction.
Input selection problem: The quality of data sets has been affected by many factors. These factors
include but are not limited to temporal resolution (e.g., monthly vs. hourly), number of predictors,
or noise in the data. Therefore, it is very important to select the appropriate input and preprocess
the data. This review found that the vast majority of researchers chose inputs based on their domain
knowledge or in any arbitrary manner. Such input selection methods have some limitations because
they neither analyze the relationship between input and output, nor consider the performance of the
model. Some studies use cross-correlation to explore the relationship between inputs and outputs.
It is a linear approach, which is contrary to the premise of using a nonlinear neural network model.
Researchers can use nonlinear analysis methods such as mutual information to select inputs.
Output strategies problem: A variety of output strategies were adopted in the reviewed
papers—the quantity of which is 18—because researchers hope to select the most suitable through
comparison to illustrate the relationship between input(s) and output(s), which is good practice.
Multivariate-Input-Other (multi)-Output is the most popular output strategy which represents the
case where the output(s) at a specific point is learned the historical information from other variables
(more than one). Few studies have considered the spatial relationship between exogenous variables.
This may be due to the fact that external variables do not influence the outcome of the forecast most
of the time. However, researchers must be aware that exogenous variables can have a significant
impact on predictions at some point. For example, the effect of water circulation on dissolved oxygen.
A recent research used the mechanism of attention to simultaneously explore the relationship between
temporal and spatical, and applied it to DO prediction. Researchers can use this method for reference
to further explore the spatial relationship of other water quality variables.
Forecasting length problem: At present, the research mainly focuses on the short-term prediction,
and the research on the long-term prediction is relatively limited. The reason for this phenomenon
is that with the increase in the prediction length, the uncertainty factors also increase, which leads
to the accumulation of errors and thus reduces the accuracy of the prediction. Researchers can
adopt appropriate strategies to solve such problems in forecating field, such as Recursive, DirRec,
and Multiple Output Strategies [168].
Data dividing problem: At present, researchers tend to use ad-hoc method to divide the training
set data into 70% to 90% of the total data. The most common percentage of the training, validation,
Appl. Sci. 2020, 10, 5776 24 of 49

and testing is 60%, 20%, 20%, and 50%, 25%, 25%. Such methods based on the expertise of researchers
or divide data in arbitrary ways has certain universality. However, this approach has not promoted
the development of data partitioning methods. It is always difficult to determine the number of K
for common K-fold cross-validation, as the results may have a considerable bias [169]. Therefore,
leave-one-out cross-validation, the most extreme form of K-fold cross-validation, should be encouraged
for use because it has been shown to provide a good estimation of the model’s true generalization
capabilities in the case of fewer training data or more model parameters despite the limited usage.
Data preprocessing problem: Most studies use the normalization method for pre-processing data,
but it does not disclose specific details. This is probably due to the use of built-in functions to deal
with normalization in many platforms. However, this basic information should be clearly defined,
because different scaling ranges have different effects on the final result of the model. In the face
of missing data, researchers will simply delete it. This approach is not worth advocating because
data is precious. Researchers can adopt appropriate populating strategies to deal with missing data.
Some imputation methods besides linear interpolation—such as the improved mean value method
that can solve the breakpoint phenomenon of linear interpolation, and designing filling schemes such
as missing–refilling schemes or gap-filling to obtain continuous records—are worthy of exploring.
The restricted condition of the model forecasting methods using prediction values to fill the missing
vales is that the data are appropriately and normally distributed. Therefore, it is uncertain whether
the above method can be applied to other prediction tasks that do not meet the above conditions.
Existing literature has shown that the identification of error and abnormal data is a difficult task
because they are difficult to define in water quality prediction. How to deal with such data still needs
further exploration by researchers.

6.3. Model Is the Core


Model structure determination problem: Most researchers use a trial-and-error method to
determine the ANN structure, which does not fundamentally promote the further development of the
model. This review summarizes some empirical formulas to determine the number of neurons in the
hidden layer that future researchers can apply to their studies. This review does not reveal the science
behind these formulas or the conditions under which they apply. To some extent, the use of these
empirical formulas contributes to the determination of the model structure, because researchers build
on previous studies rather than stay at the level of trial-and-error with no rules to find. Global methods
can obtain topology and network weights, which have been developed to some extent in recent
years. Compared with the trial-and-error method, the global method has a sound theoretical basis.
Researchers can further study and improve global methods.
Activation function determination problem: Most of the time, researchers choose S-shaped
functions because they create a random nonlinear mapping between the input and output. The essential
reason is that the S-type transfer function is differentiable, continuous, and monotonically increasing
in its domain. Purelin is used more frequently in the output layer than other functions because its
output can be arbitrary rather than limited to a small range compared to the sigmoid function.
Model training problem: The reason for developing so many subclasses of training algoriths
is that researchers want to use the appropriate matrix (e.g., Hessian matrix, Jacobian matrix) to
accomplish the computing tasks easily. The Quasi-Newton method is suited for the situation that
the matrix (whether Hessian or Jacobian) is difficult or even impossible to compute. In water quality
prediction, the deterministic methods are more mature than the stochastic methods. One possible
reason is that the former only looks for a single parameter vector, while the latter looks for the model
parameter distribution, so the latter parameter is more uncertain. The online learning algorithm has
the characteristics of real-time and rapid adjustment model which is suitable for prediction tasks.
However, its application in water quality prediction is still very limited. Therefore, the algorithm is
worthy of further study.
Appl. Sci. 2020, 10, 5776 25 of 49

Model structure selection problem: Many researchers utilized MLP architectures in ANN to
complete prediction tasks between 2008–2019. This result is as same as the conclusion of the review
between 1999 to 2007, which may be due to the fact that the MLPs architecture has the advantage of
being easy to use, and they can approximate any relationship between input(s) and output(s) through
the typical three layers [81]. Global methods (see Table 9), obtaining topology structure and network
weights, are drawing the attention of researchers—in contrast to the previous review [27]. It must
be noted that the GA, PSO, and ABC methods are typical examples of evolution-related methods.
In general, evolutionary methods are combined with ANNs to meet different constraints.
Much effort has been made regarding the data-intensive methods, while the model-intensive and
technique-intensive approaches were implemented relatively infrequently. Wavelet analyses were
widely used in data-intensive methods, while the decomposing approaches were used less. This may
be because wavelet analysis has the ability to extract the trends, discontinuities, and breakdown points
of the original data. Furthermore, it is also able to process signals by compressing or denoising.
In recent years, CNN, as a new feedforward neural network method, has been used in water
quality prediction. However, its application is rather limited. Researchers can further expand CNN’s
reach. RNN has good memory ability, so it can make full use of historical information and lay
a solid foundation for realizing long-term prediction of water quality. Hybrid Models should be
further developed because they are not a substitute for traditional technologies, but a combination
of their strengths. Researchers can refer to the ensemble approaches, transfer learning technology,
and evidence theory in the literature to improve the prediction accuracy and generalization ability,
and accommodate uncertainty.

Author Contributions: Conceptualization and review framework, Y.C.; original draft preparation and writing,
L.S.; review and editing, Y.L., L.Y.; supervision, D.L. All authors have read and agreed to the published version of
the manuscript.
Funding: This research was funded by the Next Generation Precision Aquaculture: R&D on intelligent
measurement, control technology (Project Number:2017YFE0122 100).
Acknowledgments: The authors would like to thank the anonymous reviewers for their valuable and insightful
comments on an earlier version of this manuscript.).
Conflicts of Interest: The authors declare no conflict of interest.
Appl. Sci. 2020, 10, 5776 26 of 49

Appendix A

Table A1. Details of the reviewed papers.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
BOD; SS, TN, 364 samples Train: 67%,
Feedforward [75] WWTP(Turkey) NA Q Category 2 daily ANN, MLR NA
TP (1 year) test:33%
Mamasin dam
DO, EC; SS,
Feedforward [76] reservoir RF AODD Category 2 No details No details No details ANN(MLP) NA
TN, WT
(Turkey)
Singapore
S, DO, Chl-a;; 32 samples ANN(MLP),
Feedforward [40] coastal waters NA NA Category 3 No details No details 1
WT (5 months) GRNN
(Singapore)
Feitsui Reservoir Train: 75%,
Feedforward [19] Chl-a; NA Bands Category 2 No details No details ANN(MLP) NA
(China) test:25%
ANN,
Pyeongchang No details
Feedforward [90] DO, TOC; WT NA Q Category 3 5 minutes No details MNN, 12,24
river (Korea) (3 months)
ANFIS
Train:57%,
Feitsui Reservoir No details
Feedforward [170] Chl-a; NA Bands Category 2 No details validate: 29%, ANN(MLP) NA
(China) (7 years)
test: 14%
BOD; COD,
Train:60%,
Melen River WT, DO, Chl-a, No details
Feedforward [91] NA F, Ns Category 2 monthly validate: 20%, ANN(MLP) NA
(Turkey) NH3 -N, NO3 , (over 6 years)
test: 20%
NO2
Moshui River mineral No details Train:80%,
Feedforward [92] COD, NH3 -N;; NA Category 0 No details BPNN NA
(China) oil;; (5 years) test: 20%
Train:50%,
Doce River WT, pH, EC, 232samples
Feedforward [93] NA other ions Category 2 No details validate: 25%, ANN NA
(Brazil) TN (3 years)
test: 25%
pH, DO;; WT,
Train:80%,
Feedforward [94] NA (China) S, NH3 -N, NA NA Category 3 500 samples No details BPNN NA
test: 20%
NO2
DO, BOD; pH,
Train:60%,
Gomti river TA, TH, TS, 500 samples
Feedforward [95] RF NA Category 2 monthly validate: 20%, ANN NA
(India) COD, NH3 -N, (10 years)
test: 20%
NO3 , P
Appl. Sci. 2020, 10, 5776 27 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Pyeongchang No details
Feedforward [96] TOC;; Precip Q;; Category 3 No details No details ANN NA
River (Korea) (7 years)
Groundwater other 7 Train:56%,
Feedforward [97] NO2 , COD;; NA Category 3 97 samples No details ANN NA
(China) variables test: 44%
DO; BOD,
Omerli Lake No details ANN,
Feedforward [98] NH3 -N, NO3 , NA NA Category 2 No details No details NA
(Turkey) (17 years) MLR, NLR
NO2 , P
Changle River DO, TN, TP;; No details
Feedforward [99] RF F, FTT Category 3 monthly No details BPNN NA
(China) WT (18months)
Sangamon River No details Train:50%,
Feedforward [105] NO3 ;; AT, Precip Q Category 3 weekly ANN 1
(USA) (6 years) test: 50%
Surface water other 12 Train:67%,
Feedforward [106] Chl-a; NA Category 2 110 samples No details ANN(MLP) NA
(Turkey) variables test: 33%
DO; pH, WT,
Gruˇza reservoir 180samples Train:84%,
Feedforward [107] CL, TP, NO2 , NA Fe, Mn Category 2 No details ANN NA
(Serbia) (3 years) test: 16%
NH3 -N, EC
Train:57%,
No details
Feedforward [108] The tank (China) DO;; pH, S, WT AT NA Category 3 1 minute validate: 29%, ANN 30
(22 months)
test: 14%
WL, T,
Groundwater No details Train:29%,
Feedforward [109] S; EC NA Pumping, Category 2 No details ANN NA
(India) (7 years) test: 71%
Rainp
BOD; COD, SS, Train:50%,
Feedforward [110] WWTP(China) NA Oil Category 2 No details No details RBFNN 5
pH, NH3 –N test: 50%
Mg, Cl, Na, ANN,
818samples
Groundwater NO3 ; pH, EC, K, HCO3 , Train:70%, test: Linear
Feedforward [10] NA Category 2 (nearly 30 minutes NA
(Iran) TDS, TH SO4 , Ca, 30% regression
17days)
ICs (LR)
Q, other
975samples MLP, RBF,
Feedforward [77] Wells (Palestine) NO3 ; NA five Category 2 No details No details NA
(16 years) GRNN
variables
Upstream and 2063, 4765 Train:50%, RBFNN,
DO; pH, WT,
Feedforward [112] downstream NA Q Category 2 samples daily validate:25%, ANN(MLP), NA
EC
(USA) (18 years) test: 25% MLR,
Train:45%,
Feedforward [50] WWTP (Korea) DO;; NH3 -N NA NA Category 3 1900 samples No details validate:5%, test: MNN NA
50%
Appl. Sci. 2020, 10, 5776 28 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Eastern Black Train:75%,
144 samples
Feedforward [79] Sea Basin SS; Tur NA NA Category 1 fortnightly validate:8%, ANN(MLP) NA
(1 year)
(Turkey) test: 17%
DO, BOD, Train:80%,
Kinta River 255 samples
Feedforward [113] NH3 -N, pH, NA NA Category 2 No details validate:10%, ANN(MLP) NA
(Malaysia) (7 months)
COD, Tur;; test: 10%
45,594
Power station AT, AP, WD, other 8 Train:70%,
Feedforward [78] WT; Category 2 samples 10 minutes ANN(MLP) 12
(New Zealand) WS variables test: 30%
(2 years)
Train:70%,
Yuan-Yang Lake SR, AP, RH, No details
Feedforward [114] WT; ST Category 2 10 minutes validate & ANN(MLP) 1
(China) AT, WS, WD (2 months)
test: 30%
BOD, NH3 -N,
Experimental NO3 , P; DO, 195samples Train: 62%,
Feedforward [22] NA RP Category 2 No details ANN NA
system (UK) WT, pH, EC, (4 years) test: 38%
TSS, Tur
DO, TP, SD, Category 2
Lake Fuxian
Feedforward [11] Chl-a;; TN, WT, NA Month; and No details No details No details ANN NA
(China)
pH Category 3
ANN,
Category 1 more than Support
Doiraj River
Feedforward [115] SS; RF Q and 3000 samples daily No details vector 1
(Iran)
Category 2 (11 years) regression
(SVR)
ANN,
Train:60%, Multiple
Lake Abant DO, Chl-a; WT, 6674 samples
Feedforward [116] NA MDHM Category 2 15 minutes validate:15%, nonlinear NA
(Turkey) EC (86 days)
test: 25% regression
(MNLR)
The test set is
Johor River, approximately ANN(MLP),
No details
Feedforward [37] Sayong River TDS, EC, Tur; NA NA Category 1 No details 10–40 % of the RBFNN, NA
(5 years)
(Malaysia) size of the LR
training data set
BOD, COD;
Mine water Train:79%,
Feedforward [158] WT, pH, DO, NA other Category 2 73 samples No details ANN NA
(India) test: 21%
TSS
Appl. Sci. 2020, 10, 5776 29 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
DO; pH, NO3 , Train:60%,
Heihe River 164 samples
Feedforward [160] NH3-N, EC, NA Cl, Ca Category 2 monthly validate:20%, ANN(MLP) NA
(China) (over 6 years)
TA, TH test: 20%
Na, CL,
Train:70%,
Danube River DO; WT, pH, SO4, HCO3 , 1512 samples
Feedforward [117] Category 2 No details validate:20%, GRNN NA
(Serbia) NO3 , EC other 11 (9 years)
test: 10%
variables
Category 1
Stream Harsit 132 samples
Feedforward [80] SS; Tur NA TCC, TIC and No details No details ANN(MLP) NA
(Turkey) (11months)
Category 2
DO; WT, pH, BPNN,
Feitsui Reservoir 400 samples
Feedforward [118] EC, Tur, SS, TH, NA NA Category 2 No details No details ANFIS, NA
(China) (20 years)
TA, NH3 -N MLR
Form
Train:90%,
attributes, 982
Feedforward [163] Stream (USA) WT; AT Category 2 daily validate & ANN(MLP) NA
forested (6 months)
test: 10%
land cover
The Bahr Hadus Train:80%, CCNN,
Feedforward [49] DO, TDS;; NA NA Category 0 No details monthly NA
drain (Egypt) test: 20% BPNN
DO, COD,
ANN(MLP),
Karoon River BOD; EC, pH, 200 samples Train:80%,
Feedforward [161] NA Ca, Mg, Na Category 2 monthly RBFNN, NA
(Iran) Tur, NO3 , NO2 , (17 years) test: 20%
ANFIS
P
EMS
Manawatu River (Energy, Train: 70%,
Feedforward [121] NO3 ; NA Category 1 144 samples weekly RBFNN NA
(New Zealand) Mean, test: 30%
Skewness)
HELM,
BOD; DO, pH, Train: 83%, Bayesian
Feedforward [119] WWTP (China) NA F, TNs Category 2 360 samples daily NA
SS test: 17% approach,
ELM
Tur; NH3 -N,
Nalón river No details Train: 90%,
Feedforward [35] EC, DO, pH, NA NA Category 2 15 minutes ANN(MLP) NA
(Spain) (1 year) test: 10%
WT
Groundwater SAR, SO4; 124 samples Train: 84.1%,
Feedforward [128] pH, TDS, TH NA Category 2 monthly ANN NA
(Turkey) CL (1 year) test: 15.9%
Appl. Sci. 2020, 10, 5776 30 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Train:60%,
Johor River DO; WT, pH, No details ANN(MLP),
Feedforward [89] NA NA Category 2 monthly validate: 25%, NA
(Malaysia) NO3 , NH3 -N (10 year) ANFIS
test: 15%
The Taipei Water
No details
Feedforward [129] Source Domain Tur; RF NA Category 2 No details No details BPNN NA
(1 year)
(China)
ANN(MLP),
Category 2 Train:65%,
Mashhad plain CL; Lon, ANFIS,
Feedforward [130] EC; NA and 122 samples No details validate: 20%, NA
(Iran) Lat geostatistical
Category 3 test: 15%
models
DO; pH, EC, ANN,
Tai Po River 252 samples Train:85%,
Feedforward [122] WT, NH3 -N, NA CL Category 2 No details ANFIS, NA
(China) (21 years) test: 15%
TP, NO2 , NO3 MLR
DOP
DO, BOD, Alk, (dissolved
Ireland Rivers 3001 samples
Feedforward [137] TH;; WT, pH, NA oxygen Category 2 No details No details ANN NA
(Ireland) (No details)
EC percentage),
CL;;
Twostatistical
databases other 20 159 samples Train:88%, GRNN,
Feedforward [42] BOD; DO NA Category 2 No details NA
(European variables (9 years) test: 12% MLR
countries)
HCO3 , SO4 , Train:60%,
Maroon River WT, Tur, pH, No details ANN(MLP),
Feedforward [81] NA CL, Na, K, Category 2 monthly validate: 15%, NA
(Iran) EC, TDS, TH; (20 years) RBFNN
Mg, Ca test: 35%
Na, Mg,
River
CO3 2− , 1320 samples RBFNN,
Feedforward [36] Zayanderud TSS; pH, TH NA Category 2 monthly No details NA
HCO3 , CL, (10 years) TDNN
(Iran)
Ca
Ardabil plain No details Train:71%,
Feedforward [9] EC, TDS; RF RO, WL Category 2 6 months ANN, MLR 1
(Iran) (17 years) test: 29%
BOD; WT, DO, more than
Train:72%,
Danube River pH, NH3 -N, other 8 32,000
Feedforward [25] NA Category 2 No details validate: 18%, GRNN NA
(Serbia) COD, EC, NO3 , variables samples
test: 10%
TH, TP (years)
Appl. Sci. 2020, 10, 5776 31 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Category 0 Train and
Hydrometric No details ANN(MLP),
Feedforward [82] SS;; NA Q and daily test:80%, 1
stations (USA) (8 years) SVR, MLR
Category 3 validate:20%
Category 0 Train:70%,
Surma River No details RBFNN,
Feedforward [138] BOD, COD;; NA NA and No details validate: 15%, NA
(Angladesh) (3 years) MLP
Category 3 test: 15%
Train: more than
Groundwater S; EC, TDS, No details ANN(MLP),
Feedforward [85] NA Mg, Ca, Na Category 2 No details 50%, test: less NA
(Palestine) NO3 (11 years) SVM
than 50%
GRNN,
More than
River Danube DO; pH, WT, ANN(MLP),
Feedforward [24] NA RO Category 2 151 samples monthly No details NA
(Hungary) EC RBFNN,
(6 years)
MLR
Langat River DO, BOD,
No details Train:80%,
Feedforward [60] and Klang River COD, SS, pH, NA NA Category 2 monthly RBFNN NA
(10 years) validate: 20%
(Malaysia) NH3 -N;
Eight United
35,064
States Geological DO; WT, EC, Train:70%, ELM, 1, 12, 24, 48,
Feedforward [47] NA YMDH Category 2 samples hourly
Survey stations Tur, pH test: 30% ANN(MLP) 72, 168
(4 years)
(USA)
DO; WT, pH, Train and
other BPNN,
Feedforward [162] Rivers (China) BOD, NH3 -N, NA Category 2 969 samples No details validate: 80%, NA
variables SVM, MLR
TN, TP test: 20%
DO, BOD, Train:60%,
Syrenie Stawy CL; other No details
Feedforward [86] COD, TN, TP, NA Category 2 monthly validate: 20%, ANN(MLP) NA
Ponds (Poland) ions (19 months)
TA test: 20%
Category 1 ANN(MLP),
Delaware River DO; pH, EC, 2063 samples Train:75%,
Feedforward [83] NA Q and daily RBFNN, NA
(USA) WT (6 years) test: 25%
Category 2 SVM
Na, K, Ca,
Train:50%,
Zayandeh-rood NO3 ; EC, pH, Mg, SO4 ,
Feedforward [84] NA Category 2 No details No details validate: 30%, ANN(MLP) NA
River (Iran) TH CL,
test: 20%
bicarbonate
Appl. Sci. 2020, 10, 5776 32 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Train:60%,
Saint John River TSS, COD, 39 samples BPNN,
Feedforward [59] NA NA Category 2 No details validate: 20%, NA
(Canada) BOD, DO, Tur; (3 days) SVM
test: 20%
CL, Na, 13,800
Karkheh River
Feedforward [164] BOD; TDS, EC NA SO4 , Mg, Category 2 samples No details No details ANN NA
(Iran)
SAR, Ca (5 years)
COD; WT, DO,
Xuxi River 110 samples
Feedforward [159] TN, TP, NA NA Category 2 No details No details MLP NA
(China) (8 hours)
NH3 -N, SD, SS
DO; pH, WT,
No details Train:72%,
Danube River EC, BOD, five metal monthly or
Feedforward [102] NA Category 2 (6 years; validate: 18%, BPNN NA
(Serbia) COD, SS, P, ions fortnightly
7 years) test: 10%
NO3 , TA, TH
Train:66%,
Sufi Chai river Q, Other 4 144 samples
Feedforward [131] TDS; NA Category 2 monthly validate: 17%, ANN(MLP) NA
(Iran) variables (12 years)
test: 17%
More than RBFNN,
River Tisza DO; WT, EC, Train:67%,
Feedforward [127] NA RO Category 2 1300 samples No details GRNN, 12
(Hungary) pH test: 33%
(6 years) MLR
SAR;
HCO3 , CL,
Karoon River TH; EC, TDS, No details ANN(MLP),
Feedforward [171] NA SO4 , Ca, Category 2 No details No details NA
(Iran) pH (49 years) RBFNN
Mg, Na, K,
TAC
BPNN,
DO;; BOD,
Yamuna River No details Train:75%, SVM,
Feedforward [32] COD, pH, WT, NA Q Category 3 monthly NA
(India) (4 years) test: 25% ANFIS,
NH3 -N
ARIMA
Chl-a; TP, TN, 1087 samples Train:75%, MLP,
Feedforward [88] Lakes (USA) NA SD Category 2 No details NA
Tur (6 years) test: 25% ANFIS
ANN,
ANFIS,
Karoun River BOD, COD; six mental 200 samples
Feedforward [139] NA Category 2 No details No details Least NA
(Iran) EC, Tur, pH ions (16 years)
Squares
SVM(LSSVM)
Appl. Sci. 2020, 10, 5776 33 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Train:55%,
TN, TP; pH,
Feedforward [133] Lakes (USA) NA NA Category 2 1217 samples No details validate: 22%, ANN, LR NA
EC, Tur
test: 23%
ELM,
Three rivers No details
Feedforward [48] WT; AT Q, DOY Category 2 No details No details ANN(MLP), NA
(USA) (8 years)
MLR
CCNN,
St. Johns River DO; NH3 -N, 232 samples half a Train:75%, DWT,
Feedforward [63] NA CL Category 2 NA
(USA) TDS, pH, WT (12 years) month test: 25% VMD-MLP,
MLP
Train:69%,
Talkheh Rud No details Elman,
Recurrent [111] TDS; NA Q Category 1 No details validate & 1
River (Iran) (13 years) ANN(MLP)
test: 31%
Hyriopsis Train and
816 samples
Recurrent [3] Cumingii ponds DO;; pH, WT SR, WS, AT NA Category 3 No details validate:80%, Elman NA
(34 days)
(China) test: 20%
Elman,
Danube River DO; WT, pH, monthly or Train: 85%, GRNN,
Recurrent [41] NA Q Category 2 61 samples NA
(Serbia) EC semi-monthly test: 15% BPNN,
MLR
Systematical
dynamic-neural
Chou-Shui River No details modeling
Recurrent [167] pH, Alk NA As;; Ca Category 3 No details No details NA
(China) (8 years) (SDM),
BPNN,
NARX
13,744 Train:60%, TLRN,
Yenicaga Lake DO; WT, EC, WL, DOY,
Recurrent [55] NA Category 2 samples 15 minutes validate: 15%, RNN, NA
(Turkey) pH hour
(573 days) test: 25% TDNN
TP;; EC, SS,
NARX,
Dahan River pH, DO, BOD, 280 samples Train:75%,
Recurrent [12] NA Coli Category 3 monthly BPNN, 1
(China) COD, WT, (11 years) test: 25%
MLR
NH3 -N
Appl. Sci. 2020, 10, 5776 34 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
LSTM,
Taihu Lake 657 samples Train:90%,
Recurrent [6] DO, TP;; NA NA Category 0 monthly BPNN, NA
(China) (7 years) test: 10%
OS-ELM
BOD, TP;; Category 2 Train:45%,
Recurrent [38] WWTP(China) COD, TSS, pH, NA ORP and 5000 samples No details validate: 15%, RESN NA
DO, WT Category 3 test: 40%
Mariculturebase WT, pH; EC, S, 710 samples Train:86%, LSTM,
Recurrent [66] NA NA Category 2 5 minutes >32
(China) Chl-a, Tur, DO (21 days) test: 14% RNN
Marine
Train:86%,
Recurrent [67] aquaculture base pH, WT;; NA NA Category 0 710 samples No details SRU NA
test: 14%
(China)
Geum River No details Train:70%, RNN,
Recurrent [53] BOD, COD, SS; AT, WS WL, Q Category 2 daily 1
basin (Korea) (10 years) test: 30% LSTM
Train:65%,
Recurrent [165] Lakes (USA) WT;; NA NA Category 0 1520 samples No details LSTM NA
test: 35%
Category 0 TL-FNN,
Reservoir Chl-a;; WT, pH, 1440 samples
Recurrent [153] NA ORP and 5 minutes No details RNN, NA
(China) EC, DO, Tur (5 days)
Category 2 LSTM
10,060
Two gauged Train: 70–90%,
Recurrent [134] SS;; NA Q Category 1 samples daily WANN NA
stations (USA) test: 30–10%
(30 years)
Agricultural Category 1 26,355
Train: 66.67%, SOM-MLP,
Recurrent [135] catchment NO3 , SS; RF Q and samples daily NA
test: 33.33% MLP
(France) Category 2 (1 year)
u
Train:50%,
Four streams No details GA-ANN,
Recurrent [140] WT; SR, AT NA Category 2 10 minutes validate: 25%, NA
(USA) (4 years) BPNN,
test: 25%
RBFNN
18,368
(TN),1050(TP) GA-BP,
Chaohu Lake Train:86%,
Hybrid [141] TP, TN, Chl-a; NA Bands Category 2 samples No details BPNN, NA
(China) test: 14%
(more than RBFNN
3 years)
Category 1
Two stations 730 samples Train:50%, ANN-differential
Hybrid [142] SS;; NA Q and daily NA
(USA) (2 years) test: 50% evolution
Category 3
Appl. Sci. 2020, 10, 5776 35 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
B¨uy ¨ uk ARIMA-ANN,
108 samples Train:67%,
Hybrid [71] Menderes river WT, DO, B;; NA NA Category 0 monthly ANN, NA
(9 years) test: 33%
(Turkey) ARIMA
Karkheh water quality No details
Hybrid [143] NA NA Category 2 No details No details PSO-ANN NA
reservoir (Iran) variables (6 months)
DO; COD, other two SOM-RBFNN,
Hybrid [1] WWTP(China) NA Category 3 No details daily No details NA
BOD, SS variables ANN(MLP)
DO;; WT, pH, total
13,846
Bangkok canals BOD, COD, SS, coliform, Train: 70%, FCM-MLP,
Hybrid [144] NA Category 3 samples monthly 1
(Thailand) NH3 -N, TP, hydrogen test: 30% MLP
(5 years)
NO2 , NO3 , sulfide
Chl-a; WT, pH,
Lake WANN,
DO, SD, TP, No details
Hybrid [56] Baiyangdian Precip, Evap WL, LV, Sth Category 2 monthly No details ANN, NA
TN, NH3 –N, (10 years)
(China) ARIMA
BOD, COD
BWNN,
Songhua River No details Train:71%, ANN,
Hybrid [64] DO, NH3 -N;; NA NA Category 0 monthly 1
(China) (7 years) test: 29% WANN,
ARIMA
Gazacoastal
NO3 ; EC, TDS, CL, SO4 , No details
Hybrid [136] aquifer Category 2 No details No details K-means-ANN NA
NO3 , Ca, Mg, Na (10 year)
(Palestine)
k-means-MLP,
Arima-RBF,
Train:50%, ANN(MLP),
COD; SS, pH, 265 samples
Hybrid [43] WWTP (Turkey) NA Q Category 2 daily validate:25%, MLR, NA
WT (3 years)
test: 25% RBFNN,
GRNN,
ANFIS
Train:67%,
Yangtze River 480 samples
Hybrid [70] DO, NH3 -N;; NA NA Category 0 weekly validate & ARIMA-RBFNN 1
(China) (9 years)
test: 33%
DO, EC, pH, VP,
Taihu Lake NH3 -N, TN, petroleum, Train:75%,
Hybrid [120] NA Category 2 2680 samples No details PCA-GA-BPNN NA
(China) COD, TP, BOD, other 11 test: 25%
COD; variables
Appl. Sci. 2020, 10, 5776 36 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Category 0
and Train:70%,
Gauging station DO, WT, S;; 650, 540 daily, WANN,
Hybrid [62] NA NA Category 2 validate: 15%, 1, 2, 3
(Iran) Tur, Chl-a samples hourly ANN
and test: 15%
Category 3
Category 0
Two gauging 1974 samples Train:75%,
Hybrid [172] SS;; NA Q and daily WANN NA
stations (USA) (8 years) test: 25%
Category 3
ANN,
River Yamuna 120 samples Train:92.5%,
Hybrid [173] COD;; NA NA Category 0 monthly ANFIS, 9
(India) (10 years) test: 7.5%
WANFIS
MLP,
ANFIS,
WNN,
Q, Product-Unit
Two catchments No details
Hybrid [100] WT; AT declination Category 2 daily No details ANNs 1, 3, 5
(Poland) (10 years)
of the Sun (PUNN),
ensemble
aggregation
approach
South San Train:60%, WANN,
No details
Hybrid [7] Francisco bay Chl-a;; NA NA Category 0 monthly validate: 20%, MLR, 1
(20 years)
(USA) test: 20% GA-SVR
Category 0
Asi River 274 samples Train:75%, WANN,
Hybrid [72] EC;; NA Q and No details NA
(Turkey) (23 years) test: 25% ANN
Category 3
Category 0 Train:80%,
Klamath River DO;; pH, WT, WANN,
Hybrid [146] NA NA and No details monthly validate: 10%, NA
(USA) EC, SD ANN, MLR
Category 2 test: 10%
Prawn culture 1152 samples Train:87.5%, EMD-BPNN,
Hybrid [147] WT; NA NA Category 0 10 minutes 1
ponds (China) (8 days) test: 12.5% BPNN
BOD; COD, SS, 598 samples Chaos
Hybrid [44] WWTP(China) NA NA Category 2 daily No details NA
DO, pH (19 months) Theory-PCA-ANN
Appl. Sci. 2020, 10, 5776 37 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
WANN,
wavelet-gene
Train:70%, expression
Charlotte harbor No details
Hybrid [174] TN; NA NA Category 0 monthly validate: 15%, programing 1
marine waters (13 years)
test: 15% (WGEP),
TDNN,
GEP, MLR
Groundwater EC, Tur, pH, No details Train:80%,
Hybrid [73] NA Cu Category 2 No details PCA-ANN NA
(Iran) NO2 , NO3 (8 years) test: 20%
WT, DO, pH, Train:80%,
Downstream No details
Hybrid [17] EC, TN, TP, NA NA Category 0 daily validate: 10%, Ensemble-ANN 1
(China) (13 months)
Tur, Chl-a; test: 10%
Category 0
and Train:80%,
Karaj River WANN,
Hybrid [104] NO3 ; NA CL; Q Category 1 No details monthly validate: 10%, NA
(Iran) ANN, MLR
and test: 10%
Category 3
Crab ponds SR, WS, AT, 700 samples Train:71%, RBFNN-IPSO-LSSVM,
Hybrid [148] DO;; WT NA Category 3 20 minutes 3
(China) AH (22 days) test: 29% BPNN
Guanting
DO, COD, No details
Hybrid [149] reservoirs NA NA Category 0 weekly No details Kalman-BPNN 2
NH3 -N;; (18 weeks)
(China)
A
Category 0
Toutle River 2000 samples least-square
Hybrid [101] SS;; NA Q and daily No details NA
(USA) (8 years) ensemble
Category 3
models-WANN
Train:70%,
Hybrid [69] WWTP (China) DO; pH NA NA Category 2 50 samples No details FNN-WNN NA
test: 30%
WANN,
Clackamas River 1623 samples Train:78%, WMLR,
Hybrid [52] DO;; WT NA Q Category 3 daily 1, 31
(USA) (6 years) test: 22% ANN(MLP),
MLR
Appl. Sci. 2020, 10, 5776 38 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Chl-a; WT, pH;;
Representative other 17 No details Train:80%,
Hybrid [123] NH3 -N, TN, NA Category 3 No details GA-BP NA
lakes (China) variables (3 years) test: 20%
TP, DO, BOD
PSO-WNN,
Miyun reservoir DO, COD, 5000 samples Train:98%, WNN,
Hybrid [16] NA NA Category 0 weekly NA
(China) NH3 -N; (2 years) test: 2% BPNN,
SVM
Aji-Chay River 315 samples Train:90%, WA-ELM,
Hybrid [126] EC;; NA NA Category 0 monthly 1, 2, 3
(Iran) (26 years) test: 10% ANFIS
Train:50%,
Yangtze River DO, CODMn , 65 samples IABC-BPNN,
Hybrid [4] NA NA Category 3 daily validate: 16%, NA
(China) BOD;; (2 months) BPNN
test: 34%
COD; COD, SS, WANN,
Hybrid [33] WWTP(China) NA NA Category 2 250 samples No details No details NA
pH, NH3 -N ANN(MLP)
The Stream
pH, EC, DO, No details
Hybrid [175] Veszprémi-Séd NA NA Category 2 yearly No details DE-ANN NA
Tur;; (7 years)
(Hungary)
SAE-LSTM,
Shrimp pond DO; WT, AT, AH, AP, 2880 samples Train:75%, SAE-BPNN,
Hybrid [54] NA Category 2 10 minutes 18, 36, 72
(China) NH3 -N, pH WS (20 days) test: 25% LSTM,
BPNN
WANN,
Four basins No details Train:80%,
Hybrid [124] TDS; EC NA Na, CL Category 2 No details GEP, NA
(Iran) (20 years) test: 20%
WANFIS
Category 0
pH, DO, Tur; No details Train:80%, WANN,
Hybrid [125] Blue River (USA) NA Q and daily 1
WT (4 years) test: 20% WGEP
Category 3
WANN,
Chattahoochee 730 samples Train:75%, ANN,
Hybrid [157] pH;; NA Q Category 3 daily 1, 2, 3
River (USA) (2 years) test: 20% WMLR,
MLR
Morava River WT, EC; SS, No details
Hybrid [176] NA other ions Category 2 15 days No details PCA-ANN NA
Basin (Serbia) DO (10 years)
IGRA-LSTM,
Tai Lake, Victoria DO;; WT, pH, No details Train:80%,
Hybrid [151] Precip NA Category 3 No details BPNN, NA
Bay (China) NO2 , TP (7 years) test: 20%
ARIMA
Appl. Sci. 2020, 10, 5776 39 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
WWTP (Saudi
Hybrid [46] C, DO, SS, pH NA CL;; Category 3 774 samples No details No details PCA-ELM NA
Arabia)
Train:70%,
Prespa Lake 363 samples CEEMDAN-VMD
Hybrid [5] DO, Chl-a;; NA NA Category 0 daily validate: 15%, NA
(Greece) (11 months) -ELM)
test: 15%
No details Train:4/9,
The Warta River WANN(MLP),
Hybrid [87] WT;; AT NA Category 3 (22 to 27 daily validate: 2/9, 1
(Poland) MLP
years) test: 1/3
Ashi River DO, NH3 -N, 846 samples more than Train:70%,
Hybrid [152] NA NA Category 0 IGA-BPNN 1
(China) Tur;; (4 hours) 4 months test: 30%
DS-RNN,
Qiantang River Train:70%, RNN,
Hybrid [15] pH, TP, DO;; NA NA Category 0 1448 samples No details NA
(China) test: 30% BPNN,
SVR
WANFIS,
The Johor river NH3 -N, SS, COD Mn , No details MLP,
Hybrid [132] NA Category 2 No details No details NA
(Malaysia) pH; Tur, WT, Mg, Na (1 year) RBFNN,
ANFIS
Bates–Granger
(BG)-least
Hilo Bay (the No details square
Hybrid [103] Chl-a, S;; NA NA Category 0 daily No details 1, 3, 5
Pacific Ocean) (5 years) based
ensemble
(LSE)-WANN
COD, TP, pH, CL,
23,268
TN; DO, oil-related Train:80%,
Hybrid [154] WWTP (China) NA Category 2 samples hourly PSO-LSTM 1
NH3 -N, BOD, quality test: 20%
(4 years)
TH indicators
Beihai Lake pH, Chl-a, DO, No details Train:70%,
Hybrid [68] NA HA;; Category 3 30 minutes PSO-GA-BPNN 12
(China) BOD, EC; (5 days) test: 30%
460 samples Train:95%,
Hybrid [26] River (China) COD;; NA NA Category 0 12 hours LSTM-RNN 1
(14 months) test: 5%
Appl. Sci. 2020, 10, 5776 40 of 49

Table A1. Cont.

Authors Water Quality Meteorological Other Output Prediction


Categories Locations Dataset Time Step Data Dividing Methods
(Year) Variables Factors Factors Strategy Lengths
Zhejiang
Institute of AT, AH, WS, 5006 samples Train:80%, 6, 12, 48,
Hybrid [45] DO; WT SM, ST Category 4 10 minutes attention-RNN
Freshwater WD, SR, AP (1 year) test: 20% 144, 288
Fisheries (China)
grey
Taihu Lake pH; DO, COD, 28 samples Train:75%, theory-GRNN,
Hybrid [39] NA NA Category 2 Weekly 1
(China) NH3 -N (6 months) test: 25% BPNN,
RBFNN
TP; WT, TSS,
Wastewater other 3 1000 samples Train:80%,
Emerging [58] pH, NH3 -N, NA Category 2 No details SODBN NA
factory (China) variables (4 months) test: 20%
NO3 , DO
Recirculating Train:67%,
DO;; EC, pH, 4500 samples CNN,
Emerging [57] Aquaculture NA NA Category 3 10 minutes validate: 11%, 18
WT (13 months) BPNN
Systems (China) test: 22%
The contents before the “;” symbol were the output variables; The contents before the “;;” symbol were output and predictors; NA represents blank content.
Appl. Sci. 2020, 10, 5776 41 of 49

References
1. Han, H.G.; Qiao, J.F.; Chen, Q.L. Model predictive control of dissolved oxygen concentration based on a
self-organizing RBF neural network. Control Eng. Pract. 2012, 20, 465–476. [CrossRef]
2. Zheng, F.; Tao, R.; Maier, H.R.; See, L.; Savic, D.; Zhang, T.; Chen, Q.; Assumpção, T.H.; Yang, P.; Heidari, B.;
et al. Crowdsourcing Methods for Data Collection in Geophysics: State of the Art, Issues, and Future
Directions. Rev. Geophys. 2018, 56, 698–740. [CrossRef]
3. Liu, S.; Yan, M.; Tai, H.; Xu, L.; Li, D. Prediction of dissolved oxygen content in aquaculture of hyriopsis
cumingii using elman neural network. IFIP Adv. Inf. Commun. Technol. 2012, 370 AICT, 508–518. [CrossRef]
4. Chen, S.; Fang, G.; Huang, X.; Zhang, Y. Water quality prediction model of a water diversion project based
on the improved artificial bee colony-backpropagation neural network. Water 2018, 10, 806. [CrossRef]
5. Fijani, E.; Barzegar, R.; Deo, R.; Tziritis, E.; Konstantinos, S. Design and implementation of a hybrid model
based on two-layer decomposition method coupled with extreme learning machines to support real-time
environmental monitoring of water quality parameters. Sci. Total Environ. 2019, 648, 839–853. [CrossRef]
[PubMed]
6. Wang, Y.; Zhou, J.; Chen, K.; Wang, Y.; Liu, L. Water quality prediction method based on LSTM neural
network. In Proceedings of the 2017 12th International Conference on Intelligent Systems and Knowledge
Engineering, ISKE 2017, Nanjing, China, 24–26 November 2017. [CrossRef]
7. Rajaee, T.; Boroumand, A. Forecasting of chlorophyll-a concentrations in South San Francisco Bay using five
different models. Appl. Ocean Res. 2015, 53, 208–217. [CrossRef]
8. Araghinejad, S. Data-Driven Modeling: Using MATLAB®in Water Resources and Environmental Engineering;
Springer: Berlin, Germany, 2014; ISBN 978-94-007-7505-3.
9. Nourani, V.; Alami, M.T.; Vousoughi, F.D. Self-organizing map clustering technique for ANN-based
spatiotemporal modeling of groundwater quality parameters. J. Hydroinformatics 2016, 18, 288–309. [CrossRef]
10. Zare, A.H.; Bayat, V.M.; Daneshkare, A.P. Forecasting nitrate concentration in groundwater using artificial
neural network and linear regression models. Int. Agrophysics 2011, 25, 187–192.
11. Huo, S.; He, Z.; Su, J.; Xi, B.; Zhu, C. Using Artificial Neural Network Models for Eutrophication Prediction.
Procedia Environ. Sci. 2013, 18, 310–316. [CrossRef]
12. Chang, F.J.; Chen, P.A.; Chang, L.C.; Tsai, Y.H. Estimating spatio-temporal dynamics of stream total phosphate
concentration by soft computing techniques. Sci. Total Environ. 2016, 562, 228–236. [CrossRef]
13. Zhang, G.; Patuwo, B.E.; Hu, M.Y. Forecasting with artificial neural networks: The state of the art. Int. J.
Forecast. 1998, 14, 35–62. [CrossRef]
14. Anmala, J.; Meier, O.W.; Meier, A.J.; Grubbs, S. GIS and artificial neural network-based water quality model
for a stream network in the upper green river basin, Kentucky, USA. J. Environ. Eng. 2015, 141, 1–15.
[CrossRef]
15. Li, L.; Jiang, P.; Xu, H.; Lin, G.; Guo, D.; Wu, H. Water quality prediction based on recurrent neural network
and improved evidence theory: A case study of Qiantang River, China. Environ. Sci. Pollut. Res. 2019, 26,
19879–19896. [CrossRef] [PubMed]
16. Zhang, L.; Zou, Z.; Shan, W. Development of a method for comprehensive water quality forecasting and its
application in Miyun reservoir of Beijing, China. J. Environ. Sci. 2017, 56, 240–246. [CrossRef]
17. Seo, I.W.; Yun, S.H.; Choi, S.Y. Forecasting Water Quality Parameters by ANN Model Using Pre-processing
Technique at the Downstream of Cheongpyeong Dam. Procedia Eng. 2016, 154, 1110–1115. [CrossRef]
18. Heddam, S. Modelling hourly dissolved oxygen concentration (DO) using dynamic evolving neural-fuzzy
inference system (DENFIS)-based approach: Case study of Klamath River at Miller Island Boat Ramp, OR,
USA. Environ. Sci. Pollut. Res. 2014, 21, 9212–9227. [CrossRef]
19. Wang, T.S.; Tan, C.H.; Chen, L.; Tsai, Y.C. Applying artificial neural networks and remote sensing to estimate
chlorophyll-a concentration in water body. In Proceedings of the 2008 2nd International Sympoisum
Intelligent Information Technology Application IITA, Shanghai, China, 20–22 December 2008; pp. 540–544.
[CrossRef]
20. Maier, H.R.; Dandy, G.C. Neural networks for the prediction and forecasting of water resources variables:
A review of modelling issues and applications. Environ. Model. Softw. 2000, 15, 101–124. [CrossRef]
21. Loke, E.; Warnaars, E.A.; Jacobsen, P.; Nelen, F.; Do Céu Almeida, M. Artificial neural networks as a tool in
urban storm drainage. Water Sci. Technol. 1997, 36, 101–109. [CrossRef]
Appl. Sci. 2020, 10, 5776 42 of 49

22. Tota-Maharaj, K.; Scholz, M. Artificial neural network simulation of combined permeable pavement and
earth energy systems treating storm water. J. Environ. Eng. 2012, 138, 499–509. [CrossRef]
23. Nour, M.H.; Smith, D.W.; El-Din, M.G.; Prepas, E.E. The application of artificial neural networks to flow
and phosphorus dynamics in small streams on the Boreal Plain, with emphasis on the role of wetlands.
Ecol. Modell. 2006, 191, 19–32. [CrossRef]
24. Csábrági, A.; Molnár, S.; Tanos, P.; Kovács, J. Application of artificial neural networks to the forecasting
of dissolved oxygen content in the Hungarian section of the river Danube. Ecol. Eng. 2017, 100, 63–72.
[CrossRef]
25. Šiljić Tomić, A.N.; Antanasijević, D.Z.; Ristić, M.; Perić-Grujić, A.A.; Pocajt, V.V. Modeling the BOD of Danube
River in Serbia using spatial, temporal, and input variables optimized artificial neural network models.
Environ. Monit. Assess. 2016, 188. [CrossRef] [PubMed]
26. Ye, Q.; Yang, X.; Chen, C.; Wang, J. River Water Quality Parameters Prediction Method Based on LSTM-RNN
Model. In Proceedings of the 2019 Chinese Control and Decision Conference CCDC, Nanchang, China,
3–5 June 2019; pp. 3024–3028. [CrossRef]
27. Maier, H.R.; Jain, A.; Dandy, G.C.; Sudheer, K.P. Methods used for the development of neural networks
for the prediction of water resource variables in river systems: Current status and future directions.
Environ. Model. Softw. 2010, 25, 891–909. [CrossRef]
28. Pu, F.; Ding, C.; Chao, Z.; Yu, Y.; Xu, X. Water-quality classification of inland lakes using Landsat8 images by
convolutional neural networks. Remote Sens. 2019, 11, 1674. [CrossRef]
29. Nourani, V.; Hosseini Baghanam, A.; Adamowski, J.; Kisi, O. Applications of hybrid wavelet-Artificial
Intelligence models in hydrology: A review. J. Hydrol. 2014, 514, 358–377. [CrossRef]
30. Wu, W.; Dandy, G.C.; Maier, H.R. Protocol for developing ANN models and its application to the
assessment of the quality of the ANN model development process in drinking water quality modelling.
Environ. Model. Softw. 2014, 54, 108–127. [CrossRef]
31. Cabaneros, S.M.; Calautit, J.K.; Hughes, B.R. A review of artificial neural network models for ambient air
pollution prediction. Environ. Model. Softw. 2019, 119, 285–304. [CrossRef]
32. Elkiran, G.; Nourani, V.; Abba, S.I. Multi-step ahead modelling of river water quality parameters using
ensemble artificial intelligence-based approach. J. Hydrol. 2019, 577, 123962. [CrossRef]
33. Cong, Q.; Yu, W. Integrated soft sensor with wavelet neural network and adaptive weighted fusion for water
quality estimation in wastewater treatment process. Measurement 2018, 124, 436–446. [CrossRef]
34. Humphrey, G.B.; Maier, H.R.; Wu, W.; Mount, N.J.; Dandy, G.C.; Abrahart, R.J.; Dawson, C.W. Improved
validation framework and R-package for artificial neural network models. Environ. Model. Softw. 2017, 92,
82–106. [CrossRef]
35. Iglesias, C.; Martínez Torres, J.; García Nieto, P.J.; Alonso Fernández, J.R.; Díaz Muñiz, C.; Piñeiro, J.I.;
Taboada, J. Turbidity Prediction in a River Basin by Using Artificial Neural Networks: A Case Study in
Northern Spain. Water Resour. Manag. 2014, 28, 319–331. [CrossRef]
36. Gholamreza, A.; Afshin, M.-D.; Shiva, H.A.; Nasrin, R. Application of artificial neural networks to predict
total dissolved solids in the river Zayanderud, Iran. Environ. Eng. Res. 2016, 21, 333–340. [CrossRef]
37. Najah, A.; El-Shafie, A.; Karim, O.A.; El-Shafie, A.H. Application of artificial neural networks for water
quality prediction. Neural Comput. Appl. 2012, 22, 187–201. [CrossRef]
38. Zhao, J.; Zhao, C.; Zhang, F.; Wu, G.; Wang, H. Water Quality Prediction in the Waste Water Treatment
Process Based on Ridge Regression Echo State Network. IOP Conf. Ser. Mater. Sci. Eng. 2018, 435. [CrossRef]
39. Zhai, W.; Zhou, X.; Man, J.; Xu, Q.; Jiang, Q.; Yang, Z.; Jiang, L.; Gao, Z.; Yuan, Y.; Gao, W. Prediction of water
quality based on artificial neural network with grey theory. IOP Conf. Ser. Earth Environ. Sci. 2019, 295.
[CrossRef]
40. Palani, S.; Liong, S.Y.; Tkalich, P. An ANN application for water quality forecasting. Mar. Pollut. Bull. 2008,
56, 1586–1597. [CrossRef] [PubMed]
41. Antanasijević, D.; Pocajt, V.; Povrenović, D.; Perić-Grujić, A.; Ristić, M. Modelling of dissolved oxygen
content using artificial neural networks: Danube River, North Serbia, case study. Environ. Sci. Pollut. Res.
2013, 20, 9006–9013. [CrossRef]
42. Aleksandra, Š.; Antanasijevi, D. Perić-Grujić, A.; Ristić, M.; Pocajt, V. Artificial neural network modelling
of biological oxygen demand in rivers at the national level with input selection based on Monte Carlo
simulations. Environ. Sci. Pollut. Res. 2015, 22, 4230–4241. [CrossRef]
Appl. Sci. 2020, 10, 5776 43 of 49

43. Ay, M.; Kisi, O. Modelling of chemical oxygen demand by usinAg ANNs, ANFIS and k-means clustering
techniques. J. Hydrol. 2014, 511, 279–289. [CrossRef]
44. Qiao, J.; Hu, Z.; Li, W. Soft measurement modeling based on chaos theory for biochemical oxygen demand
(BOD). Water 2016, 8, 581. [CrossRef]
45. Liu, Y.; Zhang, Q.; Song, L.; Chen, Y. Attention-based recurrent neural networks for accurate short-term and
long-term dissolved oxygen prediction. Comput. Electron. Agric. 2019, 165, 104964. [CrossRef]
46. Djerioui, M.; Bouamar, M.; Ladjal, M.; Zerguine, A. Chlorine Soft Sensor Based on Extreme Learning Machine
for Water Quality Monitoring. Arab. J. Sci. Eng. 2019, 44, 2033–2044. [CrossRef]
47. Heddam, S.; Kisi, O. Extreme learning machines: A new approach for modeling dissolved oxygen (DO)
concentration with and without water quality variables as predictors. Environ. Sci. Pollut. Res. 2017, 24,
16702–16724. [CrossRef] [PubMed]
48. Zhu, S.; Heddam, S.; Wu, S.; Dai, J.; Jia, B. Extreme learning machine-based prediction of daily water
temperature for rivers. Environ. Earth Sci. 2019, 78, 1–17. [CrossRef]
49. Elbisy, M.S.; Ali, H.M.; Abd-Elall, M.A.; Alaboud, T.M. The use of feed-forward back propagation and
cascade correlation for the neural network prediction of surface water quality parameters. Water Resour.
2014, 41, 709–718. [CrossRef]
50. Baek, G.; Cheon, S.P.; Kim, S.; Kim, Y.; Kim, H.; Kim, C.; Kim, S. Modular neural networks prediction model
based A 2/O process control system. Int. J. Precis. Eng. Manuf. 2012, 13, 905–913. [CrossRef]
51. Ding, D.; Zhang, M.; Pan, X.; Yang, M.; He, X. Modeling extreme events in time series prediction.
In Proceedings of the ACM SIGKDD International Conference Knowledge Discovery & Data Mining,
Anchorage, AK, USA, 4–8 August 2019; Volume 1, pp. 1114–1122. [CrossRef]
52. Khani, S.; Rajaee, T. Modeling of Dissolved Oxygen Concentration and Its Hysteresis Behavior in Rivers
Using Wavelet Transform-Based Hybrid Models. Clean-Soil Air Water 2017, 45. [CrossRef]
53. Lim, H.; An, H.; Kim, H.; Lee, J. Prediction of pollution loads in the Geum River upstream using the recurrent
neural network algorithm. Korean J. Agrcultural Sci. 2019, 46, 67–78. [CrossRef]
54. Li, Z.; Peng, F.; Niu, B.; Li, G.; Wu, J.; Miao, Z. Water Quality Prediction Model Combining Sparse
Auto-encoder and LSTM Network. IFAC-PapersOnLine 2018, 51, 831–836. [CrossRef]
55. Evrendilek, F.; Karakaya, N. Monitoring diel dissolved oxygen dynamics through integrating wavelet
denoising and temporal neural networks. Environ. Monit. Assess. 2014, 186, 1583–1591. [CrossRef]
56. Wang, F.; Wang, X.; Chen, B.; Zhao, Y.; Yang, Z. Chlorophyll a simulation in a lake ecosystem using a
model with wavelet analysis and artificial neural network. Environ. Manag. 2013, 51, 1044–1054. [CrossRef]
[PubMed]
57. Ta, X.; Wei, Y. Research on a dissolved oxygen prediction method for recirculating aquaculture systems based
on a convolution neural network. Comput. Electron. Agric. 2018, 145, 302–310. [CrossRef]
58. Qiao, J.; Wang, G.; Li, X.; Li, W. A self-organizing deep belief network for nonlinear system modeling.
Appl. Soft Comput. J. 2018, 65, 170–183. [CrossRef]
59. Sharaf El Din, E.; Zhang, Y.; Suliman, A. Mapping concentrations of surface water quality parameters
using a novel remote sensing and artificial intelligence framework. Int. J. Remote Sens. 2017, 38, 1023–1042.
[CrossRef]
60. Hameed, M.; Sharqi, S.S.; Yaseen, Z.M.; Afan, H.A.; Hussain, A.; Elshafie, A. Application of artificial
intelligence (AI) techniques in water quality index prediction: A case study in tropical region, Malaysia.
Neural Comput. Appl. 2017, 28, 893–905. [CrossRef]
61. Zhang, Y.F.; Fitch, P.; Thorburn, P.J. Predicting the trend of dissolved oxygen based on the kPCA-RNN model.
Water 2020, 12, 585. [CrossRef]
62. Alizadeh, M.J.; Kavianpour, M.R. Development of wavelet-ANN models to predict water quality parameters
in Hilo Bay, Pacific Ocean. Mar. Pollut. Bull. 2015, 98, 171–178. [CrossRef]
63. Zounemat-Kermani, M.; Seo, Y.; Kim, S.; Ghorbani, M.A.; Samadianfard, S.; Naghshara, S.; Kim, N.W.;
Singh, V.P. Can decomposition approaches always enhance soft computing models? Predicting the dissolved
oxygen concentration in the St. Johns River, Florida. Appl. Sci. 2019, 9, 2534. [CrossRef]
64. Wang, Y.; Zheng, T.; Zhao, Y.; Jiang, J.; Wang, Y.; Guo, L.; Wang, P. Monthly water quality forecasting and
uncertainty assessment via bootstrapped wavelet neural networks under missing data for Harbin, China.
Environ. Sci. Pollut. Res. 2013, 20, 8909–8923. [CrossRef]
Appl. Sci. 2020, 10, 5776 44 of 49

65. Lee, K.J.; Yun, S.T.; Yu, S.; Kim, K.H.; Lee, J.H.; Lee, S.H. The combined use of self-organizing map technique
and fuzzy c-means clustering to evaluate urban groundwater quality in Seoul metropolitan city, South Korea.
J. Hydrol. 2019, 569, 685–697. [CrossRef]
66. Hu, Z.; Zhang, Y.; Zhao, Y.; Xie, M.; Zhong, J.; Tu, Z.; Liu, J. A water quality prediction method based on
the deep LSTM network considering correlation in smart mariculture. Sensors 2019, 19, 1420. [CrossRef]
[PubMed]
67. Liu, J.; Yu, C.; Hu, Z.; Zhao, Y.; Xia, X.; Tu, Z.; Li, R. Automatic and accurate prediction of key water quality
parameters based on SRU deep learning in mariculture. In Proceedings of the 2018 IEEE International
Conference on Advanced Manufacturing ICAM 2018, Yunlin, Taiwan, 16–18 November 2018; pp. 437–440.
[CrossRef]
68. Yan, J.; Xu, Z.; Yu, Y.; Xu, H.; Gao, K. Application of a hybrid optimized bp network model to estimatewater
quality parameters of Beihai Lake in Beijing. Appl. Sci. 2019, 9, 1863. [CrossRef]
69. Huang, M.; Zhang, T.; Ruan, J.; Chen, X. A New Efficient Hybrid Intelligent Model for Biodegradation
Process of DMP with Fuzzy Wavelet Neural Networks. Sci. Rep. 2017, 7, 1–9. [CrossRef] [PubMed]
70. Deng, W.; Wang, G.; Zhang, X.; Guo, Y.; Li, G. Water quality prediction based on a novel hybrid model of
ARIMA and RBF neural network. In Proceedings of the 2014 IEEE 3rd International Conference Cloud
Computing Intelligence System, Shenzhen, China, 27–29 November 2014; pp. 33–40. [CrossRef]
71. Ömer Faruk, D. A hybrid neural network and ARIMA model for water quality time series prediction.
Eng. Appl. Artif. Intell. 2010, 23, 586–594. [CrossRef]
72. Ravansalar, M.; Rajaee, T. Evaluation of wavelet performance via an ANN-based electrical conductivity
prediction model. Environ. Monit. Assess. 2015, 187. [CrossRef]
73. Sakizadeh, M.; Malian, A.; Ahmadpour, E. Groundwater Quality Modeling with a Small Data Set. Groundwater
2016, 54, 115–120. [CrossRef]
74. Lin, Q.; Yang, W.; Zheng, C.; Lu, K.; Zheng, Z.; Wang, J.; Zhu, J. Deep-learning based approach for forecast of
water quality in intensive shrimp ponds. Indian J. Fish. 2018, 65, 75–80. [CrossRef]
75. Dogan, E.; Ates, A.; Yilmaz, C.; Eren, B. Application of Artificial Neural Networks to Estimate Wastewater
Treatment Plant Inlet Biochemical Oxygen Demand. Environ. Prog. 2008, 27, 439–446. [CrossRef]
76. Elhatip, H.; Kömür, M.A. Evaluation of water quality parameters for the Mamasin dam in Aksaray City in the
central Anatolian part of Turkey by means of artificial neural networks. Environ. Geol. 2008, 53, 1157–1164.
[CrossRef]
77. Al-Mahallawi, K.; Mania, J.; Hani, A.; Shahrour, I. Using of neural networks for the prediction of nitrate
groundwater contamination in rural and agricultural areas. Environ. Earth Sci. 2012, 65, 917–928. [CrossRef]
78. Hong, Y.S.T. Dynamic nonlinear state-space model with a neural network via improved sequential learning
algorithm for an online real-time hydrological modeling. J. Hydrol. 2012, 468–469, 11–21. [CrossRef]
79. Bayram, A.; Kankal, M.; Önsoy, H. Estimation of suspended sediment concentration from turbidity
measurements using artificial neural networks. Environ. Monit. Assess. 2012, 184, 4355–4365. [CrossRef]
[PubMed]
80. Bayram, A.; Kankal, M.; Tayfur, G.; Önsoy, H. Prediction of suspended sediment concentration from water
quality variables. Neural Comput. Appl. 2014, 24, 1079–1087. [CrossRef]
81. Tabari, H.; Talaee, P.H. Reconstruction of river water quality missing data using artificial neural networks.
Water Qual. Res. J. Canada 2015, 50, 326–335. [CrossRef]
82. Zounemat-Kermani, M.; Kişi, Ö.; Adamowski, J.; Ramezani-Charmahineh, A. Evaluation of data driven
models for river suspended sediment concentration modeling. J. Hydrol. 2016, 535, 457–472. [CrossRef]
83. Olyaie, E.; Zare Abyaneh, H.; Danandeh Mehr, A. A comparative analysis among computational intelligence
techniques for dissolved oxygen prediction in Delaware River. Geosci. Front. 2017, 8, 517–527. [CrossRef]
84. Ostad-Ali-Askari, K.; Shayannejad, M.; Ghorbanizadeh-Kharazi, H. Artificial neural network for modeling
nitrate pollution of groundwater in marginal area of Zayandeh-rood River, Isfahan, Iran. KSCE J. Civ. Eng.
2017, 21, 134–140. [CrossRef]
85. Alagha, J.S.; Seyam, M.; Md Said, M.A.; Mogheir, Y. Integrating an artificial intelligence approach with
k-means clustering to model groundwater salinity: The case of Gaza coastal aquifer (Palestine). Hydrogeol. J.
2017, 25, 2347–2361. [CrossRef]
Appl. Sci. 2020, 10, 5776 45 of 49

86. Miller, T.; Poleszczuk, G. Prediction of the Seasonal Changes of the Chloride Concentrations in Urban Water
Reservoir. Ecol. Chem. Eng. S 2017, 24, 595–611. [CrossRef]
87. Graf, R.; Zhu, S.; Sivakumar, B. Forecasting river water temperature time series using a wavelet–neural
network hybrid modelling approach. J. Hydrol. 2019, 578. [CrossRef]
88. Luo, W.; Zhu, S.; Wu, S.; Dai, J. Comparing artificial intelligence techniques for chlorophyll-a prediction in
US lakes. Environ. Sci. Pollut. Res. 2019, 26, 30524–30532. [CrossRef] [PubMed]
89. Najah Ahmed, A.; Binti Othman, F.; Abdulmohsin Afan, H.; Khaleel Ibrahim, R.; Ming Fai, C.;
Shabbir Hossain, M.; Ehteram, M.; Elshafie, A. Machine learning methods for better water quality prediction.
J. Hydrol. 2019, 578. [CrossRef]
90. Yeon, I.S.; Kim, J.H.; Jun, K.W. Application of artificial intelligence models in water quality forecasting.
Environ. Technol. 2008, 29, 625–631. [CrossRef] [PubMed]
91. Dogan, E.; Sengorur, B.; Koklu, R. Modeling biological oxygen demand of the Melen River in Turkey using
an artificial neural network technique. J. Environ. Manag. 2009, 90, 1229–1235. [CrossRef] [PubMed]
92. Miao, Q.; Yuan, H.; Shao, C.; Liu, Z. Water quality prediction of moshui river in china based on BP neural
network. In Proceedings of the 2009 International Conference Computing Intelligent Natural Computing
CINC, Wuhan, China, 6–7 June 2009; pp. 7–10. [CrossRef]
93. Oliveira Souza da Costa, A.; Ferreira Silva, P.; Godoy Sabará, M.; Ferreira da Costa, E. Use of neural networks
for monitoring surface water quality changes in a neotropical urban stream. Environ. Monit. Assess. 2009,
155, 527–538. [CrossRef] [PubMed]
94. Shen, X.; Chen, M.; Yu, J. Water environment monitoring system based on neural networks for shrimp
cultivation. In Proceedings of the 2009 International Conference Artifitial Intelligence and Computional
Intelligence AICI, Shanghai, China, 7–8 November 2009; pp. 427–431. [CrossRef]
95. Singh, K.P.; Basant, A.; Malik, A.; Jain, G. Artificial neural network modeling of the river water quality—A case
study. Ecol. Modell. 2009, 220, 888–895. [CrossRef]
96. Yeon, I.S.; Jun, K.W.; Lee, H.J. The improvement of total organic carbon forecasting using neural networks
discharge model. Environ. Technol. 2009, 30, 45–51. [CrossRef]
97. Zuo, J.; Yu, J.T. Application of neural network in groundwater denitrification process. In Proceedings of the
2009 Asia-Pacific Conference Information Processing APCIP, Shenzhen, China, 18–19 July 2009; pp. 79–82.
[CrossRef]
98. Akkoyunlu, A.; Akiner, M.E. Feasibility Assessment of Data-Driven Models in Predicting Pollution Trends of
Omerli Lake, Turkey. Water Resour. Manag. 2010, 24, 3419–3436. [CrossRef]
99. Chen, D.; Lu, J.; Shen, Y. Artificial neural network modelling of concentrations of nitrogen, phosphorus and
dissolved oxygen in a non-point source polluted river in Zhejiang Province, southeast China. Hydrol. Process.
2010, 24, 290–299. [CrossRef]
100. Piotrowski, A.P.; Napiorkowski, M.J.; Napiorkowski, J.J.; Osuch, M. Comparing various artificial neural
network types for water temperature prediction in rivers. J. Hydrol. 2015, 529, 302–315. [CrossRef]
101. Alizadeh, M.J.; Jafari Nodoushan, E.; Kalarestaghi, N.; Chau, K.W. Toward multi-day-ahead forecasting of
suspended sediment concentration using ensemble models. Environ. Sci. Pollut. Res. 2017, 24, 28017–28025.
[CrossRef] [PubMed]
102. Šiljić Tomić, A.; Antanasijević, D.; Ristić, M.; Perić-Grujić, A.; Pocajt, V. Application of experimental design
for the optimization of artificial neural network-based water quality model: A case study of dissolved
oxygen prediction. Environ. Sci. Pollut. Res. 2018, 25, 9360–9370. [CrossRef] [PubMed]
103. Shamshirband, S.; Jafari Nodoushan, E.; Adolf, J.E.; Abdul Manaf, A.; Mosavi, A.; Chau, K. Ensemble models
with uncertainty analysis for multi-day ahead forecasting of chlorophyll a concentration in coastal waters.
Eng. Appl. Comput. Fluid Mech. 2019, 13, 91–101. [CrossRef]
104. Rajaee, T.; Benmaran, R.R. Prediction of water quality parameters (NO3 , CL) in Karaj river by using a
combination of Wavelet Neural Network, ANN and MLR models. J. Water Soil 2016, 30, 15–29. [CrossRef]
105. Markus, M.; Hejazi, M.I.; Bajcsy, P.; Giustolisi, O.; Savic, D.A. Prediction of weekly nitrate-N fluctuations in a
small agricultural watershed in Illinois. J. Hydroinformatics 2010, 12, 251–261. [CrossRef]
106. Merdun, H.; Çinar, Ö. Utilization of two artificial neural network methods in surface water quality modeling.
Environ. Eng. Manag. J. 2010, 9, 413–421. [CrossRef]
Appl. Sci. 2020, 10, 5776 46 of 49

107. Ranković, V.; Radulović, J.; Radojević, I.; Ostojić, A.; Čomić, L. Neural network modeling of dissolved oxygen
in the Gruža reservoir, Serbia. Ecol. Modell. 2010, 221, 1239–1244. [CrossRef]
108. Zhu, X.; Li, D.; He, D.; Wang, J.; Ma, D.; Li, F. A remote wireless system for water quality online monitoring
in intensive fish culture. Comput. Electron. Agric. 2010, 71, S3. [CrossRef]
109. Banerjee, P.; Singh, V.S.; Chatttopadhyay, K.; Chandra, P.C.; Singh, B. Artificial neural network model as a
potential alternative for groundwater salinity forecasting. J. Hydrol. 2011, 398, 212–220. [CrossRef]
110. Han, H.G.; Chen, Q.L.; Qiao, J.F. An efficient self-organizing RBF neural network for water quality prediction.
Neural Netw. 2011, 24, 717–725. [CrossRef]
111. Asadollahfardi, G.; Taklify, A.; Ghanbari, A. Application of Artificial Neural Network to Predict TDS in
Talkheh Rud River. J. Irrig. Drain. Eng. 2012, 138, 363–370. [CrossRef]
112. Ay, M.; Kisi, O. Modeling of dissolved oxygen concentration using different neural network techniques in
Foundation Creek, El Paso County, Colorado. J. Environ. Eng. 2012, 138, 654–662. [CrossRef]
113. Gazzaz, N.M.; Yusoff, M.K.; Aris, A.Z.; Juahir, H.; Ramli, M.F. Artificial neural network modeling of the
water quality index for Kinta River (Malaysia) using water quality variables as predictors. Mar. Pollut. Bull.
2012, 64, 2409–2420. [CrossRef] [PubMed]
114. Liu, W.C.; Chen, W.B. Prediction of water temperature in a subtropical subalpine lake using an artificial
neural network and three-dimensional circulation models. Comput. Geosci. 2012, 45, 13–25. [CrossRef]
115. Kakaei Lafdani, E.; Moghaddam Nia, A.; Ahmadi, A. Daily suspended sediment load prediction using
artificial neural networks and support vector machines. J. Hydrol. 2013, 478, 50–62. [CrossRef]
116. Karakaya, N.; Evrendilek, F.; Gungor, K.; Onal, D. Predicting diel, diurnal and nocturnal dynamics of
dissolved oxygen and chlorophyll-a using regression models and neural networks. Clean-Soil Air Water 2013,
41, 872–877. [CrossRef]
117. Antanasijević, D.; Pocajt, V.; Perić-Grujić, A.; Ristić, M. Modelling of dissolved oxygen in the danube river
using artificial neural networks and Monte carlo simulation uncertainty analysis. J. Hydrol. 2014, 519,
1895–1907. [CrossRef]
118. Chen, W.B.; Liu, W.C. Artificial neural network modeling of dissolved oxygen in reservoir.
Environ. Monit. Assess. 2014, 186, 1203–1217. [CrossRef]
119. Han, H.G.; Wang, L.D.; Qiao, J.F. Hierarchical extreme learning machine for feedforward neural network.
Neurocomputing 2014, 128, 128–135. [CrossRef]
120. Ding, Y.R.; Cai, Y.J.; Sun, P.D.; Chen, B. The use of combined neural networks and genetic algorithms for
prediction of river water quality. J. Appl. Res. Technol. 2014, 12, 493–499. [CrossRef]
121. Faramarzi, M.; Yunus, M.A.M.; Nor, A.S.M.; Ibrahim, S. The application of the Radial Basis Function Neural
Network in estimation of nitrate contamination in Manawatu river. In Proceedings of the 2014 International
Conference Computional Science Technology ICCST, Kota Kinabalu, Malaysia, 27–28 August 2014; pp. 1–5.
[CrossRef]
122. Nemati, S.; Fazelifard, M.H.; Terzi, Ö.; Ghorbani, M.A. Estimation of dissolved oxygen using data-driven
techniques in the Tai Po River, Hong Kong. Environ. Earth Sci. 2015, 74, 4065–4073. [CrossRef]
123. Li, X.; Sha, J.; Wang, Z.L. Chlorophyll-A Prediction of lakes with different water quality patterns in China
based on hybrid neural networks. Water 2017, 9, 524. [CrossRef]
124. Montaseri, M.; Zaman Zad Ghavidel, S.; Sanikhani, H. Water quality variations in different climates of Iran:
Toward modeling total dissolved solid using soft computing techniques. Stoch. Environ. Res. Risk Assess.
2018, 32, 2253–2273. [CrossRef]
125. Rajaee, T.; Jafari, H. Utilization of WGEP and WDT models by wavelet denoising to predict water quality
parameters in rivers. J. Hydrol. Eng. 2018, 23. [CrossRef]
126. Barzegar, R.; Asghari Moghaddam, A.; Adamowski, J.; Ozga-Zielinski, B. Multi-step water quality forecasting
using a boosting ensemble multi-wavelet extreme learning machine model. Stoch. Environ. Res. Risk Assess.
2018, 32, 799–813. [CrossRef]
127. Csábrági, A.; Molnár, S.; Tanos, P.; Kovács, J.; Molnár, M.; Szabó, I.; Hatvani, I.G. Estimation of dissolved
oxygen in riverine ecosystems: Comparison of differently optimized neural networks. Ecol. Eng. 2019, 138,
298–309. [CrossRef]
Appl. Sci. 2020, 10, 5776 47 of 49

128. Klçaslan, Y.; Tuna, G.; Gezer, G.; Gulez, K.; Arkoc, O.; Potirakis, S.M. ANN-based estimation of groundwater
quality using a wireless water quality network. Int. J. Distrib. Sens. Netw. 2014, 2014, 1–8. [CrossRef]
129. Yang, T.M.; Fan, S.K.; Fan, C.; Hsu, N.S. Establishment of turbidity forecasting model and early-warning
system for source water turbidity management using back-propagation artificial neural network algorithm
and probability analysis. Environ. Monit. Assess. 2014, 186, 4925–4934. [CrossRef]
130. Khashei-Siuki, A.; Sarbazi, M. Evaluation of ANFIS, ANN, and geostatistical models to spatial distribution
of groundwater quality (case study: Mashhad plain in Iran). Arab. J. Geosci. 2015, 8, 903–912. [CrossRef]
131. Yousefi, P.; Naser, G.; Mohammadi, H. Surface water quality model: Impacts of influential variables. J. Water
Resour. Plan. Manag. 2018, 144, 1–10. [CrossRef]
132. Najah, A.; El-Shafie, A.; Karim, O.A.; El-Shafie, A.H. Performance of ANFIS versus MLP-NN dissolved
oxygen prediction models in water quality monitoring. Environ. Sci. Pollut. Res. 2014, 21, 1658–1670.
[CrossRef] [PubMed]
133. Sinshaw, T.A.; Surbeck, C.Q.; Yasarer, H.; Najjar, Y. Artificial Neural Network for Prediction of Total Nitrogen
and Phosphorus in US Lakes. J. Environ. Eng. 2019, 145, 1–11. [CrossRef]
134. Partal, T.; Cigizoglu, H.K. Estimation and forecasting of daily suspended sediment data using wavelet-neural
networks. J. Hydrol. 2008, 358, 317–331. [CrossRef]
135. Anctil, F.; Filion, M.; Tournebize, J. A neural network experiment on the simulation of daily nitrate-nitrogen
and suspended sediment fluxes from a small agricultural catchment. Ecol. Modell. 2009, 220, 879–887.
[CrossRef]
136. Alagha, J.S.; Said, M.A.M.; Mogheir, Y. Modeling of nitrate concentration in groundwater using artificial
intelligence approach-a case study of Gaza coastal aquifer. Environ. Monit. Assess. 2014, 186, 35–45.
[CrossRef]
137. Salami, E.S.; Ehteshami, M. Simulation, evaluation and prediction modeling of river water quality properties
(case study: Ireland Rivers). Int. J. Environ. Sci. Technol. 2015, 12, 3235–3242. [CrossRef]
138. Ahmed, A.A.M. Prediction of dissolved oxygen in Surma River by biochemical oxygen demand and chemical
oxygen demand using the artificial neural networks (ANNs). J. King Saud Univ. Eng. Sci. 2017, 29, 151–158.
[CrossRef]
139. Najafzadeh, M.; Ghaemi, A. Prediction of the five-day biochemical oxygen demand and chemical oxygen
demand in natural streams using machine learning methods. Environ. Monit. Assess. 2019, 191. [CrossRef]
140. Sahoo, G.B.; Schladow, S.G.; Reuter, J.E. Forecasting stream water temperature using regression analysis,
artificial neural network, and chaotic non-linear dynamic models. J. Hydrol. 2009, 378, 325–342. [CrossRef]
141. Wu, M.; Zhang, W.; Wang, X.; Luo, D. Application of MODIS satellite data in monitoring water quality
parameters of Chaohu Lake in China. Environ. Monit. Assess. 2009, 148, 255–264. [CrossRef]
142. Kişi, Ö. River suspended sediment concentration modeling using a neural differential evolution approach.
J. Hydrol. 2010, 389, 227–235. [CrossRef]
143. Afshar, A.; Kazemi, H. Multi objective calibration of large scaled water quality model using a hybrid particle
swarm optimization and neural network algorithm. KSCE J. Civ. Eng. 2012, 16, 913–918. [CrossRef]
144. Areerachakul, S.; Sophatsathit, P.; Lursinsap, C. Integration of unsupervised and supervised neural networks
to predict dissolved oxygen concentration in canals. Ecol. Modell. 2013, 261–262, 1–7. [CrossRef]
145. Gazzaz, N.M.; Yusoff, M.K.; Ramli, M.F.; Juahir, H.; Aris, A.Z. Artificial Neural Network Modeling of the
Water Quality Index Using Land Use Areas as Predictors. Water Environ. Res. 2015, 87, 99–112. [CrossRef]
[PubMed]
146. Heddam, S. Simultaneous modelling and forecasting of hourly dissolved oxygen concentration (DO) using
radial basis function neural network (RBFNN) based approach: A case study from the Klamath River,
Oregon, USA. Model. Earth Syst. Environ. 2016, 2, 1–18. [CrossRef]
147. Liu, S.; Xu, L.; Li, D. Multi-scale prediction of water temperature using empirical mode decomposition with
back-propagation neural networks. Comput. Electr. Eng. 2016, 49, 1–8. [CrossRef]
148. Yu, H.; Chen, Y.; Hassan, S.; Li, D. Dissolved oxygen content prediction in crab culture using a hybrid
intelligent method. Sci. Rep. 2016, 6, 1–10. [CrossRef]
149. Zhao, Y.; Zou, Z.; Wang, S. A Back Propagation Neural Network Model based on kalman filter for water
quality prediction. In Proceedings of the International Conference Natrual Computation, Zhangjiajie, China,
15–17 August 2015; pp. 149–153. [CrossRef]
Appl. Sci. 2020, 10, 5776 48 of 49

150. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization:
Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [CrossRef]
151. Zhou, J.; Wang, Y.; Xiao, F.; Wang, Y.; Sun, L. Water quality prediction method based on IGRA and LSTM.
Water 2018, 10, 1148. [CrossRef]
152. Jin, T.; Cai, S.; Jiang, D.; Liu, J. A data-driven model for real-time water quality prediction and early warning
by an integration method. Environ. Sci. Pollut. Res. 2019, 26, 30374–30385. [CrossRef]
153. Tian, W.; Liao, Z.; Wang, X. Transfer learning for neural network model in chlorophyll-a dynamics prediction.
Environ. Sci. Pollut. Res. 2019, 26, 29857–29871. [CrossRef] [PubMed]
154. Yan, J.; Chen, X.; Yu, Y.; Zhang, X. Application of a parallel particle swarm optimization-long short term
memory model to improve water quality data. Water 2019, 11, 1317. [CrossRef]
155. Chu, H.B.; Lu, W.X.; Zhang, L. Application of artificial neural network in environmental water quality
assessment. J. Agric. Sci. Technol. 2013, 15, 343–356.
156. Rajaee, T.; Ebrahimi, H.; Nourani, V. A review of the arti fi cial intelligence methods in groundwater level
modeling. J. Hydrol. 2019, 572, 336–351. [CrossRef]
157. Rajaee, T.; Ravansalar, M.; Adamowski, J.F.; Deo, R.C. A New Approach to Predict Daily pH in Rivers Based
on the “à trous” Redundant Wavelet Transform Algorithm. Water. Air. Soil Pollut. 2018, 229. [CrossRef]
158. Verma, A.K.; Singh, T.N. Prediction of water quality from simple field parameters. Environ. Earth Sci. 2013,
69, 821–829. [CrossRef]
159. Ruben, G.B.; Zhang, K.; Bao, H.; Ma, X. Application and Sensitivity Analysis of Artificial Neural Network for
Prediction of Chemical Oxygen Demand. Water Resour. Manag. 2018, 32, 273–283. [CrossRef]
160. Wen, X.; Fang, J.; Diao, M.; Zhang, C. Artificial neural network modeling of dissolved oxygen in the Heihe
River, Northwestern China. Environ. Monit. Assess. 2013, 185, 4361–4371. [CrossRef]
161. Emamgholizadeh, S.; Kashi, H.; Marofpoor, I.; Zalaghi, E. Prediction of water quality parameters of Karoon
River (Iran) by artificial intelligence-based models. Int. J. Environ. Sci. Technol. 2014, 11, 645–656. [CrossRef]
162. Li, X.; Sha, J.; Wang, Z. liang A comparative study of multiple linear regression, artificial neural network and
support vector machine for the prediction of dissolved oxygen. Hydrol. Res. 2017, 48, 1214–1225. [CrossRef]
163. DeWeber, J.T.; Wagner, T. A regional neural network ensemble for predicting mean daily river water
temperature. J. Hydrol. 2014, 517, 187–200. [CrossRef]
164. Ahmadi, A.; Fatemi, Z.; Nazari, S. Assessment of input data selection methods for BOD simulation using
data-driven models: A case study. Environ. Monit. Assess. 2018, 190. [CrossRef] [PubMed]
165. Read, J.S.; Jia, X.; Willard, J.; Appling, A.P.; Zwart, J.A.; Oliver, S.K.; Karpatne, A.; Hansen, G.J.A.; Hanson, P.C.;
Watkins, W.; et al. Process-Guided Deep Learning Predictions of Lake Water Temperature. Water Resour. Res.
2019, 55, 9173–9190. [CrossRef]
166. Bowden, G.J.; Dandy, G.C.; Maier, H.R. Input determination for neural network models in water resources
applications. Part 1—Background and methodology. J. Hydrol. 2005, 301, 75–92. [CrossRef]
167. Chang, F.J.; Chen, P.A.; Liu, C.W.; Liao, V.H.C.; Liao, C.M. Regional estimation of groundwater arsenic
concentrations through systematical dynamic-neural modeling. J. Hydrol. 2013, 499, 265–274. [CrossRef]
168. Bontempi, G.; Ben Taieb, S.; Le Borgne, Y.A. Machine learning strategies for time series forecasting. In Business
Intelligence; Spriger: Berlin/Heidelberg, Germany, 2013; ISBN 9783642363177.
169. Kohavi, R. A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection.
In Proceedings of the 14th International Joint Conference on Artificial Intelligence, IJCAI’95,
Morgan Kaufmann, United States, Montreal, QC, Canada, 20–25 August 1995; pp. 1137–1143.
170. Chen, L.; Hsu, H.H.; Kou, C.H.; Yeh, H.C.; Wang, T.S. Applying Multi-temporal Satellite Imageries to
Estimate Chlorophyll-a Concentration in Feitsui Reservoir using ANNs. IJCAI Int. Jt. Conf. Artif. Intell. 2009,
345–348. [CrossRef]
171. Ebadati, N.; Hooshmandzadeh, M. Water quality assessment of river using RBF and MLP methods of artificial
network analysis (case study: Karoon River Southwest of Iran). Environ. Earth Sci. 2019, 78, 1–12. [CrossRef]
172. Olyaie, E.; Banejad, H.; Chau, K.W.; Melesse, A.M. A comparison of various artificial intelligence approaches
performance for estimating suspended sediment load of river systems: A case study in United States.
Environ. Monit. Assess. 2015, 187. [CrossRef]
Appl. Sci. 2020, 10, 5776 49 of 49

173. Parmar, K.S.; Bhardwaj, R. River Water Prediction Modeling Using Neural Networks, Fuzzy and Wavelet
Coupled Model. Water Resour. Manag. 2015, 29, 17–33. [CrossRef]
174. Rajaee, T.; Shahabi, A. Evaluation of wavelet-GEP and wavelet-ANN hybrid models for prediction of total
nitrogen concentration in coastal marine waters. Arab. J. Geosci. 2016, 9. [CrossRef]
175. Dragoi, E.N.; Kovács, Z.; Juzsakova, T.; Curteanu, S.; Cretescu, I. Environmental assesment of surface waters
based on monitoring data and neuro-evolutive modelling. Process Saf. Environ. Prot. 2018, 120, 136–145.
[CrossRef]
176. Voza, D.; Vuković, M. The assessment and prediction of temporal variations in surface water quality—A case
study. Environ. Monit. Assess. 2018, 190. [CrossRef] [PubMed]

© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).

You might also like