Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000. 10.1109/ACCESS.2022.11635
This research work is supported in part by the U.S. Army Research Office (ARO) under agreement number W911NF-20-2-0266 and the U.S. Office of the Under Secretary of Defense for Research and Engineering (OUSD(R&E)) under agreement number FA8750-15-2-0119. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright notation thereon.
Corresponding author: Lijun Qian (e-mail: liqian@pvamu.edu).
Underwater Acoustic Communication Channel Modeling using Reservoir Computing
Abstract
Underwater acoustic (UWA) communications have been widely used but greatly impaired due to the complicated nature of the underwater environment. In order to improve UWA communications, modeling and understanding the UWA channel is indispensable. However, there exist many challenges due to the high uncertainties of the underwater environment and the lack of real-world measurement data. In this work, the capability of reservoir computing and deep learning has been explored for modeling the UWA communication channel accurately using real underwater data collected from a water tank with disturbance and from Lake Tahoe. We leverage the capability of reservoir computing for modeling dynamical systems and provided a data-driven approach to modeling the UWA channel using Echo State Network (ESN). In addition, the potential application of transfer learning to reservoir computing has been examined. Experimental results show that ESN is able to model chaotic UWA channels with better performance compared to popular deep learning models in terms of mean absolute percentage error (MAPE), specifically, ESN has outperformed deep neural network by 2% and as much as 40% in benign and chaotic UWA respectively.
Index Terms:
Channel modeling, Deep Learning, Echo State Network, Reservoir Computing, Time Series Prediction, Underwater Acoustic Communication=-15pt
I Introduction
Underwater wireless communication has rapidly grown in importance for numerous ocean monitoring and information exchange applications in civil and military use in recent years [1]. Acoustic technology has also been shown as a useful tool for a wide range of underwater activities and applications. Ocean exploration, scientific data collection, and transmission are some of the most prevalent applications for underwater acoustic (UWA) communications. Furthermore, underwater communications have benefited the maritime sector by making process management and monitoring easier and more efficient [2]. Underwater operations including undersea marine biology study, undersea mining, pipeline laying, underwater maintenance, and geological surveys have fueled the increased demand for underwater channel and environment research [3]. In general, the behavior of the channel has a significant impact on acoustic signal transmission, therefore having a deep understanding of the channel characteristics is critical for implementing an effective underwater communication system [2].
The underwater environment presents a unique set of challenges for wireless communications [4], and UWA channels are widely regarded as one of the most challenging communication media now in use [1]. While low frequencies are excellent for acoustic propagation, the bandwidth available for communication is extremely limited. Furthermore, a UWA channel has low physical link quality and high latency, and it suffers from large multipath delay spread and frequency selective fading [5], making modeling of the UWA channel quite challenging [6, 7, 4, 8, 9]. A typical UWA communication scenario is depicted in Figure 1. As illustrated in the figure, variations in sound velocity, roughness of the ocean bed, multi-path propagation of acoustic signals, and ambient ocean acoustic noises created by aquatic creatures and human activities make it even more challenging to model UWA channel [2, 8].
Many physics-based UWA channel models have been developed. The most commonly used one is the BELLHOP model, which is an open-source beam/ray-tracing model for predicting acoustic pressure fields in the underwater environment [8]. The BELLHOP ray model is an intuitive and straightforward means for modeling sound propagation in the ocean among the various existing mathematical UWA channel models based on ray, normal-mode, and parabolic curve [1]. The majority of these models, however, are based on mathematical assumptions and approximations rather than real underwater communication data. As a result, they do not work well in reality [1].

Machine learning has seen considerable success in fields such as image and voice recognition, language processing, medical diagnosis, and wireless communications. This is due largely to its capacity to learn and intelligently respond to changing and complex operating conditions, such as those found in the UWA communication channel. Specifically, it is shown that modeling the UWA channel by replicating the effect of real water environment characteristics on the channel is effective [1]. However, there has not been much research work done in the domain of UWA communications using machine learning because of the complex nature of the underwater environment and the lack of sufficient and high-quality data. This motivated us to leverage the capability of collecting real-world data of UWA communications from Lake Tahoe in Reno, Nevada, and apply a data-driven approach to modeling the UWA channel using machine learning on the collected datasets [11].
RC is explored to model UWA channel in this paper. It has been shown that RC is capable of modeling dynamical systems [12, 13] and predicting chaos [14]. RC is a time-dependent data processing paradigm influenced by neuroscience [15]. It is a type of recurrent neural network model in which the recurrent component is initialized randomly and subsequently fixed thus incurring less computation and reducing training time [16, 17]. Despite this significant simplification, the recurring element of the model, the reservoir has a huge number of dynamic properties that can be used to solve a wide range of problems [18].

In this work, we seek to leverage these properties of RC and provide a data-driven approach to modeling the UWA communication channel using the Echo State Network (ESN), which is one of the two pioneering RC approaches. Using this approach significantly reduced the implementation complexity and the training time. The contributions of this paper are:
-
1.
A data-driven approach for UWA channel modeling is proposed to take advantage of the real-world experimental UWA datasets and avoid the unrealistic assumptions made by physics-based mathematical models.
-
2.
UWA channel modeling using ESN, an approach of RC and transfer learning have been carried out. Observations and insights are provided based on the experimental results.
-
3.
The effects of different setups of the ESN on the model performance, in terms of the reservoir initialization method, the size of the reservoir, the activation function, and the regression algorithms used at the readout layer, have been examined and suggestions are made to improve the performance of ESN for UWA channel modeling.
-
4.
A novel approach of designing the reservoir using a pre-trained deep learning model as the reservoir has been proposed in this study. Experimental results demonstrate that ESN using pre-trained deep learning models as reservoir outperform the deep learning models for modeling the UWA channel. Although this may not be advantageous if the reservoir of the ESN can be designed properly using randomized weights, it provides a systematic way to set up the reservoir, which is valuable because there does not exist a systematic way for the design of reservoir for diverse real-world applications.
-
5.
Transfer learning using simulated radio frequency (RF) data and Bellhop-based simulated UWA channel data has been performed. It is observed that the performance of transfer learning using RF data is poor because of the significant differences in characteristics between RF channel and UWA channel. On the contrary, the performance of transfer learning using Bellhop-generated data is pretty good, although the Bellhop model itself is a simplified mathematical model and it does not take into account the various uncertainties in a real-world scenario.
The remainder of this paper is structured as follows. Some related works are discussed in Section II. Section III describes the data generation and collection process and gives the description of the datasets used in this work. Section IV discusses RC and the ESN approach used in this experiment. Detailed experimental results and analysis are given in Section VI and observations and insights from the results are discussed. Section VII concludes the paper.
II Related Works
Many mathematical models have been built and used for different purposes in theUWA domain including but not limited to investigating and tracking channel properties, sound propagation characteristics, the behavior of acoustic signals for different transmission frequencies, and the computation of some channel parameters such as the route loss. For example, in [1], a BELLHOP ray model was used to model the UWA channel in order to examine sound propagation characteristics while taking into account the rough nature of sea surfaces and bottoms for various oceanic conditions. It was also used to examine the behavior of an acoustic signal with transmission frequencies in the range of 9K to 90KHz in [19]. Channel properties for an autonomous underwater vehicle (AUV) wireless communication system were mathematically quantified by modeling the UWA channel in [3]. The model was created using the product, signal-to-noise ratio (SNR), and band selection, where represents attenuation and represents ambient noise. In [9], simulations using ray-theory-based multipath Rayleigh underwater channel models for shallow and deep waters are carried out to investigate transmission losses between transceivers, the effects of bit error rate, maximum internode distances for different networks and depths, the effect of weather season, and variability of ocean environmental factors. The authors in [20] proposed a channel model for tracking dynamic UWA channels by using the channel’s correlation as the state-space model in the Kalman filter in order to improve tracking. The authors of [6] calculated the channel route loss, changed the log-distance model to produce a model suitable for an underwater IoT network, and created an empirical channel model for medium-distance UWA channels based on real measurement data. In [21], a non-stationary two-dimensional wideband channel model was designed for UWA communication and evaluated with measurement data.
Various deep learning models have been proposed to model UWA communication channels. For example, in [22], a deep learning network based signal detection was employed for full-duplex cognitive UWA communication with self interface cancellation. Automatic modulation classification of underwater communication signals using a combination of the convolutional neural network (CNN) and LSTM [23]. A similar task has been done in [24] using blind equalization in conjunction with a CNN. Because it is difficult to identify modulation during actual communication due to the complex and unstable nature of UWA communication systems, several machine learning methods were used in [25] to classify the modulation type in their quest to find an efficient link adaptation method based on channel quality of an underwater communications network. In [26], the authors employed a DNN to create a deep learning-based receiver for single carrier communication in a UWA channel utilizing data from the sea. When compared to the traditional channel-estimate based decision feedback equalizer, the DNN based receiver consistently performed better. The authors in [27] also utilized the DNN to estimate channel parameters based on data from the Bellhop Ray model simulation of the UWA environment. When compared to traditional channel estimation methods such as least square and minimum mean square error (MMSE), the DNN outperformed the LS algorithm and is comparable to the MMSE algorithm in terms of bit error rate and normalized mean square error. A deep learning-based UWA orthogonal frequency-division multiplexing (OFDM) communication system was constructed by representing the receiver as a DNN in [28] and [29]. The deep learning UWA communication systems could easily recover the transmitted symbols after training without using explicit channel estimation and equalization. In [30], the authors developed a depth learning-based underwater target recognition approach employing CNN and an extreme learning machine for UWA target classification and recognition.
RC has been applied in wireless communications for predicting wireless channel or state conditions, symbol detection, and measuring the channel SNR. For example, in [31], the performance of an extreme learning machine and an ESN for forecasting wireless channel conditions was compared. These two methods were used to forecast the SNR for single-input single-output systems in both pico-cellular and micro-cellular contexts. For multiple-input multiple-output orthogonal frequency-division multiplexing (MIMO-OFDM) systems, an ESN-based symbol detector was used in [32]. The efficiency of the adopted symbol detector outperforms traditional symbol detectors based on channel estimation methods in terms of BER performance according to simulation results. In [33], a new RC-based detector called windowed ESN was designed for MIMO-OFDM symbol detection. This resulted in significant improvements in interference cancellation and nonlinear compensation, as well as the ability to improve short-term memory fundamentally. The authors in [15] looked at a simplified fading channel model, defined the transmission properties of satellite communication channels, and devised an ESN-based approach for measuring channel SNR. For the categorization of multivariate time series, the authors in [34] applied an unsupervised approach for creating multivariate time series (MTS) representations (also known as reservoir model space). The parameters of a one-step forward predictor that forecasts the future reservoir state rather than the future MTS input were used to create the reservoir model space. The results revealed that RC classifiers are substantially faster and achieve higher classification accuracy. In [35], an ESN was utilized to train an RNN to predict channel state information in a wireless OFDM system, which resulted in a significant decrease in training time, implementation, and computing complexity. RC has also been applied in UWA communications. For example, transfer learning was introduced to ESNs in [36] to develop a channel model that predicts shallow water dynamics. Experimental results showed that transfer learning helped improve the predictions.
In this work, we leveraged the capabilities of RC and deep learning to build a data-driven channel model using a real experimental UWA communications dataset collected from Lake Tahoe under various environmental conditions. Different from the existing works, the goal of this work is to model the UWA channel and be able to perform sequence-to-sequence prediction, i.e., when a sequence of transmitted data is fed into the model as input, the corresponding sequence of data is expected to be received at the receiving end of the UWA channel will be predicted. The obtained UWA channel model in this work would be very useful as a candidate plugin module when large scale simulations of UWA communications are needed, or a large amount of data need to be generated for UWA channels with high fidelity while it is difficult to obtain that kind of data from physical underwater data collections.
III UWA Dataset
III-A Real-World UWA Communications Data Collection
Underwater communication testbeds were built to collect the UWA communications data for training and further evaluate the developed learning-based channel modeling. To fully study the developed technique, a series of experiments have been conducted. It includes the lab-based experiment and the open-water test and the set-ups for these experiments are as summarized in the block diagrams in Figure 2. In the lab-based experiment, a water tank was used. For the open-water experiment, the experiments were carried out at Lake Tahoe. Lake Tahoe, as seen in Figure 3, is a large freshwater lake in the Sierra Nevada Mountains that straddles the California-Nevada state line. According to Wikipedia, it is the largest alpine lake in North America and at a maximum depth of feet (501 meters), it is the second deepest lake in the United States. Lake Tahoe is also said to be the th deepest lake in the world, and the fifth deepest in average depth. This work does not target sea or ocean environment which has its unique characteristics such as the ocean salinity and we recognize that some modifications and refining the model might be needed to properly transfer it for the ocean environment. By taking advantage of our access to Lake Tahoe, a very large lake that has a lot of similar characteristics as the undersea environment, such as waves, aquatic-life disturbances, etc. the developed model may be served as a reference model to be modified and transferred to ocean environment. In the future, authors plan to further evaluate the developed algorithm in more uncertain environments such as open sea.

For the developed underwater data acquisition system, the transducer is made watertight and other electronic systems including MCU have been packed inside a watertight enclosure tube from BlueRobotics Inc. , the manufacturer. To collect enough underwater communication data, we have deployed transducers at 1-2 meters below the water surface and the distance between Tx and Rx is around 3-5 meters. For each experiment scenario, the source signals were first coded through an oscillator along with the microcontroller (MCU) module. The coded signal strength was then enhanced through an amplifier circuit. It was important to strengthen the signals because the strength of the transmitted signals changes or drops along with the communication distance underwater. To stabilize the signal strength along the transmission path, the amplifier limiter circuit has been used to amplify or enhance the signal strength. Next, the enhanced signals were passed through a quadrature phase shift keying (QPSK) modulation block which outputs continuous signals. The continuous signals were then passed through a raised cosine transmit filter and finally to an ultrasonic ceramic transducer (200LM450) which doubles as the transmitter, as shown in Figure 5. The transmitted I/Q data has been collected at the transmitter and used as input during the training of the channel model. The transmitter and the receiver were placed horizontally apart and at a perpendicular distance below the water level.


At the receiver end, the transmitted continuous signals were captured by the receiver/ ultrasonic ceramic transducers (200LM450). The received I/Q data has been collected at the receiver and used as ground truth for the training of the channel model. The received signals are then reformulated, demodulated, and decoded through the automatic gain control (AGC) circuit as well as a series of filters including the raised cosine receive filter. Using the Schmitt trigger with an analog to digital converter (ADC) circuit, the received signals can be digitalized for further decoding in the workstation. The collected time-series data at the receiver corresponds to the channel impulse response (CIR) of the UWA channel.The receiver setup is as shown in Figure 5. The sampling rate of acquisition used is and the length of each data object is seconds. The sonar working frequency is kHz and the digital signal transfer speed is . Research has shown that deep learning requires large amount of data to obtain good performance [37]. To collect sufficient data to train our deep learning models and for our models to work well, we used a sampling rate of 1 MHz to obtain time-series data containing 60 seconds of I/Q samples.
III-B Data Characteristics
For the purpose of modeling the UWA channel using machine learning and RC models, data collected at the transmitter (ultrasonic ceramic transducer) just before the channel were used as the input to our models while the data collected at the receiver, immediately after the channel were used as the ground truth for training the models. Four different data categories were collected and used to train, evaluate and compare the performances of the trained models. The datasets are described as follows: The first category of data, subsequently referred to as Data 1, were collected using the water tank as the communication channel with no external disturbance. The second category of data collected termed Data 2 were also collected from the lake with no artificial/external disturbance introduced. The third category of data termed Data 3 were also collected from the lake but with the introduction of mild external disturbance.
The disturbance was introduced to create a more realistic underwater scenario. The fourth category, termed Data 4, was also collected from the lake but with the introduction of strong external disturbance to mimic a more chaotic underwater scenario. To model the waves or disturbance in the lake, a vibration platform was used to inject the vibration that generated waves in the lake. For all the categories of data, samples were collected using the same transmission settings and parameters. These data descriptions are summarized in Table I below.
Category | Channel | External Disturbance | Size |
Data 1 | Water Tank | None | 100,000578 |
Data 2 | Lake | None | 100,000578 |
Data 3 | Lake | Mild | 100,000578 |
Data 4 | Lake | Strong | 100,000578 |
IV Reservoir Computing
RC is a time-dependent data processing paradigm influenced by neuroscience [15]. It is a class of recurrent neural network (RNN) model in which the recurrent component is produced randomly and subsequently fixed [16, 17]. The RC methodology builds an RNN with random synaptic weights, dubbed the reservoir, in order to avoid the gradient-descent procedures of the training algorithms for a typical RNN. In [16] and [38], RC shows how an RNN with fixed connectivity can memorize and produce complicated Spatio-temporal sequences. RC has also been demonstrated to be a valuable tool for modeling and predicting dynamic systems [12, 13]. It was demonstrated in [14] that RC is capable of forecasting massive chaotic systems.
Despite this significant simplification, the recurring element of the model (the reservoir) has a huge number of dynamic properties that can be used to solve a wide range of problems [18]. It is also pertinent to state that RC involves less computation and thus has reduced training time dramatically. The normal workflow of solving a task using RC requires handling two key steps: (1) designing a suitable reservoir for the specific task under consideration, and (2) determining a readout function that will adequately map the state of the reservoir to a target output [39]. One of the concerns of RC is that the design is mainly driven by a succession of randomized model-building stages, leaving researchers to rely on a series of trials and errors [40]. In this work, a novel approach to designing the reservoir using a pre-trained deep learning model as the reservoir has been proposed. Although this may not be advantageous if the reservoir of the ESN can be designed properly using randomized weights, it provides a systematic way to set up the reservoir, which is valuable because there does not exist a systematic way for the design of reservoir for diverse real-world applications.
There are two popular RC approaches: the Liquid State Machines (LSM) and ESN [41, 42, 43]. Both architectures attempt to model biological information processing using similar principles [44]. In this work, ESN is chosen because of its close relationship to RNN/LSTM and allows direct performance comparison between these different models.
IV-A Echo State Network
ESNs are dynamical artificial neural networks and belong to the general class of RNN. This approach is prominent and is based on the discovery that if a random RNN has certain algebraic features, then training a linear readout from it is typically enough to provide outstanding performance in practical applications [45]. ESNs have a topology of nonlinear processing elements that is densely interconnected and recurrent, forming a “reservoir” that stores information about the history of input and output patterns. The outputs of these internal processing elements are referred to as the “echo states”. The titles of the echo states stem from the input values echoing throughout the reservoir’s states due to the reservoir’s recurrent nature [16, 46]. These echo states are fed into a memoryless but adaptive, usually linear, readout network, which generates the network output. The architecture of a typical ESN is shown in Figure 6. ESN has the unique property of just training the memoryless readout, whereas the recurrent topology has fixed connection weights. This reduces the complexity of RNN training to simple regression while maintaining the recurrent topology, but at the same time, it imposes significant constraints on the overall architecture that has yet to be resolved [44].

As shown in Figure 6, ESN has three layers; the input layer, the dynamic reservoir, and the output layer. The input weight matrix connects the input layer to the dynamic reservoir. The internal weights of the dynamic reservoir, define the linkages inside the reservoir. The output weight matrix connects the dynamic reservoir to the output layer. Feedback weights are used to feed the output back into the dynamic reservoir. The fundamental structural distinction between an ESN and the conventional RNN is the connectivity of neurons within the dynamic reservoir [32] and one major advantage of RC over regular RNN is that simple regression algorithms may be used to alter output weights [45].
Through the weighted input connections, the input layer of neurons delivers the stimulus to stimulate the reservoir. Through the weighted feedback connections, the output layer of neurons transmits teacher-forced outputs to the reservoir. The reservoir is trained to generate weighted connections from the reservoir to the output based on the input stimuli and feedback from the teacher-forced outputs [35]. Specific values are used to weigh the connections between each layer of neurons and between neurons in the reservoir. Let us consider a recurrent discrete-time network with input units, internal processing elements, also known as nodes and output units. The value of the input unit at time, is , the value of the internal units is and those of the output units are . An matrix, , defines the weights of connections from the input layer to the dynamic reservoir , which is an matrix for connection between the nodes. Also, an matrix, defines the connection from the reservoir nodes or processing elements to the output units. Lastly, , which is an matrix defines the connection weights of the feedback from the output layer to the reservoir [44]. Only the output weights , are computed during training, while the rest of the connection weights are generated randomly and fixed throughout the training and testing stages[35]. In Figure 6, assuming is the input vector at time step , the activations of hidden nodes, also known as the echo states, are updated according to equation (1).
(1) |
where is an hyperbolic tangent activation function of the hidden unit, , and are the matrices of hidden-hidden, input-hidden, and output-hidden connections, respectively [44]. Equation (1) can thus be rewritten as
(2) |
This equation is also known as the state transition equation.
IV-B Training the Readout Layer
As mentioned in Section IV-A, when training the ESN, only the output connection matrix is updated while other connection matrices are kept constant. The training process involves driving the reservoir with the input time series data, to generate the corresponding states, using Equation (2). All the states are then collected into a matrix and the target data (ground truth) are collected into another matrix . The output weights are then computed in closed form using Equation (3) given by
(3) |
where is an identity matrix and is the Tikhonov regularizer which is a fixed positive number used to determine the sensitivity of the system [48]. Since we are dealing with batches of sequence data, Equation (3) becomes
(4) |
where and . The output from the readout layer is also computed using the simple output layer equation given by
(5) |
where is the output nonlinear activation function [45]. The task of training the readout is then reduced to a simple linear regression problem of minimizing the squared error. The regression model minimizes the mean square error between predictions, and the ground truth, , i.e., .
IV-C Echo State Property
In RC, one important criterion that must be satisfied is the echo state condition, especially when working with ESN. In essence, this condition states that the effect of a prior state and a previous input on a future state should fade away or vanish with time. In other words, it is stated that the dynamics of the ESN should be uniquely controlled by the input [44]. The echo state condition is defined in terms of the spectral radius, of the reservoir weight matrix, . Specifically, assuming are the eigenvalues of the reservoir matrix , then its spectral radius, is defined as
(6) |
The echo state condition is satisfied if is scaled such that its spectral radius is close to or inferior to as expressed in Equation (7) given by
(7) |
The spectral radius is the largest absolute eigenvalue of the matrix and is a crude way of measuring how much memory the reservoir can hold, with small values indicating a short memory and large values indicating a longer memory, up to the point of over-amplification, when the echo state condition no longer holds [49, 47]. Another important consideration with the echo state property is that it must guarantee that the memory capacity is not reduced to zero since one of the advantages of using RNNs is their capacity to have a memory of the inputs. The reservoir output should be able to recreate the input with a steps delay [46]. Spectral radius of W can be specified by a user and used backward in the design or initialization of the reservoir so that we could guarantee the performance. It is a design choice, and it does not depend on data.

One of the foci of this paper is using ESN for modeling the UWA communication channel. A known challenge with working with ESN is that there is no predefined systemic way of designing the dynamic reservoir for a specific application or use case. As of today, the design of the ESN relies heavily on the selection of the spectral radius. A suggested method of producing an appropriate reservoir, according to [39], is to optimize their dynamics for the range of activities to be expected, such that the readout layer may simply extract the information it requires. The dynamic reservoir, alongside and are usually randomly generated at network initialization [45], [41] and stay fixed or left untrained during the network’s lifetime [34]. However, there are many different weight matrices with the same spectral radius that can be generated, and they don’t all have the same performance with respect to the mean square error (MSE) or mean absolute percentage error (MAPE) for functional approximation. According to [44], different randomizations with the same spectral radius perform differently on the same problem. To improve the performance of ESNs, many simple methods have been proposed. Some of these methods include increasing non-linearity by augmenting non-linear expansion with polynomial functions of reservoir activities, increasing reservoir size, averaging predictions from many reservoirs, introducing delay lines into the readout system, providing neurons with a diversity of time constants, and having the reservoir adapt to input statistics via intrinsic plasticity [50]. Also, the behavior of the reservoir, according to [18] can be controlled by modifying the spectral radius , the sparsity (or the probability of connection), which is the percentage of non-zero connections, and the number of hidden units in the reservoir, .
In this study, we investigated the effect of different initialization approaches for the reservoir with respect to the spectral radius, sparsity, and the number of hidden units and then compared their performances in terms of the MAPE. We also propose the use of pre-trained deep learning models e.g. LSTM, DNN, etc. into the ESN architecture to act as the reservoir instead of using some randomized vectors as the dynamic reservoir, thus leaving us with a modified ESN architecture such as one shown in Figure 7. The pre-trained deep learning models were trained with the same datasets. This architecture makes use of the weights from the pre-trained model and as with standard architecture of the ESN, only the output connections are modified during the learning process i.e training only occurs in the readout layer. The two established deep learning models experimented with are the deep neural network (DNN) and the long short term memory (LSTM). The LSTM is a variant of RNN with the ability to learn input data with long-term dependencies [51] while the DNN is a deep learning architecture with more than one hidden layer that are fully connected [52].
V Transfer Learning
Transfer learning can be defined as the process of improving learning in a new task or domain by transferring knowledge acquired from another related task or domain that has already been trained. In other words, transfer learning relates to situations in which what has been learned in one domain is used to improve generalization in another domain [53]. In transfer learning, a base model is first trained on a base dataset and task. The learned features are then transferred to the target domain to be trained on a target dataset and task. This method is more likely to succeed if the features are generic, i.e., applicable to both the base and target domains, rather than being specific to the base domain. This approach is known as the “pre-trained model approach”. The purpose of transfer learning is to quickly obtain the learning model by leveraging commonalities between tasks. In this study, transfer learning for RC has been considered. Specifically, transfer learning using simulated RF data and Bellhop based simulated UWA channel data has been performed. In this case, the base domain is RF wireless communications or simulated UWA communications, and the target domain is real-world UWA communications.
VI Results and Discussion
In this section, we provide the experimental set up such as the hyperparameters’ settings in section VI.A. Then the detailed performance of ESN is evaluated in section VI.B. Performance evaluations of modified ESN and transfer learning are carried out in section VI.C and section VI.D, respectively. Together they provide a comprehensive performance analysis of the proposed schemes.
VI-A Computing Experimental Setup
Multiple sets of experiments were carried out and model performances are evaluated using the MAPE. The MAPE value is the average of all the absolute percentage errors in predictions. To compute the MAPE, percentage errors are added together without respect to sign, as shown in equation (8) where is the actual value and is the predicted value.
(8) |
MAPE provides a fairly intuitive interpretation in terms of relative error when evaluating regression problems, and it is preferable in the assessment since it provides the error in terms of percentages, avoiding the problem of positive and negative errors canceling each other out. The better the prediction, the smaller the MAPE and values recorded in the result tables below are the average value from multiple repeated experiments.
Table No. | DNN | LSTM | ESN | PMESN |
IV, XII | = 6 | = 6 | ||
XIII, XV | ||||
XVI, XVIII | ||||
We started the set of experiments by training popular deep learning models such as DNN and LSTM and compare their performance with a typical setup of an ESN. The experimental results are given in Tables IV, XII, XIII, XV, XVI, XVIII. The hyper-parameters used for each of the models built in these experiments are summarized in Table II, where the first column gives the index of the tables that contain the experimental results using these hyper-parameters. Here is the learning rate, is the number of hidden layers, is the number of nodes per hidden layer, is the batch size and is the number of epoch used.
In order to study the effects of various hyper-parameters on ESN performance, a set of in-depth ESN experiments have been done. The experimental results are given in Tables VII, VIII, IX, X, XI. The hyper-parameters used for the different setups of the ESNs are listed in Table III. The first row gives the index of the tables that contain the experimental results using these hyper-parameters for ESNs. In the first column, is the initialization method used, is the spectral radius, is the reservoir size, is the activation function used, and is the regression method used for training the output layer. is Xavier initialization method, is normalized Xavier initialization (gloriot) method, is HE initialization method, is ridge regression, is linear regression, is lasso regression and is hyperbolic tangent.
VI-B Performance Evaluation of the ESN
Two major categories of experiments, with several sub-experiments, were carried out here to build and evaluate ESN models on different data quality and with different network setups. These experiments include (1) comparison of performance between DL models and ESN; and (2) in-depth performance evaluation of ESN with different setups.
VI-B1 Comparison of performance between DL models and ESN
Deep learning models and ESNs were built, trained, and evaluated using Data 3 (data from experiments in Lake Tahoe with mild disturbance) and Data 4 (data from experiments in Lake Tahoe with strong disturbance, hence has the poorest quality). Model parameters used are provided in Table II. Results from these experiments were recorded in Table IV. It is observed that ESN outperforms DNN and LSTM, as expected. Moreover, it is clear that ESN outperforms DNN and LSTM in MAPE by a big margin when the dynamical system that we try to model becomes more chaotic, and the data quality deteriorated. The average times taken to train each of the models using the same data size on a GPU using a single core are also recorded. It is observed that ESN takes a much shorter time to train compared to DNN and LSTM. For example, it takes an average of 2.3 hours to train the LSTM model and an average of 6.7 minutes to train the ESN model on the same dataset.
MAPE(%) | Training Time | |||
Model | Data size | Data 3 | Data 4 | (secs) |
DNN | 100,000578 | 4.19 | 56.54 | 1719 |
LSTM | 100,000578 | 3.14 | 52.10 | 8226 |
ESN | 100,000578 | 2.15 | 15.96 | 407 |
VI-B2 In-depth performance evaluation of ESN with different setups
This set of experiments seeks to explore and investigate the effect of different ESN setups, using hyper-parameters listed in Table III, on the performance of our ESN channel models. The experiments carried out here are itemized below.
-
a)
Using different Reservoir Initialization Methods: As mentioned in Section IV, RC is not principled enough as there is no systematic way of defining the dynamic reservoir. The reservoir is usually randomly generated at network initialization. In addition to randomly initializing the reservoir, we explored the option of using some other weight initialization approaches used in general deep learning [54, 55, 56] and investigated how they affect the performance of the ESN. For the four categories of data, we ran experiments using the random, Xavier, normalized Xavier (gloriot), and HE initialization methods while keeping all other parameters constant. The model performances when each of the initialized methods were used are recorded in Table V. This experiment, however, was carried out without taking into consideration the echo state property.
TABLE V: Model Performance Comparison for Different Initialization Approaches MAPE(%) Random Xavier Norm. Xavier HE Data 1 32.54 6.11 14.82 9.57 Data 2 23.06 5.83 12.08 10.79 Data 3 12.70 2.15 6.24 5.92 Data 4 46.20 15.96 27.31 32.19 -
b)
Enforcing the Echo State Condition: An investigation into the spectral radii of the matrices generated by each of the initialization approaches used above revealed, as shown in TableVI, that only the Xavier initialized reservoir matrix satisfied the echo state property at all instances. Both Gloriot and HE initialization methods have spectral radius, greater than one for different reservoir sizes. In this experiment, we ensured that the echo state property is satisfied by normalizing the matrices such that the spectral radius is less than one and the results from this experiment are as recorded in Table VII.
TABLE VI: Spectral Radius Check for different Proposed Initialization Approaches Spectral Radius Reservoir Size, N Xavier Norm. Xavier HE 50 0.63 1.11 1.46 100 0.59 1.03 1.41 150 0.60 1.05 1.49 300 0.61 1.02 1.47 578 0.60 1.09 1.44 600 0.58 1.03 1.44 1200 0.59 1.01 1.44 2400 0.59 1.02 1.43 TABLE VII: Model Performance Comparison for Different Initialization Approaches with normalized Spectral Radius [] MAPE(%) Random Xavier Norm. Xavier HE Data 1 20.24 6.05 6.02 5.91 Data 2 14.19 5.56 5.87 6.41 Data 3 2.86 2.16 2.17 2.17 Data 4 16.2 15.80 16.59 16.25 -
c)
Using different Spectral Radii: Here, we varied the spectral radius such that it is greater than 0 but less than 1 at an incremental step of 0.1, ensuring the echo state property is still obeyed while keeping every other parameter (the initialization method, the number of nodes in the reservoir and the activation function used) constant. The MAPE values are as recorded in Table VIII. Figure 8 is a graphical plot of the MAPE values for different spectral radii across all the data categories.
TABLE VIII: Model Performance Comparison for different Spectral Radius MAPE(%) Data 1 Data 2 Data 3 Data 4 0.1 6.00 5.98 2.15 16.37 0.2 5.87 7.49 2.15 16.41 0.3 6.02 5.60 2.16 16.35 0.4 6.00 5.72 2.15 16.64 0.5 5.87 7.19 2.15 15.96 0.6 5.75 5.65 2.16 16.14 0.7 6.04 6.30 2.15 16.47 0.8 6.06 5.61 2.16 16.17 0.9 6.02 5.91 2.15 16.12 Figure 8: MAPE Plot for Different Spectral Radius -
d)
Using different Reservoir Sizes: Next, we investigated the effect of varying the number of nodes in the dynamic reservoir, on the performance of the model while keeping other parameters (spectral radius, initialization method, and the activation function) constant. Results from this experiment are as recorded in Table IX. Figure 9 shows a graphical plot of the MAPE values for the different reservoir sizes across all the data categories.
TABLE IX: Model Performance Comparson for Different Sizes of the Reservoir MAPE(%) Reservoir Size, N Data 1 Data 2 Data 3 Data 4 50 17.02 26.17 3.29 16.06 100 10.69 15.55 2.56 16.10 150 8.68 11.91 2.32 16.10 300 6.73 7.77 2.17 16.03 578 5.70 5.86 2.15 15.96 600 5.91 5.97 2.21 16.05 1200 5.99 6.08 2.54 16.15 2400 6.34 12.13 2.87 17.43 Figure 9: MAPE Plot for Different Reservoir Sizes, N -
e)
Using different Activation Functions: Instead of the hyperbolic tangent () used in the state transition equation (2), the effect of using some other activation functions are investigated. In addition to tanh, we also experimented with ReLU and Sigmoid activation functions. Using the Sigmoid and ReLU activation functions required normalizing the input data since the sigmoid function only gives an output between to and that of the ReLU function is between and . Then the outputs of the activation functions are ‘unnormalized’. Results from these experiments are recorded in Table X.
TABLE X: Model Performance Comparison for Different Activation Functions MAPE(%) Data Size tanh ReLU Sigmoid Data 1 100,000578 6.05 6.66 7.14 Data 2 100,000578 5.56 5.79 6.02 Data 3 100,000578 2.16 2.27 2.38 Data 4 100,000578 15.96 17.21 18.11 -
f)
Using different Regression Models: In minimizing error at the readout layer, the only layer where training occurs in the network, we also investigated the effect of three different regression models on the performance of the network - ridge, linear, and lasso regression, while keeping every other parameter constant. The results from these experiments are recorded in Table XI. A sample plot of ESN predicted data is as shown in Figure 10.
TABLE XI: Model Performance Comparison for Different Regression Model Used at the Readout Layer MAPE(%) Ridge Linear Lasso Data 1 5.91 6.00 5.86 Data 2 5.89 5.97 6.01 Data 3 2.15 2.18 2.21 Data 4 15.96 16.12 16.81 Figure 10: Sample Plot of ESN Predicted Data
Summary of results in Section VI-B: Observations from these experimental results are summarized below:
-
•
ESN models perform better than the popular deep learning models such as DNN and LSTM, especially when the data quality becomes poor as seen in Table IV when we used the dataset with the poorest quality. It is observed that ESN takes a much shorter time to train compared to DNN and LSTM, as expected.
- •
-
•
RC is very robust against different spectral radii, activation functions, and regression methods used for training the output layer. For instance, there were no significant differences in performance of the models in Tables VIII, X and XI when different spectral radius, activation function and regression model respectively were used.
-
•
When the size of the reservoir, that is the number of nodes in the reservoir, matches the size of the input sequence into an ESN model, the model performs better. For example, in Table IX the models gave the best performance, across all data categories, when the reservoir size was set to 578, which corresponds to the size of the sequence of the input data used in these experiments.
VI-C Performance of the Modified ESN
In a bid to improve the performance of the ESN, we experimented with the use of pre-trained deep learning models as a replacement for the reservoir, as highlighted in Figure 7, instead of randomized using Data 3 and examine how it affects the performance of the network. Specifically, we used pre-trained DNN (PDNN) and LSTM (PLSTM), deep learning models, for these experiments, and the results are shown in Table XII.
Data 3 | |
Model | MAPE(%) |
DNN | 4.19 |
LSTM | 3.14 |
ESN | 2.15 |
ESN + PDNN | 3.63 |
ESN + PLSTM | 2.90 |
When compared to the performances of the deep learning models, it is observed that the ESN with pre-trained deep models as reservoir did show some improvements in performance. There was a 13.42% improvement in the performance of the DNN and a 7.64% improvement in that of LSTM but did not do better than the conventional ESN model. Thus, using pre-trained deep models as a replacement for the reservoir in the ESN might not be necessary.
VI-D Transfer Learning Performance
Two categories of experiments were carried out to investigate if transfer learning would help improve the performance of the ESNs as stated in Section V. Under each category, three experiments were carried out.
VI-D1 Model trained using RF data
This experiment seeks to leverage trained models from the RF domain using simulated RF data and transferring them into real UWA communication domains. RF datasets were generated using MATLAB simulation of an additive white gaussian noise (AWGN) channel. Just as with UWA data generation, signal bits were generated and passed through a quadrature phase-shift keying (QPSK) block before sending to a raised cosine transmit filter (RCTx) block. The output of the RCTx block was first saved and then fed to the AWGN block as the transmitted signal. The output of the AWGN channel is also saved as the received signal used for training our models. These datasets were used in the sub-experiments itemized below.
-
a)
Training and testing models with RF data: Deep learning models and an ESN were trained to model the AWGN channel and tested directly using solely the generated RF data. These deep models were saved and then used as the reservoir in training the readout layer of our proposed ESN architecture and also evaluated using the RF testing data. The results from these experiments are as recorded in Table XIII.
TABLE XIII: Model Performance when models were trained and evaluated using simulated RF Data RF Data Model Data size MAPE(%) DNN 100,000578 6.20 LSTM 100,000578 5.00 Echo ESN 100,000578 3.50 ESN + RFP-DNN 100,000578 5.34 ESN + RFP-LSTM 100,000578 4.82 -
b)
Using RF pre-trained models and testing on Data 3: Models from the experiment above were saved and then evaluated directly using Data 3 test dataset. The results from this experiment are as shown in Table XIV where RFP-DNN, RFP-LSTM, and RFP-ESN represent RF-Pretrained DNN, RF-Pretrained LSTM, and RF-Pretrained ESN models respectively.
TABLE XIV: Model Performance when RF-Pretrained Models from Table XIII were evaluated directly with Data 3. Data 3 Model Data size MAPE(%) RFP-DNN 100,000578 17.21 RFP-LSTM 100,000578 10.78 RFP-ESN 100,000578 68.87 ESN + RFP-DNN 100,000578 18.13 ESN + RFP-LSTM 100,000578 12.51 -
c)
Refining RF pre-trained models and testing on Data 3: Next, we fine-tuned the pre-trained models from sub-experiment i by re-training the models with Data 3 training data and re-evaluating the resultant model with Data 3 test data. Results from this experiments are as recorded in Table XV.
TABLE XV: Model Performance when RF-Pretrained models from Table XIII were re-trained with Data3 and evaluated with Data 3. Data 3 Model Data size MAPE(%) BHP-DNN 100,000578 10.43 BHP-LSTM 100,000578 6.32 BHP-ESN 100,000578 23.21
VI-D2 Using Bellhop Simulated data
The next set of experiments seeks to import pre-trained models built from simulated datasets generated using the BELLHOP ray mathematical model [1]. Although the Bellhop model is a well established mathematical model for UWA channel modeling [19], it failed to perform well in a real-world scenario. When used to predict the received signals, with real UWA data from Data 3 as the input to the model, the average MAPE value is 57.30%. As with the RF data, these simulated Bellhop datasets are used in the experiments itemized below.
-
a)
Training and testing models with Bellhop data: Deep learning models and an ESN were trained and tested using these BELLHOP generated data. The results from these experiments are as recorded in Table XVI.
TABLE XVI: Model performance when models were trained and evaluated using Bellhop generated data. Bellhop Generated Data Model Data size MAPE(%) DNN 100,000578 4.88 LSTM 100,000578 3.62 ESN 100,000578 3.14 ESN + BHP-DNN 100,000578 4.02 ESN + BHP-LSTM 100,000578 3.46 -
b)
Using Bellhop pre-trained models and testing on Data 3: The pre-trained models, from Table XVI, were saved and then evaluated directly using Data 3 test data. These deep models were saved and then used as the reservoir in training the readout layer of our proposed ESN architecture and also evaluated using the bellhop testing data. The results from these experiments are as shown in Table XVII where BHP-DNN, BHP-LSTM, and BHP-ESN mean Bellhop-Pretrained DNN, Bellhop-Pretrained LSTM, and Bellhop-Pretrained ESN, respectively.
TABLE XVII: Model Performance when Bellhop-Pretrained models from Table XVI are evaluated directly with Data 3. Data 3 Model Data size MAPE(%) BHP-DNN 100,000578 6.26 BHP-LSTM 100,000578 4.92 BHP-ESN 100,000578 3.98 ESN + BHP-DNN 100,000578 5.81 ESN + BHP-LSTM 100,000578 4.10 -
c)
Refining Bellhop pre-trained models and testing on Data 3: Pre-trained models from Table XVI were fine-tuned by re-training the models with Data 3 training data and re-evaluating the resultant model with Data 3. Results from these experiments are as recorded in Table XVIII.
TABLE XVIII: Model Performance when Bellhop-Pretrained models from Table XVI were re-trained with Data 3 and evaluated with Data 3 testing data. Data 3 Model Data size MAPE(%) BHP-DNN 100,000578 4.91 BHP-LSTM 100,000578 3.23 BHP-ESN 100,000578 2.51
Summary of Results of Section VI-D:
-
•
Transfer learning does not perform well when transferring models trained with very different data or domains. For example, using models built with simulated RF data and transferring directly to underwater data did not give a good result as observed in Table XIV.
-
•
On the other hand, transfer learning tends to perform reasonably well when the base domain is related to the target domain. For instance, when models trained with simulated Bellhop data were transferred and evaluated directly on underwater data, there was a notable improvement in performance as seen in Table XVII.
- •
VII Conclusion
Two key objectives were pursued in this work. The first objective is to mitigate the approximations and unrealistic assumptions made by the mathematical models for UWA channel models by providing a data-driven approach to UWA channel modeling. To achieve this first goal, data generation and collection experiments were carried out in a water tank and in Lake Tahoe under different levels of disturbances. Then the obtained datasets were used to train deep learning models (DNN and LSTM) as well as the ESN to obtain models with high fidelity in a real-world scenario. It is observed that ESN performs better than DNN and LSTM in terms of prediction accuracy and has much less computational cost in terms of training time. The performance gap becomes larger when the data quality gets worse (more chaotic data), which is in agreement with previous studies in the literature that RC is very effective in modeling chaotic dynamical systems. The second objective is to examine how RC performs under various setups including different initialization methods of the reservoir, various spectral radii, the reservoir size, the activation function used, and the regression method used for training the output layer. It is shown that ESN is quite robust to these different settings as long as the spectral radius satisfy the echo state property. It is also observed that when the size of the reservoir matches the size of the input sequence, the ESN model performs better. Furthermore, the performance of transfer learning in ESN is also evaluated. Although Bellhop mathematical model has a very poor performance by itself, when ESN models trained with simulated Bellhop data were transferred and evaluated on real-world underwater data, there was a noticeable improvement in ESN performance. Lastly, a modified ESN implementation was examined where the reservoir of the ESN is replaced with a pre-trained deep learning model. Though this approach gives a slightly better performance when compared to pre-trained deep learning models, it is not as efficient as the conventional ESN.
In this study, the trained ESN model, using real-world UWA communications data, is a data-driven black box model that performs sequence-to-sequence or point-to-point prediction very well with time-series I/Q data as the input to the model. It could be used to carry out large scale simulations of UWA communications or obtain a large amount of high fidelity data when it is difficult to obtain that kind of data from physical underwater data collections. While it provides a more realistic model when compared to mathematical models, it may also be over specified due to the peculiarities of the measurement environment. It could also be generalized to model a network setting with multiple transceivers or could be refined to model the ocean environment or any other environment with time-series data. For instance, we can generalize the developed channel model to a network setting with multiple transmitting-receiving (Tx/Rx) pairs, as long as the multiple Tx/Rx pairs are orthogonal. If we consider multiple simultaneous transmissions in non-orthogonal UWA channels, new channel model must be trained based on new datasets that capture the mutual interference. The model can also be trained in the reverse direction to recover the transmitted signals based on the received signals, which may be an alternative implementation for the software-based receiver. This is one of our future efforts.
Acknowledgment
This research work is supported in part by the U.S. Army Research Office (ARO) under agreement number W911NF-20-2-0266 and the U.S. Office of the Under Secretary of Defense for Research and Engineering (OUSD(R&E)) under agreement number FA8750-15-2-0119. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright notation thereon. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of the U.S. Army Research Office (ARO) or the Office of the Under Secretary of Defense for Research and Engineering (OUSD(R&E)) or the U.S. Government.
References
- [1] R. Jiang, S. Cao, C. Xue, and L. Tang, “Modeling and analyzing of underwater acoustic channels with curvilinear boundaries in shallow ocean,” in 2017 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), 2017, pp. 1–6.
- [2] A. Petroni and M. Biagi, “Underwater acoustic channel modeling in harbors,” in OCEANS 2019 - Marseille, 2019, pp. 1–4.
- [3] H. P. Yoong, K. B. Yeo, K. T. K. Teo, and W. L. Wong, “Modeling of acoustic channel for underwater wireless communication system in auv application,” in 2012 UKSim 14th International Conference on Computer Modelling and Simulation, 2012, pp. 603–607.
- [4] M. W. Khan, Y. Zhou, and G. Xu, “Modeling of acoustic propagation channel in underwater wireless sensor networks,” in The 2014 2nd International Conference on Systems and Informatics (ICSAI 2014), 2014, pp. 586–590.
- [5] N. Morozs, W. Gorma, B. T. Henson, L. Shen, P. D. Mitchell, and Y. V. Zakharov, “Channel modeling for underwater acoustic network simulation,” IEEE Access, vol. 8, pp. 136 151–136 175, 2020.
- [6] H.-K. Lee and B. Lee, “An underwater acoustic channel modeling for internet of things networks,” Wireless Personal Communications, vol. 116, 02 2021.
- [7] H. Yu, X. Li, and G. Li, “Underwater acoustic channel tracking based on parameter fault detection algorithm,” in 2019 IEEE 8th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), 2019, pp. 469–472.
- [8] M. Stojanovic and J. Preisig, “Underwater acoustic communication channels: Propagation models and statistical characterization,” IEEE Communications Magazine, vol. 47, no. 1, pp. 84–89, 2009.
- [9] M. C. Domingo, “Overview of channel models for underwater wireless communication networks,” Phys. Commun., vol. 1, no. 3, p. 163–182, Sep. 2008. [Online]. Available: https://doi.org/10.1016/j.phycom.2008.09.001
- [10] Y. Luo, L. Pu, M. Zuba, Z. Peng, and J.-H. Cui, “Cognitive acoustics: making underwater communications environment-friendly,” in Conference: In Proceedings of the International Conference on Underwater Networks & Systems (WUWNet), 01 2014.
- [11] O. Onasami, D. Adesina, and L. Qian, “Underwater acoustic communication channel modeling using deep learning,” in The 15th International Conference on Underwater Networks & Systems, ser. WUWNet’21. New York, NY, USA: Association for Computing Machinery, 2021. [Online]. Available: https://doi.org/10.1145/3491315.3491323
- [12] F. Triefenbach, A. Jalalvand, B. Schrauwen, and j.-p. Martens, “Phoneme recognition with large hierarchical reservoirs.” in Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010, 01 2010, pp. 2307–2315.
- [13] L. Appeltant, M. Soriano, G. Van der Sande, J. Danckaert, S. Massar, J. Dambre, B. Schrauwen, C. Mirasso, and I. Fischer, “Information processing using a single dynamical node as complex system,” Nature communications, vol. 2, p. 468, 09 2011.
- [14] J. Pathak, B. Hunt, M. Girvan, Z. Lu, and E. Ott, “Model-free prediction of large spatiotemporally chaotic systems from data: A reservoir computing approach,” Physical Review Letters, vol. 120, 01 2018.
- [15] T. Liu, B. Sun, Z. Li, and Z. Dou, “Research on snr prediction technology based on reservoir computing in wireless transmission of satellite communication system,” in 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Oct 2017, pp. 1–5.
- [16] M. Lukosevicius and H. Jaeger, “Reservoir computing approaches to recurrent neural network training. computer science review 3, 127-149,” Computer Science Review, vol. 3, pp. 127–149, 08 2009.
- [17] S. Scardapane and D. Wang, “Randomness in neural networks: An overview,” Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol. 7, 01 2017.
- [18] F. M. Bianchi, S. Scardapane, S. Lokse, and R. Jenssen, “Reservoir computing approaches for representation and classification of multivariate time series,” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 5, pp. 2169–2179, 2021.
- [19] S. Gul, S. S. H. Zaidi, R. Khan, and A. B. Wala, “Underwater acoustic channel modeling using bellhop ray tracing method,” in 2017 14th International Bhurban Conference on Applied Sciences and Technology (IBCAST), 2017, pp. 665–670.
- [20] Q. Huang, W. Li, W. Zhan, Y. Wang, and R. Guo, “Dynamic underwater acoustic channel tracking for correlated rapidly time-varying channels,” IEEE Access, vol. 9, p. 50485–50495, 2021. [Online]. Available: http://dx.doi.org/10.1109/ACCESS.2021.3069336
- [21] X. Zhu, C.-X. Wang, and R. Ma, “A 2d non-stationary channel model for underwater acoustic communication systems,” 2021.
- [22] J. Wang, S. Ma, Y. Cui, H. Sun, M. Zhou, B. Wang, J. Li, and L. Liu, “Signal detection for full-duplex cognitive underwater acoustic communications with sic using model-driven deep learning network,” in 2019 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), 2019, pp. 1–6.
- [23] D. Li-Da, W. Shi-Lian, and Z. Wei, “Modulation classification of underwater acoustic communication signals based on deep learning,” in 2018 OCEANS - MTS/IEEE Kobe Techno-Oceans (OTO), 2018, pp. 1–4.
- [24] C. N. Marcoux, B. Chandna, D. Egnor, and B. Blair, “Blind equalization and automatic modulation classification of underwater acoustic signals,” Proceedings of Meetings on Acoustics, vol. 35, no. 1, p. 055003, 2018. [Online]. Available: https://asa.scitation.org/doi/abs/10.1121/2.0000952
- [25] M. Alamgir, M. N. Sultana, and K. Chang, “Link adaptation on an underwater communications network using machine learning algorithms: Boosted regression tree approach,” IEEE Access, vol. 8, pp. 73 957–73 971, 2020.
- [26] Y. Zhang, J. Li, Y. V. Zakharov, J. Li, Y. Li, C. Lin, and X. Li, “Deep learning based single carrier communications over time-varying underwater acoustic channel,” IEEE Access, vol. 7, pp. 38 420–38 430, 2019.
- [27] R. Jiang, X. Wang, S. Cao, J. Zhao, and X. Li, “Deep neural networks for channel estimation in underwater acoustic ofdm systems,” IEEE Access, vol. 7, pp. 23 579–23 594, 2019.
- [28] Y. Zhang, J. Li, Y. Zakharov, X. Li, and J. Li, “Deep learning based underwater acoustic ofdm communications,” Applied Acoustics, vol. 154, pp. 53 – 58, 2019. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0003682X18307400
- [29] Y. Zhang, J. Li, Y. Zakharov, D. Sun, and J. Li, “Underwater acoustic ofdm communications using deep learning,” in The 2nd Franco-Chinese Acoustic Conference (FCAC) (31/10/18), October 2018. [Online]. Available: https://eprints.soton.ac.uk/426097/
- [30] H. Gang, K. Wang, Y. Peng, M. Qiu, J. Shi, and L. Liu, “Deep learning methods for underwater target feature extraction and recognition,” Computational Intelligence and Neuroscience, vol. 2018, pp. 1–10, 03 2018.
- [31] M. B. Stojanovic, N. M. Sekulovic, and A. S. Panajotovic, “A comparative performance analysis of extreme learning machine and echo state network for wireless channel prediction,” in 2019 14th International Conference on Advanced Technologies, Systems and Services in Telecommunications (TELSIKS), Oct 2019, pp. 356–359.
- [32] S. Mosleh, L. Liu, C. Sahin, R. Zheng, and Y. Yi, “Brain-inspired wireless communications: Where reservoir computing meets mimo-ofdm,” IEEE Transactions on Neural Networks and Learning Systems, vol. PP, pp. 1–15, 12 2017.
- [33] Z. Zhou, L. Liu, and H.-H. Chang, “Learning for detection: Mimo-ofdm symbol detection through downlink pilots,” IEEE Transactions on Wireless Communications, vol. 19, no. 6, pp. 3712–3726, June 2020.
- [34] F. M. Bianchi, S. Scardapane, S. Løkse, and R. Jenssen, “Reservoir computing approaches for representation and classification of multivariate time series,” 2020.
- [35] W. Danesh, C. Zhao, B. T. Wysocki, M. J. Medley, N. N. Thawdar, and Y. Yi, “Channel estimation in wireless ofdm systems using reservoir computing,” in 2015 IEEE Symposium on Computational Intelligence for Security and Defense Applications (CISDA), 2015, pp. 1–5.
- [36] X. Chen, B. T. Nadiga, and I. Timofeyev, “Predicting shallow water dynamics using echo-state networks with transfer learning,” 2021.
- [37] N. Thompson, K. Greenewald, K. Lee, and G. Manso, “The computational limits of deep learning,” 07 2020.
- [38] D. Sussillo and L. Abbott, “Generating coherent patterns of activity from chaotic neural networks,” Neuron, vol. 63, pp. 544–557, 2009.
- [39] W. Singer, The Cerebral Cortex: A Delay-Coupled Recurrent Oscillator Network?, K. Nakajima and I. Fischer, Eds. Singapore: Springer Singapore, 2021. [Online]. Available: https://doi.org/10.1007/978-981-13-1687-6_1
- [40] A. Rodan and P. Tino, “Minimum complexity echo state network,” IEEE Transactions on Neural Networks, vol. 22, no. 1, pp. 131–144, Jan 2011.
- [41] D. Kleyko, E. P. Frady, M. Kheffache, and E. Osipov, “Integer echo state networks: Efficient reservoir computing for digital hardware,” 2020.
- [42] W. Maass, T. Natschlager, and H. Markram, “Real-time computing without stable states: A new framework for neural computation based on perturbations,” Neural computation, vol. 14, pp. 2531–60, 12 2002.
- [43] H. Jaeger, “Adaptive nonlinear system identification with echo state networks,” NIPS, 06 2003.
- [44] M. Ozturk, D. Xu, and J. Principe, “Analysis and design of echo state networks,” Neural computation, vol. 19, pp. 111–38, 02 2007.
- [45] M. Cernansky and P. Tino, “Predictive modeling with echo state networks,” in ICANN, 2008.
- [46] A. Rosenberg, “Reservoir computing by felix grezes,” 2014.
- [47] N. Trouvain, L. Pedrelli, T. T. Dinh, and X. Hinaut, “ReservoirPy: an Efficient and User-Friendly Library to Design Echo State Networks,” May 2020, working paper or preprint. [Online]. Available: https://hal.inria.fr/hal-02595026
- [48] C. Gallicchio and A. Micheli, “Architectural and markovian factors of echo state networks,” Neural networks : the official journal of the International Neural Network Society, vol. 24, pp. 440–56, 02 2011.
- [49] H. Palangi, L. Deng, and R. Ward, “Learning input and recurrent weight matrices in echo state networks,” 11 2013.
- [50] A. Laan and R. Vicente, “Echo state networks with multiple readout modules,” bioRxiv, 2015.
- [51] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, pp. 1735–80, 12 1997.
- [52] Y. Bengio, “Learning deep architectures for ai,” Foundations and Trends® in Machine Learning, vol. 2, no. 1, pp. 1–127, 2009. [Online]. Available: http://dx.doi.org/10.1561/2200000006
- [53] M. Inubushi and S. Goto, “Transferring reservoir computing: Formulation and application to fluid physics,” pp. 193–199, 09 2019.
- [54] S. K. Kumar, “On weight initialization in deep neural networks,” 2017.
- [55] H. Li, M. Krcek, and G. Perin, A Comparison of Weight Initializers in Deep Learning-Based Side-Channel Analysis, 10 2020, pp. 126–143.
- [56] W. Boulila, M. Driss, M. Al-Sarem, F. Saeed, and M. Krichen, “Weight initialization techniques for deep learning algorithms in remote sensing: Recent trends and future perspectives,” 02 2021.
![]() |
Oluwaseyi Onasami was born in Nigeria, West Africa. He received his Bachelor’s degree in Electronic and Electrical Engineering at the Obafemi Awolowo University, Nigeria in 2010 and is currently pursuing his masters degree in Electrical and Computer Engineering at the U.S. DOD Center of Excellence in Research and Education for Big Military Data Intelligence (CREDIT Center), Department of Electrical and Computer Engineering, Prairie View A&M University, Texas, USA. His research interests are in the area of big data processing, deep learning and applied artificial intelligence. |
![]() |
Ming Feng received his Bachelor’s degree in communication engineering from Taiyuan University of Technology in 2011 and received M.S. degree in Electrical Engineering from Stevens Institute of Technology in 2016. He received his Ph.D. degree in Electrical and Biomedical Engineering at the University of Nevada, Reno in 2020. Currently, he worked as a research engineer in Hughes Network Systems, LLC. His research interests are in optimal resource allocation with application complex wireless networks, UAV-assisted next generation networks, etc. |
![]() |
Hao Xu received his Master’s degree in Electrical Engineering from Southeast University in 2009, and his Ph.D. degree from the Missouri University of Science and Technology (formerly, the University of Missouri-Rolla), Rolla in 2012. He currently holds an assistant professor position with the Department of Electrical and Biomedical Engineering at the University of Nevada, Reno. His research interests are in AI-based optimization for the wireless network, underwater network, and reliable communication network development. |
![]() |
Mulugeta Haile is a senior aerospace research engineer and a research team leader at the U.S. Army Research Laboratory (ARL), Aberdeen Proving Ground, Maryland. He received his graduate degrees in Electrical and Aerospace Engineering from the University of Florida, Gainesville. Dr. Haile has led several research projects in aerospace mechanics and artificial intelligence (AI). He is the founder of the intelligent mechanics (iMechanics) lab at ARL. He has authored several research papers in top journals, conference proceedings, and book chapters and has over 23 years of research and development experience. His current research interests are in embodied intelligence, scientific machine learning, and AI-assisted creativity. |
![]() |
Lijun Qian (SM’08) is Regents Professor and holds the AT&T Endowment in the Department of Electrical and Computer Engineering at Prairie View A&M University (PVAMU), a member of the Texas A&M University System, Prairie View, Texas, USA. He is also the Director of the Center of Excellence in Research and Education for Big Military Data Intelligence (CREDIT Center). He received BS from Tsinghua University, MS from Technion-Israel Institute of Technology, and PhD from Rutgers University. Before joining PVAMU, he was a member of technical staff of Bell-Labs Research at Murray Hill, New Jersey. He was a visiting professor of Aalto University, Finland. His research interests are in the area of big data processing, artificial intelligence, wireless communications and mobile networks, network security and intrusion detection, and computational and systems biology. |