1. Introduction
To configure, monitor, and control many applications in Internet of Things (IoT) networks, accurate localization of every sensor will be a key enabling technology for private 5G applications and beyond, especially for industries [
1]. Many systems with localization algorithms have been developed by means of wireless sensor networks for both indoor and outdoor environments. To achieve a higher localization accuracy, additional hardware implementations are utilized by most of the existing localization solutions, which increase the cost and considerably limit location-based applications. Consequently, conventional localization methods such as global positioning systems (GPSs) are not suitable, as their direct implementation in IoT networks involves prohibitive demands for sophisticated equipment and substantial energy consumption and cannot meet the required accuracy level, for instance, in industrial setups. These limitations have significantly restricted the practical scalability of IoT networks. To overcome these challenges, massive wireless connections in IoT networks are leveraged for the cooperative location estimation of IoT devices, which is called connectivity-based localization [
2,
3,
4,
5,
6,
7,
8,
9]. This approach not only tackles the limitations of the conventional localization methods but also enhances the localization accuracy, resulting in more robust and energy-efficient localization within IoT networks.
From an algorithmic standpoint, multidimensional scaling (MDS) is a widely employed collection of statistical methods which are extensively utilized to create mappings of items based on their distance, i.e., dissimilarity [
10,
11,
12,
13,
14,
15]. MDS methods are able to represent complex datasets in spaces with lower dimensions; meanwhile, they maintain the dissimilarity relations among the items in the original datasets. Here, the Euclidean distance matrix (EDM) is the key information for implementing the MDS technique, which is constructed using pairwise distance measurements. The EDM serves as a useful description of the point sets and a solid foundation for localization algorithm design due to its effective description of the point sets. However, in real-world environments, EDM data are inevitably prone to contamination due to errors resulting from different sources, such as the measurement resolution, signal quality, network asynchronization, non line-of-sight (NLoS) conditions, and so on. In fact, the contamination in the EDM degrades the localization performance of the MDS, particularly in large-scale IoT networks. Errors in the EDM may lead to inaccurate and imprecise location estimations, which will result in a less reliable and robust connectivity-based localization process. Hence, it is critical to address the EDM contamination issue to ensure accurate and efficient localization in large-scale IoT networks.
To mitigate the effects of noisy measurements on the EDM, denoising techniques have been extensively utilized, and they aim to denoise a noisy EDM by resolving properly designed optimization problems such as semi-definite relaxation [
16,
17] and low-rank tensor completion [
18]. However, they rely on solving complex optimization problems, which may limit their efficiency and effectiveness. Developments in artificial intelligence (AI) methods, particularly neural network (NN)-based denoising methods such as [
19,
20,
21,
22], have become promising in leveraging statistical inference as a novel and potentially more robust alternative for dealing with noisy measurements in EDM-based localization.
The existing NN-based EDM denoising techniques require computations that scale with the square of the total number of nodes since both the input and output of the NN framework are based on pairwise distances. This is due to the combinatorial nature of measuring and generating the distances between node pairs. As the number of nodes increases, the computational complexity grows quadratically, leading to the need for a large training dataset size and extensive NN models.
In addition, the existing NN-based denoising methods may not be suitable for IoT networks due to their resource limitations. Although major research efforts have focused on big data analysis and deep neural networks, it is crucial to consider that most IoT devices face severe limitations in terms of their data acquisition capabilities, computational power, and memory size. Hence, the successful implementation of efficient NN-based algorithms that can handle big data while considering the resource limitations of IoT devices is a critical challenge to address.
In this paper, we propose a novel denoiser for a noisy EDM, referred to as DAE-EDMR (Readers can understand our framework more easily by referring to
Figure 1. Our method is based on the mathematical fact that the EDM measured from
N nodes in a
k-dimensional space must have a rank of at most
. Abstractly, this implies that
N-choose-2 pairwise distances can be rearranged into a
k-dimensional structure. Consequently, if the pairwise distances are not accurately measured, reconstructing these segments would require embedding them into a higher-dimensional space. Using this concept, we optimize the NN model for EDM denoising by utilizing the eigenvalues to capture how the nodes are volumetrically distributed. A detailed explanation is given in
Section 2), which leverages a denoising autoencoder (DAE). To increase the efficiency of the NN operations, we exploit the low-rank property of the EDM, which is bounded to only a
number of eigenvalues, where
k is the dimension of the Euclidean space, independent of the number of sensor nodes [
23]. Leveraging this valuable mathematical observation, we design our NN model by inputting and outputting vectors of the eigenvalues of the noisy and original EDMs, respectively. Through these inputting/outputting rules, the proposed scheme achieves remarkable denoising results, even with a relatively small training dataset. The efficiency of the proposed denoiser is attributed to the utilization of the low-rank property of the EDM, which helps the NN model to establish better inference with limited training data. In fact, the combination of the DAE and EDM is highly attractive due to the complementary features of these two techniques. On the one hand, the DAE is a powerful NN framework for manifold learning, which enables it to effectively capture intricate data patterns and facilitate efficient denoising. On the other hand, the EDM exhibits an extremely-low-rank property, contributing to the dimensionality reduction. Regarding the online complexity, the proposed method requires eigenvalue decomposition (EVD) of the EDM as a pre-processing step for the NN operations.
In addition, we propose a technique called truncated DAE-EDMR, i.e., T-DAE-EDMR, to enhance the robustness of DAE-EDMR to diverse environments conditions, such as changes in the number of nodes, even after NN optimization. In other words, T-DAE-EDMR offers the flexibility to be utilized in scenarios where the number of nodes varies between the training and test phases. The T-DAE-EDMR scheme involves feeding eigenvalues from the noisy EDM to the NN, which are extracted through ()-truncated EVD. In contrast, DAE-EDMR inputs N eigenvalues, making T-DAE-EDMR more versatile in accommodating different environments and simultaneously reducing the online complexity. Hence, T-DAE-EDMR is envisioned to demonstrate its superior effectiveness in environments characterized by a large number of nodes or frequent topology changes, such as vehicle-to-everything (V2X) systems.
To summarize, the main contributions of the proposed algorithms to denoising the noisy EDMs and enhancing the localization accuracy are four-fold:
Minimizing the reconstruction errors for the EDM through the mix-up of mathematical evidence and NNs: We reveal the potential to reconstruct the EDM according to the low-rank property of the ground-truth EDM and the low-dimensional representations of the NN operation, which is robust to various noise models in distance measurements.
Reducing the size of the training dataset and NN model: We develop an efficient NN framework requiring a small-sized training dataset and NN model. This is based on the novel inputting and outputting for the NN model, which consist of the eigenvalues of the noisy EDM and the ground-truth EDM, respectively.
Assisting the existing connectivity-based localization algorithms as a pre-processor: We combine our proposed scheme with the connectivity-based localization technique to validate its utility in various environments. It is verified that these joint frameworks show a superior performance compared to that of the conventional approaches. This achievement is highly remarkable, as we only need to extract the eigenvalues of an EDM, requiring marginal online complexity.
Making the NN model robust to the dynamics of wireless networks: By additionally presenting a modified model of our proposed model, we introduce an NN that can be robust to the variability in wireless networks, e.g., the number of nodes in the test phase is changed after NN optimization.
The remainder of this paper is organized as follows:
Section 2 presents the system model, the problem design, and the method for the proposed algorithms. In
Section 3, we provide numerical results to demonstrate the superiority of our proposed algorithms compared to other schemes even given the aspects of various environments. Finally,
Section 4 presents the concluding remarks.
2. The Proposed Schemes: DAE-EDMR and T-DAE-EDMR
2.1. The System Model and Problem Formulation
Consider a collection of
N nodes in a
k-dimensional Euclidean space,
, where the positions of all
N nodes,
, are randomly distributed. With the knowledge of the positions of
reference nodes, we estimate the positions of the remaining
N-
P nodes. To accomplish this, we will utilize the concept of an EDM denoted by
. It is a symmetric matrix whose
-element can be represented as follows:
where
is the true distance between nodes
i and
j.
To model practical measurements, we first consider three types of random variables due to the environment as follows:
: ranging errors dependent on the signal quality;
: ranging errors due to clock asynchronization;
: non line-of-sight (NLoS) events.
We assume that
,
, and
follow normal, uniform, and Bernoulli distributions, respectively. Hence, we can define the random variable for the bias,
, as follows:
where
is the distance bias in the event of NLoS conditions. Note that
does not follow any known probability distribution, as it is a convolution of three different distributions.
Second, we assume that the distance is measured using a grid consisting of measurement resolutions, which is determined by the ranging configuration, e.g., the time of arrival (ToA). Thus, we define the quantization function to represent the measured distance with a resolution of G, e.g., .
Third, we formulate a function to indicate whether the distance is measured or not, based on the communication capability between nodes
i and
j, as follows:
Thus, the noisy measurement of the distance between the
i-th and the
j-th nodes is defined as follows:
where
is the realization of
. Then, we can define the noisy EDM using the following expression:
Finally, the objective of this paper is to find the denoising function
, which is formulated as
2.2. Method I: DAE-EDMR
In the conventional approaches, the denoising function was set as : , more specifically : , such as in semi-definite relaxation and nonlocal patch tensor-based methods. These techniques require high computational complexity, as they perform iterative matrix multiplication operations while solving high-order optimization problems. In order to overcome this problem, NN-based techniques have recently been proposed. However, the dimension of the input and output data is large, which are the entire elements of the EDM. As a result, they exhibit low performance and require a large amount of training data and a large-sized neural network.
Considering the above problems, we propose a new method, namely DAE-EDMR, to denoise the noisy EDM in an efficient way.
2.2.1. A Denoising Process
Before describing the framework of DAE-EDMR, we will revisit the low-rank property of an EDM [
23].
Property 1. The , corresponding to the points in , is at most .
This is based on the following equality:
where
is the one-vector of size
N. According to rank characteristics, the rank of
is bounded to the summation of the ranks of each term, i.e.,
. This determines the rank of an EDM as extremely low, i.e.,
, regardless of the number of nodes. It implies that it is effective to perform denoising given the small dimension of the (potential) latent space of an NN model, i.e., the similar level of the rank of the EDM, rather than treating the distance information as a whole.
Now, let
and
denote the vectors whose elements are the descending-order eigenvalues of
and
, respectively. Thus, the optimal denoising function
:
can be constructed as follows:
Here, we design a fully connected NN framework denoted by to approximate . To this end, we will define the required terms as follows:
: The dimension of the latent space.
: The weight matrices for encoding and decoding, respectively.
: The bias vectors for encoding and decoding, respectively.
: The activation function for neural networks. At the propagation between the final hidden layer and the output layer, , i.e., an identity function. For other types of propagation between adjacent layers, , i.e., a hyperbolic tangent function. And , where is an arbitrary input vector.
With these terms (For simplicity, the description of the NN model design throughout this article is based on a single hidden layer; however, it is obvious that deeper hidden layers can be made using multiple encoding/decoding function parameters, i.e., and , where I, , and are the depth of the NN model and the i-th encoding and decoding function parameters), we define as the encoding function where the parameter is , i.e., . In addition, we define as the decoding function where the parameter is , i.e., .
Finally, we can define
consisting of the optimal encoding and decoding functions, denoted by
and
, respectively, as follows:
where
M is the number of training datasets. Let
be all of the model parameters, i.e.,
; then, it simultaneously updates every parameter in
at each iteration toward the direction of the steepest descent as follows:
where
is the gradient operator with respect to
and
is the learning rate related to the step size. Through this procedure, we can optimize the fully connected NN model
for denoising the eigenvalues.
Next, in the test phase, let
denote the
j-th denoised eigenvalues with the optimized
, and this can be obtained as follows:
Finally, the denoised EDM
can be reconstructed as follows:
where
is the matrix consisting of
eigenvectors of
.
2.2.2. The Connectivity-Based Localization Process
To obtain an estimate of
, which is denoted by
, based on the classical MDS method, we first define the geometric centering matrix as follows:
where
is the identity matrix
N by
N in size. Next, the estimated centered Gram matrix (GM) is obtained as
Recalling the fact that
, where
is the centered
, we can easily obtain
through the
k-truncated EVD of
. Based on this, we can finally obtain
through a rigid linear transform, i.e., rotation and translation, of
with the pre-knowledge of
, which are the positions of the reference nodes.
2.3. Method II: Truncated DAE-EDMR (T-DAE-EDMR)
This subsection introduces T-DAE-EDMR, which is the relaxed version of DAE-EDMR. From the previous subsection, the low-rank property of the EDM can be used more efficiently from the NN training/test point of view. Assume that the NN model is optimized in a network with N nodes. Here, we consider a scenario where the number of nodes changes at the test phase, which is frequently shown in wireless networks. If the NN model can be used flexibly under this kind of environment change, it will be a more efficient utilization.
For this reason, we newly define the optimal denoising function
:
, which can be formulated as follows:
where
is the (
)-truncated vector of
in descending order.
Again, we can design T-DAE-EDMR denoted by
with a new
as
. In order to construct T-DAE-EDMR, we only need to change the dimension of the weight matrix for encoding to
, and all the other configurations are the same as for
. Now,
can be obtained through the
M training dataset as follows:
Next, in the test phase, let
denote the
j-th denoised eigenvalues with the optimized
, and this can be written as
Finally, in the test phase, we can again obtain the
j-th denoised EDM
as follows:
After obtaining
, performing connectivity-based localization involves repeating the work in
Section 2.2.2 but replacing
with
.
Overall, Algorithms 1 and 2 describe the processes of DAE-EDMR and T-DAE-EDMR, respectively.
Algorithm 1 The DAE-EDMR process |
- 1:
[The training phase (M: number of training datasets)] - 2:
Collect the training dataset of true and noisy EDMs, i.e., and , for all , . - 3:
for to M, do - 4:
Extract the vectors of the eigenvalues of and , i.e., and . - 5:
end for - 6:
Optimize the NN-based denoiser consisting of encoding/decoding functions based on ( 9), i.e., , by inputting and outputting and , respectively. - 7:
[The test phase (L: number of test datasets)] - 8:
Collect the test dataset of true and noisy EDMs, i.e., and , for all , . - 9:
for to L, do - 10:
Make the input vector referring to step 4. - 11:
Generate by passing to . - 12:
Obtain the denoised EDM, i.e., , based on ( 12). - 13:
Implement the classical MDS with to obtain the estimate of . - 14:
end for
|
Algorithm 2 The T-DAE-EDMR process |
- 1:
[The training phase (M: number of training datasets)] - 2:
Collect the training dataset of true and noisy EDMs, i.e., and , for all , . - 3:
for to M do - 4:
Extract the eigenvalues of , i.e., . - 5:
Select the dominant eigenvalues of , i.e., . - 6:
end for - 7:
Optimize the NN-based denoiser consisting of encoding/decoding functions based on ( 16), i.e., , by inputting and outputting and , respectively. - 8:
[The test phase (L: number of test datasets)] - 9:
Collect the test dataset of true and noisy EDMs, i.e., and , for all , . - 10:
for to L, do - 11:
Make the input vector by referring to step 5. - 12:
Generate by passing to . - 13:
Obtain the denoised EDM, i.e., , based on ( 18). - 14:
Implement the classical MDS with to obtain the estimate of . - 15:
end for
|
2.4. Computational Complexity and Memory Utilization of DAE-EDMR and T-DAE-EDMR
Since the training process is performed offline and will not affect the online denoising overhead, we mainly consider the complexity of online denoising. Recalling the dimension of the latent space
and the depth of the NN model
I, FLOPs of
and
are basically required in DAE-EDMR and T-DAE-EDMR, respectively, in terms of the online complexity for the fully connected NN model’s operation. The proposed DAE-EDMR requires an additional online complexity of
FLOPs [
24] for EVD of the noisy EDM
compared to conventional NN-based works requiring online complexity for matrix multiplications (As is generally known, the FLOPs required to extract the eigenvalues are
. Additionally, in the case of the extraction of the eigenvalues and eigenvectors together, the required FLOPs are
. In our work, we only take the eigenvalues because these alone are sufficient for denoising and reducing the number of FLOPs required in offline training. Furthermore, since EDMs are inherently symmetric matrices, extracting the eigenvalues can be performed more efficiently. Investigating this aspect further could be an intriguing direction of future work, potentially leading to more computationally efficient approaches to EDM-based processing.). Additionally, T-DAE-EDMR can reduce the online complexity of DAE-EDMR to
FLOPs regarding the truncated eigenproblem. Furthermore, in contrast to traditional EDM-based NNs, which require a memory storage proportional to
due to their pairwise distance representations, our proposed model significantly reduces the memory usage by limiting the input and output dimensions to
. This allows for a more scalable and efficeint implementation, making it particularly suitable for large-scale networks.
4. Discussion and Conclusions
In this paper, we investigated the problem of denoising a contaminated Euclidean distance matrix (EDM) for high-accuracy connectivity-based localization based on the mathematical fact of an EDM, i.e., its low-rank property. Compared to conventional neural network (NN)-based algorithms with large-scale frameworks, we proposed two efficient algorithms, called denoising-autoencoder-aided EDM reconstruction (DAE-EDMR) and truncated DAE-EDMR (T-DAE-EDMR), which show a superior EDM denoising performance. Notably, the latter is designed within the NN framework, enabling it to achieve a robust performance even with a limited number of training datasets. Our contributions stem from the concept of inputting N (or dominant) eigenvalues of the noisy EDM into the NN model, with the addition of marginal online complexity for eigenvalue decomposition (EVD) of the EDMs. Furthermore, T-DAE-EDMR reinforces the robustness of DAE-EDMR to variations in the number of nodes between the training and test phases. T-DAE-EDMR inputs the dominant eigenvalues of the noisy EDM into the NN, extracted through ()-truncated EVD. This approach, as opposed to inputting N eigenvalues into DAE-EDMR, enhances the robustness to changing environments while reducing the required training dataset and off/online complexity. The proposed approach effectively leverages the linear algebraic properties of wireless localization. In particular, the use of eigenvalues to optimize the NN model for EDM denoising suggests that the volume of the convex hull formed by the distributed nodes can be interpreted as the crucial information. Our experimental results demonstrate that the proposed algorithms reduce the required training dataset’s size to nearly one-tenth of its original size while achieving more than twice the effectiveness in EDM denoising. Given the suitability of the proposed methods for massive connectivity scenarios, our approach offers useful advantages for practical deployment. Furthermore, there is an opportunity to further enhance the EDM denoising performance by incorporating deeper mathematical insights, considering not only eigenvalues but also eigenvectors. Building on these strengths, future research will focus on extending our proposed schemes to localization and tracking techniques that can more robustly adapt to time-varying environments, making them even more applicable to large-scale implementations.