Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Enhancing Security and Efficiency in IoT Assistive Technologies: A Novel Hybrid Systolic Array Multiplier for Cryptographic Algorithms
Next Article in Special Issue
Multi-Channel Speech Enhancement Using Labelled Random Finite Sets and a Neural Beamformer in Cocktail Party Scenario
Previous Article in Journal
Experimental Analysis of Heat and Flow Characteristics on Inclined and Multiple Impingement Jet Heat Transfer Using Optimized Heat Sink
Previous Article in Special Issue
Comparative Study of Minimally Invasive Microwave Ablation Applicators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Denoising-Autoencoder-Aided Euclidean Distance Matrix Reconstruction for Connectivity-Based Localization: A Low-Rank Perspective

by
Woong-Hee Lee
1,
Mustafa Ozger
2,
Ursula Challita
3 and
Taewon Song
4,*
1
Division of Electronics and Electrical Engineering, Dongguk University-Seoul, Seoul 04620, Republic of Korea
2
Department of Electronic Systems, Aalborg University, 2450 Copenhagen, Denmark
3
Ericsson AB, 16440 Stockholm, Sweden
4
Department of Internet of Things, Soonchunhyang University, Asan 31538, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(5), 2656; https://doi.org/10.3390/app15052656
Submission received: 10 January 2025 / Revised: 21 February 2025 / Accepted: 23 February 2025 / Published: 1 March 2025

Abstract

:
In contrast to conventional localization methods, connectivity-based localization is a promising approach that leverages wireless links among network nodes. Here, the Euclidean distance matrix (EDM) plays a pivotal role in implementing the multidimensional scaling technique for the localization of wireless nodes based on pairwise distance measurements. This is based on the representation of complex datasets in lower-dimensional spaces, resulting from the mathematical property of an EDM being a low-rank matrix. However, EDM data are inevitably susceptible to contamination due to errors such as measurement imperfections, channel dynamics, and clock asynchronization. Motivated by the low-rank property of the EDM, we introduce a new pre-processor for connectivity-based localization, namely denoising-autoencoder-aided EDM reconstruction (DAE-EDMR). The proposed method is based on optimizing the neural network by inputting and outputting vectors of the eigenvalues of the noisy EDM and the original EDM, respectively. The optimized NN denoises the contaminated EDM, leading to an exceptional performance in connectivity-based localization. Additionally, we introduce a relaxed version of DAE-EDMR, i.e., truncated DAE-EDMR (T-DAE-EDMR), which remains operational regardless of variations in the number of nodes between the training and test phases in NN operations. The proposed algorithms show a superior performance in both EDM denoising and localization accuracy. Moreover, the method of T-DAE-EDMR notably requires a minimal number of training datasets compared to that in conventional approaches such as deep learning algorithms. Overall, our proposed algorithms reduce the required training dataset’s size by approximately one-tenth while achieving more than twice the effectiveness in EDM denoising, as demonstrated through our experiments.

1. Introduction

To configure, monitor, and control many applications in Internet of Things (IoT) networks, accurate localization of every sensor will be a key enabling technology for private 5G applications and beyond, especially for industries [1]. Many systems with localization algorithms have been developed by means of wireless sensor networks for both indoor and outdoor environments. To achieve a higher localization accuracy, additional hardware implementations are utilized by most of the existing localization solutions, which increase the cost and considerably limit location-based applications. Consequently, conventional localization methods such as global positioning systems (GPSs) are not suitable, as their direct implementation in IoT networks involves prohibitive demands for sophisticated equipment and substantial energy consumption and cannot meet the required accuracy level, for instance, in industrial setups. These limitations have significantly restricted the practical scalability of IoT networks. To overcome these challenges, massive wireless connections in IoT networks are leveraged for the cooperative location estimation of IoT devices, which is called connectivity-based localization [2,3,4,5,6,7,8,9]. This approach not only tackles the limitations of the conventional localization methods but also enhances the localization accuracy, resulting in more robust and energy-efficient localization within IoT networks.
From an algorithmic standpoint, multidimensional scaling (MDS) is a widely employed collection of statistical methods which are extensively utilized to create mappings of items based on their distance, i.e., dissimilarity [10,11,12,13,14,15]. MDS methods are able to represent complex datasets in spaces with lower dimensions; meanwhile, they maintain the dissimilarity relations among the items in the original datasets. Here, the Euclidean distance matrix (EDM) is the key information for implementing the MDS technique, which is constructed using pairwise distance measurements. The EDM serves as a useful description of the point sets and a solid foundation for localization algorithm design due to its effective description of the point sets. However, in real-world environments, EDM data are inevitably prone to contamination due to errors resulting from different sources, such as the measurement resolution, signal quality, network asynchronization, non line-of-sight (NLoS) conditions, and so on. In fact, the contamination in the EDM degrades the localization performance of the MDS, particularly in large-scale IoT networks. Errors in the EDM may lead to inaccurate and imprecise location estimations, which will result in a less reliable and robust connectivity-based localization process. Hence, it is critical to address the EDM contamination issue to ensure accurate and efficient localization in large-scale IoT networks.
To mitigate the effects of noisy measurements on the EDM, denoising techniques have been extensively utilized, and they aim to denoise a noisy EDM by resolving properly designed optimization problems such as semi-definite relaxation [16,17] and low-rank tensor completion [18]. However, they rely on solving complex optimization problems, which may limit their efficiency and effectiveness. Developments in artificial intelligence (AI) methods, particularly neural network (NN)-based denoising methods such as [19,20,21,22], have become promising in leveraging statistical inference as a novel and potentially more robust alternative for dealing with noisy measurements in EDM-based localization. 
The existing NN-based EDM denoising techniques require computations that scale with the square of the total number of nodes since both the input and output of the NN framework are based on pairwise distances. This is due to the combinatorial nature of measuring and generating the distances between node pairs. As the number of nodes increases, the computational complexity grows quadratically, leading to the need for a large training dataset size and extensive NN models.
In addition, the existing NN-based denoising methods may not be suitable for IoT networks due to their resource limitations. Although major research efforts have focused on big data analysis and deep neural networks, it is crucial to consider that most IoT devices face severe limitations in terms of their data acquisition capabilities, computational power, and memory size. Hence, the successful implementation of efficient NN-based algorithms that can handle big data while considering the resource limitations of IoT devices is a critical challenge to address.
In this paper, we propose a novel denoiser for a noisy EDM, referred to as DAE-EDMR (Readers can understand our framework more easily by referring to Figure 1. Our method is based on the mathematical fact that the EDM measured from N nodes in a k-dimensional space must have a rank of at most k + 2 . Abstractly, this implies that N-choose-2 pairwise distances can be rearranged into a k-dimensional structure. Consequently, if the pairwise distances are not accurately measured, reconstructing these segments would require embedding them into a higher-dimensional space. Using this concept, we optimize the NN model for EDM denoising by utilizing the eigenvalues to capture how the nodes are volumetrically distributed. A detailed explanation is given in Section 2), which leverages a denoising autoencoder (DAE). To increase the efficiency of the NN operations, we exploit the low-rank property of the EDM, which is bounded to only a k + 2 number of eigenvalues, where k is the dimension of the Euclidean space, independent of the number of sensor nodes [23]. Leveraging this valuable mathematical observation, we design our NN model by inputting and outputting vectors of the eigenvalues of the noisy and original EDMs, respectively. Through these inputting/outputting rules, the proposed scheme achieves remarkable denoising results, even with a relatively small training dataset. The efficiency of the proposed denoiser is attributed to the utilization of the low-rank property of the EDM, which helps the NN model to establish better inference with limited training data. In fact, the combination of the DAE and EDM is highly attractive due to the complementary features of these two techniques. On the one hand, the DAE is a powerful NN framework for manifold learning, which enables it to effectively capture intricate data patterns and facilitate efficient denoising. On the other hand, the EDM exhibits an extremely-low-rank property, contributing to the dimensionality reduction. Regarding the online complexity, the proposed method requires eigenvalue decomposition (EVD) of the EDM as a pre-processing step for the NN operations.
In addition, we propose a technique called truncated DAE-EDMR, i.e., T-DAE-EDMR, to enhance the robustness of DAE-EDMR to diverse environments conditions, such as changes in the number of nodes, even after NN optimization. In other words, T-DAE-EDMR offers the flexibility to be utilized in scenarios where the number of nodes varies between the training and test phases. The T-DAE-EDMR scheme involves feeding k + 2 eigenvalues from the noisy EDM to the NN, which are extracted through ( k + 2 )-truncated EVD. In contrast, DAE-EDMR inputs N eigenvalues, making T-DAE-EDMR more versatile in accommodating different environments and simultaneously reducing the online complexity. Hence, T-DAE-EDMR is envisioned to demonstrate its superior effectiveness in environments characterized by a large number of nodes or frequent topology changes, such as vehicle-to-everything (V2X) systems.
To summarize, the main contributions of the proposed algorithms to denoising the noisy EDMs and enhancing the localization accuracy are four-fold:
  • Minimizing the reconstruction errors for the EDM through the mix-up of mathematical evidence and NNs: We reveal the potential to reconstruct the EDM according to the low-rank property of the ground-truth EDM and the low-dimensional representations of the NN operation, which is robust to various noise models in distance measurements.
  • Reducing the size of the training dataset and NN model: We develop an efficient NN framework requiring a small-sized training dataset and NN model. This is based on the novel inputting and outputting for the NN model, which consist of the eigenvalues of the noisy EDM and the ground-truth EDM, respectively.
  • Assisting the existing connectivity-based localization algorithms as a pre-processor: We combine our proposed scheme with the connectivity-based localization technique to validate its utility in various environments. It is verified that these joint frameworks show a superior performance compared to that of the conventional approaches. This achievement is highly remarkable, as we only need to extract the eigenvalues of an EDM, requiring marginal online complexity.
  • Making the NN model robust to the dynamics of wireless networks: By additionally presenting a modified model of our proposed model, we introduce an NN that can be robust to the variability in wireless networks, e.g., the number of nodes in the test phase is changed after NN optimization.
The remainder of this paper is organized as follows: Section 2 presents the system model, the problem design, and the method for the proposed algorithms. In Section 3, we provide numerical results to demonstrate the superiority of our proposed algorithms compared to other schemes even given the aspects of various environments. Finally, Section 4 presents the concluding remarks.

2. The Proposed Schemes: DAE-EDMR and T-DAE-EDMR

2.1. The System Model and Problem Formulation

Consider a collection of N nodes in a k-dimensional Euclidean space, X = [ x 1 , , x N ] R k × N , where the positions of all N nodes, x 1 , , x N R k , are randomly distributed. With the knowledge of the positions of P ( > k ) reference nodes, we estimate the positions of the remaining N-P nodes. To accomplish this, we will utilize the concept of an EDM denoted by D R + N × N . It is a symmetric matrix whose ( i , j ) -element can be represented as follows:
D ( i , j ) = d i j 2 = | | x i x j | | 2 2 = < x i x j , x i x j > ,
where d i j is the true distance between nodes i and j.
To model practical measurements, we first consider three types of random variables due to the environment as follows:
  • B N : ranging errors dependent on the signal quality;
  • B U : ranging errors due to clock asynchronization;
  • B N L o S : non line-of-sight (NLoS) events.
We assume that B N , B U , and B N L o S follow normal, uniform, and Bernoulli distributions, respectively. Hence, we can define the random variable for the bias, B b i a s , as follows:
B b i a s = B N + B U + R N L o S B N L o S ,
where R N L o S is the distance bias in the event of NLoS conditions. Note that B b i a s does not follow any known probability distribution, as it is a convolution of three different distributions.
Second, we assume that the distance is measured using a grid consisting of measurement resolutions, which is determined by the ranging configuration, e.g., the time of arrival (ToA). Thus, we define the quantization function Q G to represent the measured distance with a resolution of G, e.g., Q 10 ( 23 ) = 20 .
Third, we formulate a function to indicate whether the distance is measured or not, based on the communication capability between nodes i and j, as follows:
δ i j = 1 if d i j is measured , 0 otherwise .
Thus, the noisy measurement of the distance between the i-th and the j-th nodes is defined as follows:
d ˜ i j = δ i j · Q G ( d i j + e i j ) ,
where e i j is the realization of B b i a s . Then, we can define the noisy EDM using the following expression:
D ˜ = [ d ˜ i j 2 ] i , j = 1 N R + N × N .
Finally, the objective of this paper is to find the denoising function H , which is formulated as
min H   | | D H ( D ˜ ) | | F 2 .

2.2. Method I: DAE-EDMR

In the conventional approaches, the denoising function was set as H : R + N × N R + N × N , more specifically H : D ˜ D , such as in semi-definite relaxation and nonlocal patch tensor-based methods. These techniques require high computational complexity, as they perform iterative matrix multiplication operations while solving high-order optimization problems. In order to overcome this problem, NN-based techniques have recently been proposed. However, the dimension of the input and output data is large, which are the entire elements of the EDM. As a result, they exhibit low performance and require a large amount of training data and a large-sized neural network.
Considering the above problems, we propose a new method, namely DAE-EDMR, to denoise the noisy EDM in an efficient way.

2.2.1. A Denoising Process

Before describing the framework of DAE-EDMR, we will revisit the low-rank property of an EDM [23].
Property 1.
The rank ( D ) , corresponding to the points in R k , is at most k + 2 .
This is based on the following equality:
D = 1 N diag ( X T X ) T 2 X T X + diag ( X T X ) 1 N T ,
where 1 N is the one-vector of size N. According to rank characteristics, the rank of D is bounded to the summation of the ranks of each term, i.e., rank ( D ) rank ( 1 N diag ( X T X ) T ) + rank ( X T X ) + rank ( diag ( X T X ) 1 N T ) = 1 + k + 1 . This determines the rank of an EDM as extremely low, i.e., rank ( D ) k + 2 , regardless of the number of nodes. It implies that it is effective to perform denoising given the small dimension of the (potential) latent space of an NN model, i.e., the similar level of the rank of the EDM, rather than treating the distance information as a whole.
Now, let s R k + 2 and s ˜ R N denote the vectors whose elements are the descending-order eigenvalues of D and D ˜ , respectively. Thus, the optimal denoising function H : R N R k + 2 can be constructed as follows:
H = argmin H | | s H ( s ˜ ) | | 2 2 .
Here, we design a fully connected NN framework denoted by H NN to approximate H . To this end, we will define the required terms as follows:
  • N : The dimension of the latent space.
  • W R N × N , W R ( k + 2 ) × N : The weight matrices for encoding and decoding, respectively.
  • b R N , b R k + 2 : The bias vectors for encoding and decoding, respectively.
  • S : The activation function for neural networks. At the propagation between the final hidden layer and the output layer, S ( a ) = a , i.e., an identity function. For other types of propagation between adjacent layers, S ( a ) = e a e a e a + e a , i.e., a hyperbolic tangent function. And S ( a ) = ( S ( a [ 1 ] ) , , S ( a [ P ] ) ) T where a R P is an arbitrary input vector.
With these terms (For simplicity, the description of the NN model design throughout this article is based on a single hidden layer; however, it is obvious that deeper hidden layers can be made using multiple encoding/decoding function parameters, i.e., { θ i } i = 1 I and { θ i } i = 1 I , where I, θ i , and θ i are the depth of the NN model and the i-th encoding and decoding function parameters), we define f θ as the encoding function where the parameter θ is { W , b } , i.e., f θ ( s ˜ ) = S ( W s ˜ + b ) . In addition, we define g θ as the decoding function where the parameter θ is { W , b } , i.e., g θ ( f θ ( s ˜ ) ) = S ( W f θ ( s ˜ ) + b ) .
Finally, we can define H NN consisting of the optimal encoding and decoding functions, denoted by f θ and g θ , respectively, as follows:
H NN = { f θ , g θ } = argmin f θ , g θ 1 M i = 1 M | | s ( i ) g θ ( f θ ( s ˜ ( i ) ) ) | | 2 2 ,
where M is the number of training datasets. Let Θ be all of the model parameters, i.e.,  Θ : = θ θ = { W , W , b , b } ; then, it simultaneously updates every parameter in Θ at each iteration toward the direction of the steepest descent as follows:
Θ ( n + 1 ) = Θ ( n ) η Θ 1 M i = 1 M | | s ( i ) g θ ( f θ ( s ˜ ( i ) ) ) | | 2 2 ,
where Θ is the gradient operator with respect to Θ and η is the learning rate related to the step size. Through this procedure, we can optimize the fully connected NN model H NN for denoising the eigenvalues.
Next, in the test phase, let s ^ ( j ) denote the j-th denoised eigenvalues with the optimized H NN , and this can be obtained as follows:
s ^ ( j ) = H NN ( s ˜ ( j ) ) = g θ ( f θ ( s ˜ ( j ) ) ) .
Finally, the denoised EDM D ^ ( j ) can be reconstructed as follows:
D ^ ( j ) = V ˜ ( j ) diag ( s ^ ( j ) ) V ˜ ( j ) T ,
where V ˜ ( j ) R N × ( k + 2 ) is the matrix consisting of k + 2 eigenvectors of D ˜ ( j ) .

2.2.2. The Connectivity-Based Localization Process

To obtain an estimate of X , which is denoted by X ^ R k × N , based on the classical MDS method, we first define the geometric centering matrix as follows:
C = I N 1 N 1 N 1 N T R N × N ,
where I N is the identity matrix N by N in size. Next, the estimated centered Gram matrix (GM) is obtained as
G ^ c = 1 2 C D ^ C R N × N .
Recalling the fact that G ^ c = X ^ c T X ^ c , where X ^ c is the centered X ^ , we can easily obtain X ^ c through the k-truncated EVD of G ^ c . Based on this, we can finally obtain X ^ through a rigid linear transform, i.e., rotation and translation, of X ^ c with the pre-knowledge of [ x 1 , , x P ] R k × P , which are the positions of the reference nodes.

2.3. Method II: Truncated DAE-EDMR (T-DAE-EDMR)

This subsection introduces T-DAE-EDMR, which is the relaxed version of DAE-EDMR. From the previous subsection, the low-rank property of the EDM can be used more efficiently from the NN training/test point of view. Assume that the NN model is optimized in a network with N nodes. Here, we consider a scenario where the number of nodes changes at the test phase, which is frequently shown in wireless networks. If the NN model can be used flexibly under this kind of environment change, it will be a more efficient utilization.
For this reason, we newly define the optimal denoising function H T : R k + 2 R k + 2 , which can be formulated as follows:
H T = argmin H | | s H ( s ˜ T ) | | 2 2 ,
where s ˜ T R k + 2 is the ( k + 2 )-truncated vector of s ˜ in descending order.
Again, we can design T-DAE-EDMR denoted by H T NN with a new { f θ , g θ } as { f θ T , g θ T } . In order to construct T-DAE-EDMR, we only need to change the dimension of the weight matrix for encoding to W R N × ( k + 2 ) , and all the other configurations are the same as for H NN . Now, H T NN can be obtained through the M training dataset as follows:
H T NN = { f θ T , g θ T } = argmin f θ , g θ 1 M i = 1 M | | s ( i ) g θ ( f θ ( s ˜ T ( i ) ) ) | | 2 2 .
Next, in the test phase, let s ^ T ( j ) denote the j-th denoised eigenvalues with the optimized H T NN , and this can be written as
s ^ T ( j ) = H T NN ( s ˜ T ( j ) ) = g θ T ( f θ T ( s ˜ T ( j ) ) ) .
Finally, in the test phase, we can again obtain the j-th denoised EDM D ^ T ( j ) as follows:
D ^ T ( j ) = V ˜ ( j ) diag ( s ^ T ( j ) ) V ˜ ( j ) T .
After obtaining D ^ T , performing connectivity-based localization involves repeating the work in Section 2.2.2 but replacing D ^ with D ^ T
Overall, Algorithms 1 and 2 describe the processes of DAE-EDMR and T-DAE-EDMR, respectively.
Algorithm 1  The DAE-EDMR process
1:
[The training phase (M: number of training datasets)]
2:
Collect the training dataset of true and noisy EDMs, i.e., D ( i ) and D ˜ ( i ) , for all i { 1 , , M } .
3:
for  i 1 to M, do
4:
    Extract the vectors of the eigenvalues of D ( i ) and D ˜ ( i ) , i.e., s ( i ) and s ˜ ( i ) .
5:
end for
6:
Optimize the NN-based denoiser consisting of encoding/decoding functions based on (9), i.e., H NN ( = { f θ , g θ } ) , by inputting and outputting s ˜ ( i ) and s ( i ) , respectively.
7:
[The test phase (L: number of test datasets)]
8:
Collect the test dataset of true and noisy EDMs, i.e., D ( j ) and D ˜ ( j ) , for all j { 1 , , L } .
9:
for  j 1 to L, do
10:
    Make the input vector s ˜ ( j ) referring to step 4.
11:
    Generate s ^ ( j ) by passing s ˜ ( j ) to H NN .
12:
    Obtain the denoised EDM, i.e., D ^ ( j ) , based on (12).
13:
    Implement the classical MDS with D ^ ( j ) to obtain the estimate of X ( j ) .
14:
end for
Algorithm 2 The T-DAE-EDMR process
1:
[The training phase (M: number of training datasets)]
2:
Collect the training dataset of true and noisy EDMs, i.e., D ( i ) and D ˜ ( i ) , for all i { 1 , , M } .
3:
for  i 1 to M do
4:
    Extract the eigenvalues of D ( i ) , i.e., s ( i ) .
5:
    Select the dominant k + 2 eigenvalues of D ˜ ( i ) , i.e., s ˜ T ( i ) .
6:
end for
7:
Optimize the NN-based denoiser consisting of encoding/decoding functions based on (16), i.e., H T NN ( = { f θ T , g θ T } ) , by inputting and outputting s ˜ T ( i ) and s ( i ) , respectively.
8:
[The test phase (L: number of test datasets)]
9:
Collect the test dataset of true and noisy EDMs, i.e., D ( j ) and D ˜ ( j ) , for all j { 1 , , L } .
10:
for  j 1 to L, do
11:
    Make the input vector s ˜ T ( j ) by referring to step 5.
12:
    Generate s ^ T ( j ) by passing s ˜ T ( j ) to H T NN .
13:
    Obtain the denoised EDM, i.e., D ^ T ( j ) , based on (18).
14:
    Implement the classical MDS with D ^ T ( j ) to obtain the estimate of X ( j ) .
15:
end for

2.4. Computational Complexity and Memory Utilization of DAE-EDMR and T-DAE-EDMR

Since the training process is performed offline and will not affect the online denoising overhead, we mainly consider the complexity of online denoising. Recalling the dimension of the latent space N and the depth of the NN model I, FLOPs of O ( I N N ) and O ( I N ( k + 2 ) ) are basically required in DAE-EDMR and T-DAE-EDMR, respectively, in terms of the online complexity for the fully connected NN model’s operation. The proposed DAE-EDMR requires an additional online complexity of O ( N 3 ) FLOPs [24] for EVD of the noisy EDM D ˜ R + N × N compared to conventional NN-based works requiring online complexity for matrix multiplications (As is generally known, the FLOPs required to extract the eigenvalues are 4 N 3 / 3 . Additionally, in the case of the extraction of the eigenvalues and eigenvectors together, the required FLOPs are 8 N 3 / 3 . In our work, we only take the eigenvalues because these alone are sufficient for denoising and reducing the number of FLOPs required in offline training. Furthermore, since EDMs are inherently symmetric matrices, extracting the eigenvalues can be performed more efficiently. Investigating this aspect further could be an intriguing direction of future work, potentially leading to more computationally efficient approaches to EDM-based processing.). Additionally, T-DAE-EDMR can reduce the online complexity of DAE-EDMR to O ( ( k + 2 ) N 2 ) FLOPs regarding the truncated eigenproblem. Furthermore, in contrast to traditional EDM-based NNs, which require a memory storage proportional to O ( N 2 ) due to their pairwise distance representations, our proposed model significantly reduces the memory usage by limiting the input and output dimensions to k + 2 . This allows for a more scalable and efficeint implementation, making it particularly suitable for large-scale networks.

3. Simulation Results

In this section, we analyze the effect of DAE-EDMR and T-DAE-EDMR from various perspectives. We compare the performance of our proposed methods with that of the conventional algorithms, which are semi-definite relaxation [16], the nonlocal patch tensor-based method [18], and conventional NN-based methods referred to as deep-learning-based methods A [19] and B [22]. We also provide the result for the undenoised EDM to show the denoising performance of each algorithm. Overall, we will demonstrate the effectiveness of each scheme in denoising the noisy EDM and assess its impact on the localization performance.

3.1. The Simulation Setup

The performance of the proposed algorithms is evaluated via 5000 episodes. We define two error metrics: the EDM error and the localization error. The EDM error quantifies the deviation between the denoised EDM and the original EDM. Specifically, it is computed by first taking the Frobenius norm of their difference and then normalizing it by the Frobenius norm of the original EDM, with the final result averaged over all experimental trials. This metric provides a measure of the accuracy of the denoised pairwise distances. Meanwhile, the localization error represents the average localization error across all experimental trials, capturing the overall accuracy of the proposed method in reconstructing spatial coordinates.
In the simulations, N is set to 30, and the entire nodes are randomly distributed over a 100 × 100 -sized space. Furthermore, the resolution of the distance measurements is assumed to be 10. Next, we set the variables of the noisy measurements, which are defined in Section 2 A, as follows: B N N ( 0 , 5 ) , B U U ( 0 , 10 ) , B N L o S Ber ( 0.5 ) (except for Section 3C), and R N L o S = 25 . In addition, δ i j in (3) follows Ber ( 0.9 ) for all i , j , i.e., 10 % of the distance measurements are missed on average. To clearly confirm the performance in EDM denoising, all of the EDM measurements follow the above assumptions in terms of the statistical parameters for error modeling and the resolution of the distance measurements, irrespective of whether the pair measuring their distance contains the reference/target node. Additionally, except for Table 1a, the number of reference nodes (P) is 18. The classical MDS [23] is utilized as the localization method to estimate the position of the target nodes.
Throughout all cases for NN optimization, the squared error and the scaled conjugate gradient are applied as the loss function and the optimization method, respectively. We consider a fully connected NN model. Furthermore, the depth of the hidden layer and the number of perceptrons per hidden layer are 2 and 450, respectively. With the exception of Figure 2, the size of the training dataset (M) was 10,000 for all experiments. We tested on NVIDIA RTX A2000 GPU machines (Santa Clara, CA, USA).

3.2. Simulation Results According to the Number of Training Datasets

Figure 2 shows the performance of each algorithm with respect to the size of the training dataset used in the NN optimization process. Only the NN-based methods show variation in their performance with the size of the training dataset.
Despite the large size of the training dataset, such as 10 5 , the conventional deep-learning-based methods are not able to surpass the performance of the nonlocal patch tensor-based method.
This can be attributed to the challenge of extracting valuable information from the raw and unrefined training dataset. Remarkably, the T-DAE-EDMR technique outperforms the other techniques even when M is just 10 2 . In addition, DAE-EDMR shows a noticeable improvement when M is increased to 10 3 . It is noteworthy that when M is 10 5 , DAE-EDMR shows a slightly better performance than that of T-DAE-EDMR. This observation implies that extracting information from N contaminated eigenvalues, instead of solely relying on the k + 2 highest values among them, can yield advantages when a large training dataset is available.

3.3. Simulation Results According to NLoS Probability

Figure 3 illustrates the impact of the NLoS probability on the EDM and localization errors. It is assumed that the NLoS probability follows a Bernoulli distribution. As the NLoS probability increases, the error for distance measurements naturally increases, resulting in a worse localization performance. Therefore, the non-NN-based techniques are in line with this general situation. On the other hand, a different phenomenon appears from the information-theoretic perspective. In the NN-based methods, i.e., deep-learning-based methods A and B, DAE-EDMR, and T-DAE-EDMR, the EDM error takes the form of a concave function for the NLoS probability. This is related to the fact that the variance in and the entropy of Ber ( p ) are p ( 1 p ) and p log ( p ) ( 1 p ) log ( 1 p ) , respectively, which are in the form of concave functions. Interestingly, our proposed methods show a superior localization performance in the extreme case where the NLoS probability is 1 and can therefore be effectively applied in environments where NLoS conditions occur frequently. For future work, an interesting denoising task may involve modeling the distance measurement error due to NLoS conditions as a joint random variable, considering both the event occurrence and distance biasing. This approach could provide valuable insights into further enhancing the denoising techniques in the presence of errors due to NLoS conditions.

3.4. Simulation Results According to the Variation in the Cardinality of the Sensor Network

In this subsection, we deal with the dynamics of the sensor nodes. Table 1a shows the EDM and localization errors for a case where only P is changed while keeping N constant ( N = 30 ). Since the EDM measurements are independent of P, only the denoising performance of each algorithm affects the EDM error. In the case of localization errors, it can be seen that the performance improves as P increases for all algorithms, and the two algorithms we propose show an excellent performance overall.
Table 1b shows the impact of the change in the number of nodes after NN optimization is performed, with N equal to 30. In other words, this experiment serves to verify the flexibility of T-DAE-EDMR when the number of nodes, N, changes between the training and test phases. For convenience, we will refer to the changed value of N as N n e w . In the undenoised scenario, as N n e w increases, the possibility of cooperation between the nodes increases; consequently, it can be seen that the performance marginally improves. As expected, T-DAE-EDMR shows its denoising capability well when N n e w is equal to N. In addition, since the model is optimized based on the information for a situation where N is 30, the performance of T-DAE-EDMR deteriorates as the value of | N n e w N | increases. Nevertheless, our proposed T-DAE-EDMR can be considered an effectively designed algorithm that can adapt robustly to a change in the number of nodes between the training and test phases. Due to its potential, T-DAE-EDMR could be very useful in highly time-varying networks.

3.5. Simulation Results According to the Utilized Matrices

Figure 4 shows the results of applying our techniques to both an EDM and a GM, where a GM has a sharper condition in terms of its low-rank property. Under this setting, the EDM error and the localization error are investigated with respect to the size of the training dataset. A GM is created by multiplying the centering matrix on both sides of the EDM, and it is clear that it has a sharper condition than that of the EDM. In other words, rank ( GM ) k where k is the dimension of the Euclidean space. As can be seen in Figure 4, both tests show a mostly similar performance when the size of the training dataset is large enough since they contain essentially the same information. However, in cases where the training dataset is insufficient, a slight deterioration in performance is observed when the GM is applied. This can be inferred because when the GM is created through the multiplication of the EDM and the centering matrix, then some useful information is contaminated during this procedure, as the centering matrix is a singular matrix.

4. Discussion and Conclusions

In this paper, we investigated the problem of denoising a contaminated Euclidean distance matrix (EDM) for high-accuracy connectivity-based localization based on the mathematical fact of an EDM, i.e., its low-rank property. Compared to conventional neural network (NN)-based algorithms with large-scale frameworks, we proposed two efficient algorithms, called denoising-autoencoder-aided EDM reconstruction (DAE-EDMR) and truncated DAE-EDMR (T-DAE-EDMR), which show a superior EDM denoising performance. Notably, the latter is designed within the NN framework, enabling it to achieve a robust performance even with a limited number of training datasets. Our contributions stem from the concept of inputting N (or k + 2 dominant) eigenvalues of the noisy EDM into the NN model, with the addition of marginal online complexity for eigenvalue decomposition (EVD) of the EDMs. Furthermore, T-DAE-EDMR reinforces the robustness of DAE-EDMR to variations in the number of nodes between the training and test phases. T-DAE-EDMR inputs the k + 2 dominant eigenvalues of the noisy EDM into the NN, extracted through ( k + 2 )-truncated EVD. This approach, as opposed to inputting N eigenvalues into DAE-EDMR, enhances the robustness to changing environments while reducing the required training dataset and off/online complexity. The proposed approach effectively leverages the linear algebraic properties of wireless localization. In particular, the use of eigenvalues to optimize the NN model for EDM denoising suggests that the volume of the convex hull formed by the distributed nodes can be interpreted as the crucial information. Our experimental results demonstrate that the proposed algorithms reduce the required training dataset’s size to nearly one-tenth of its original size while achieving more than twice the effectiveness in EDM denoising. Given the suitability of the proposed methods for massive connectivity scenarios, our approach offers useful advantages for practical deployment. Furthermore, there is an opportunity to further enhance the EDM denoising performance by incorporating deeper mathematical insights, considering not only eigenvalues but also eigenvectors. Building on these strengths, future research will focus on extending our proposed schemes to localization and tracking techniques that can more robustly adapt to time-varying environments, making them even more applicable to large-scale implementations.

Author Contributions

Conceptualization: W.-H.L. and T.S. Methodology: W.-H.L. Software: W.-H.L. Validation: M.O. and U.C. Formal analysis: M.O. Investigation: U.C. Resources: W.-H.L. Data curation: W.-H.L. Writing—original draft preparation: W.-H.L. Writing—review and editing: W.-H.L., M.O., U.C. and T.S. Visualization: W.-H.L. Supervision: T.S. Project administration: T.S. Funding acquisition: W.-H.L. and T.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by an Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (No. RS-2023-00217885, Development of integrated interference analysis technology for improving frequency utilization efficiency) and in part by the Soonchunhyang University Research Fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting the reported results in this study can be shared upon reasonable request. Interested researchers should contact the corresponding author to discuss access. Please note that data sharing may be subject to restrictions due to privacy, ethical considerations, or intellectual property agreements with funders.

Conflicts of Interest

Author Ursula Challita was employed by the company Ericsson AB. The remaining authors declare that the re-search was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Li, Y.; Zhuang, Y.; Hu, X.; Gao, Z.; Hu, J.; Chen, L.; He, Z.; Pei, L.; Chen, K.; Wang, M.; et al. Toward Location-Enabled IoT (LE-IoT): IoT Positioning Techniques, Error Sources, and Error Mitigation. IEEE Internet Things J. 2021, 8, 4035–4062. [Google Scholar] [CrossRef]
  2. Yao, Y.; Zhang, H.; Lin, L.; Lin, G.; Shibasaki, R.; Song, X.; Yu, K. Internet of Things Positioning Technology Based Intelligent Delivery System. IEEE Trans. Intell. Transp. Syst. 2022, 24, 12862–12876. [Google Scholar] [CrossRef]
  3. Farahsari, P.S.; Farahzadi, A.; Rezazadeh, J.; Bagheri, A. A survey on indoor positioning systems for IoT-based applications. IEEE Internet Things J. 2022, 9, 7680–7699. [Google Scholar] [CrossRef]
  4. Alrajeh, N.A.; Bashir, M.; Shams, B. Localization techniques in wireless sensor networks. Int. J. Distrib. Sens. Netw. 2013, 9, 304628. [Google Scholar] [CrossRef]
  5. Chen, R.; Liu, M.; Hui, Y.; Cheng, N.; Li, J. Reconfigurable intelligent surfaces for 6G IoT wireless positioning: A contemporary survey. IEEE Internet Things J. 2022, 9, 23570–23582. [Google Scholar] [CrossRef]
  6. Shyam, S.; Juliet, S.; Ezra, K. An Emerging Paradigm in IoT-Based Indoor Positioning System. In Disruptive Technologies for Big Data and Cloud Applications: Proceedings of ICBDCC 2021; Springer Nature: Berlin/Heidelberg, Germany, 2022; pp. 613–620. [Google Scholar]
  7. Sneha, V.; Nagarajan, M. Localization in wireless sensor networks: A review. Cybern. Inf. Technol. 2020, 20, 3–26. [Google Scholar] [CrossRef]
  8. Chen, Z.; Xia, F.; Huang, T.; Bu, F.; Wang, H. A localization method for the Internet of Things. J. Supercomput. 2013, 63, 657–674. [Google Scholar] [CrossRef]
  9. Lin, X.; Bergman, J.; Gunnarsson, F.; Liberg, O.; Razavi, S.M.; Razaghi, H.S.; Rydn, H.; Sui, Y. Positioning for the Internet of Things: A 3GPP perspective. IEEE Commun. Mag. 2017, 55, 179–185. [Google Scholar] [CrossRef]
  10. Shang, Y.; Ruml, W.; Zhang, Y.; Fromherz, M.P. Localization from mere connectivity. In Proceedings of the 4th ACM International Symposium on Mobile ad Hoc Networking & Computing, Annapolis, MD, USA, 1–3 June 2003; pp. 201–212. [Google Scholar]
  11. Costa, J.A.; Patwari, N.; Hero III, A.O. Distributed weighted-multidimensional scaling for node localization in sensor networks. ACM Trans. Sens. Netw. (TOSN) 2006, 2, 39–64. [Google Scholar] [CrossRef]
  12. Lee, W.H.; Choi, J.; Lee, J.H.; Kim, Y.H.; Kim, S.C. Distributed power control-based connectivity reconstruction game in wireless localization. IEEE Commun. Lett. 2016, 21, 334–337. [Google Scholar] [CrossRef]
  13. Jin, M.; Xia, S.; Wu, H.; Gu, X. Scalable and fully distributed localization with mere connectivity. In Proceedings of the 2011 Proceedings IEEE INFOCOM, Shanghai, China, 10–15 April 2011; pp. 3164–3172. [Google Scholar]
  14. Shang, Y.; Rumi, W.; Zhang, Y.; Fromherz, M. Localization from connectivity in sensor networks. IEEE Trans. Parallel Distrib. Syst. 2004, 15, 961–974. [Google Scholar] [CrossRef]
  15. Kang, I.; Nam, H. Robust Localization System Using Vector Combination in Wireless Sensor Networks. IEEE Access 2022, 10, 73437–73445. [Google Scholar] [CrossRef]
  16. Alfakih, A.Y.; Khandani, A.; Wolkowicz, H. Solving Euclidean distance matrix completion problems via semidefinite programming. Comput. Optim. Appl. 1999, 12, 13–30. [Google Scholar] [CrossRef]
  17. Krislock, N.; Wolkowicz, H. Euclidean distance matrices and applications. In Handbook on Semidefinite, Conic and Polynomial Optimization; Springer: Berlin/Heidelberg, Germany, 2012; pp. 879–914. [Google Scholar]
  18. Zhang, L.; Song, L.; Du, B.; Zhang, Y. Nonlocal low-rank tensor completion for visual data. IEEE Trans. Cybern. 2019, 51, 673–685. [Google Scholar] [CrossRef] [PubMed]
  19. Njima, W.; Chafii, M.; Nimr, A.; Fettweis, G. Deep learning based data recovery for localization. IEEE Access 2020, 8, 175741–175752. [Google Scholar] [CrossRef]
  20. Zou, H.; Chen, C.L.; Li, M.; Yang, J.; Zhou, Y.; Xie, L.; Spanos, C.J. Adversarial learning-enabled automatic WiFi indoor radio map construction and adaptation with mobile robot. IEEE Internet Things J. 2020, 7, 6946–6954. [Google Scholar] [CrossRef]
  21. Lee, W.H.; Ozger, M.; Challita, U.; Sung, K.W. Noise learning-based denoising autoencoder. IEEE Commun. Lett. 2021, 25, 2983–2987. [Google Scholar] [CrossRef]
  22. Kim, S.; Nguyen, L.T.; Kim, J.; Shim, B. Deep learning based low-rank matrix completion for IoT network localization. IEEE Wirel. Commun. Lett. 2021, 10, 2115–2119. [Google Scholar] [CrossRef]
  23. Dokmanic, I.; Parhizkar, R.; Ranieri, J.; Vetterli, M. Euclidean distance matrices: Essential theory, algorithms, and applications. IEEE Signal Process. Mag. 2015, 32, 12–30. [Google Scholar] [CrossRef]
  24. Demmel, J.W. Applied Numerical Linear Algebra; SIAM: Philadelphia, PA, USA, 1997. [Google Scholar]
Figure 1. An illustration of an example of the proposed DAE-EDMR (Additionally, truncated DAE-EDMR (T-DAE-EDMR) is presented as a relaxed version of DAE-EDMR. This additional work is carried out to make the optimized NN model more efficient by using the dominant k + 2 eigenvalues as the input data. Details can be found in Section 2.3).
Figure 1. An illustration of an example of the proposed DAE-EDMR (Additionally, truncated DAE-EDMR (T-DAE-EDMR) is presented as a relaxed version of DAE-EDMR. This additional work is carried out to make the optimized NN model more efficient by using the dominant k + 2 eigenvalues as the input data. Details can be found in Section 2.3).
Applsci 15 02656 g001
Figure 2. Performance comparison according to the number of training datasets. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Figure 2. Performance comparison according to the number of training datasets. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Applsci 15 02656 g002
Figure 3. Performance comparison according to the NLoS probability. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Figure 3. Performance comparison according to the NLoS probability. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Applsci 15 02656 g003
Figure 4. Performance comparison according to the utilized matrices. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Figure 4. Performance comparison according to the utilized matrices. (a) NMSE between ground-truth and denoised EDMs and (b) localization error in meters.
Applsci 15 02656 g004
Table 1. The performance of each scheme under varying cardinalities of the sensor network (left: NMSE between ground-truth and denoised EDMs; right: Localization error in meters).
Table 1. The performance of each scheme under varying cardinalities of the sensor network (left: NMSE between ground-truth and denoised EDMs; right: Localization error in meters).
(a) In Cases of Variation in the Number of Reference Nodes (P) Where the Number of Reference and Target Nodes (N) is Fixed to 30.
Algorithms ∖ Variations P = 5 P = 10 P = 15 P = 20 P = 25
None (undenoised) 0.532 31.69 0.554 26.41 0.539 25.28 0.535 25.01 0.538 24.31
Semi-definite relaxation 0.466 24.70 0.458 20.21 0.457 18.69 0.470 18.31 0.477 18.10
Nonlocal patch tensor-based method 0.321 24.31 0.326 19.46 0.328 18.12 0.330 17.78 0.325 17.36
Deep-learning-based method A 0.530 29.65 0.527 25.65 0.522 22.82 0.533 22.91 0.539 22.91
Deep-learning-based method B 0.334 24.64 0.323 19.71 0.328 18.29 0.325 18.04 0.326 17.90
DAE-EDMR 0.180 17.57 0.176 15.72 0.178 13.92 0.177 13.81 0.172 12.03
T-DAE-EDMR 0.183 18.62 0.183 14.89 0.183 13.59 0.178 12.94 0.176 12.42
(b) In Cases of Variation in the Number of Nodes After NN Optimization Where the Number of Reference Nodes is Fixed to 18.
Algorithms ∖ Variations N new N = 10 N new N = 5 N new = N N new N = 5 N new N = 10
None (undenoised) 0.552 24.32 0.540 22.80 0.533 21.57 0.525 20.48 0.521 20.01
Semi-definite relaxation 0.482 21.37 0.472 20.04 0.465 18.51 0.456 17.99 0.451 17.89
Nonlocal patch tensor-based method 0.332 19.06 0.325 18.11 0.320 17.12 0.316 16.89 0.314 16.51
Deep-learning-based method AN/AN/A 0.531 24.12 N/AN/A
Deep-learning-based method BN/AN/A 0.327 18.91 N/AN/A
DAE-EDMRN/AN/A 0.177 13.22 N/AN/A
T-DAE-EDMR 0.491 20.21 0.219 15.77 0.179 13.47 0.185 14.17 0.323 17.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, W.-H.; Ozger, M.; Challita, U.; Song, T. Denoising-Autoencoder-Aided Euclidean Distance Matrix Reconstruction for Connectivity-Based Localization: A Low-Rank Perspective. Appl. Sci. 2025, 15, 2656. https://doi.org/10.3390/app15052656

AMA Style

Lee W-H, Ozger M, Challita U, Song T. Denoising-Autoencoder-Aided Euclidean Distance Matrix Reconstruction for Connectivity-Based Localization: A Low-Rank Perspective. Applied Sciences. 2025; 15(5):2656. https://doi.org/10.3390/app15052656

Chicago/Turabian Style

Lee, Woong-Hee, Mustafa Ozger, Ursula Challita, and Taewon Song. 2025. "Denoising-Autoencoder-Aided Euclidean Distance Matrix Reconstruction for Connectivity-Based Localization: A Low-Rank Perspective" Applied Sciences 15, no. 5: 2656. https://doi.org/10.3390/app15052656

APA Style

Lee, W.-H., Ozger, M., Challita, U., & Song, T. (2025). Denoising-Autoencoder-Aided Euclidean Distance Matrix Reconstruction for Connectivity-Based Localization: A Low-Rank Perspective. Applied Sciences, 15(5), 2656. https://doi.org/10.3390/app15052656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop