Optimizing Synchronization Times for Position Tracking of a Mobile Asset in GPS-denied Environments.

GLOBECOM(2022)

Cited 0|Views14
No score
Abstract
A network of distributed agents operating in a GPS-denied environment is tasked with tracking the position of a mobile asset. The agents use time-of-flight (ToF) measurements obtained from transmissions of a beacon signal by the asset to estimate the asset's position, but the estimates are noisy due to clock drift at the agents' local clocks. The clock drift can be reduced by having the asset and network of agents perform a distributed synchronization process; however, synchronization and localization cannot be performed simultaneously because the clock states are not well defined during synchronization and due to constraints on the radios' communication resources. The problem of optimizing when the agents should perform sensing or synchronization is formulated as a partially observed Markov decision process (POMDP) with continuous observations. Table-based Q-learning is used to search for an optimal policy on the belief space of the POMDP, but some form of approximation must be used because the belief space is a high-dimensional continuous space. We compare the performance of two approaches: 1) in replicated Q-learning, we learn a Q function that is a linear function of the continuous beliefs; 2) in triple-Q learning, the beliefs are replaced by a model with fewer parameters, and quantization is used on a continuous parameter. Simulations are used to compare the performance of these methods with the approach that is most commonly used, which is using periodic synchronization at an optimized, fixed rate.
More
Translated text
Key words
synchronization times,position tracking,mobile asset,gps-denied
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined