A Fast Parallel Particle Filter For Shared Memory Systems

IEEE SIGNAL PROCESSING LETTERS(2020)

引用 10|浏览9
暂无评分
摘要
Particle Filters (PFs) are Sequential Monte Carlo methods which are widely used to solve filtering problems of dynamic models under Non-Linear Non-Gaussian noise. Modern PF applications have demanding accuracy and run-time constraints that can be addressed through parallel computing. However, an efficient parallelization of PFs can only be achieved by effectively parallelizing the bottleneck: resampling and its constituent redistribution step. A pre-existing implementation of redistribute on Shared Memory Architectures (SMAs) achieves O(N/T log(2)N) time complexity over T parallel cores. This redistribute implementation is, however, highly computationally intensive and cannot be effectively parallelized due to the inherently limited number of cores of SMAs. In this paper, we propose a novel parallel redistribute on OpenMP 4.5 which takes O(N/T + log(2)N) steps and fully exploits the computational power of SMAs. The proposed approach is up to six times faster than the O(N/T log(2)N) one and its implementation on GPU provides a further three-time speed-up vs its equivalent on a 32-core CPU. We also show on an exemplary PF that our redistribution is no longer the bottleneck.
更多
查看译文
关键词
Parallel particle filters, shared memory architectures, OpenMP, resampling, redistribute
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要