Densely sampled light field reconstruction with transformers

JOURNAL OF ELECTRONIC IMAGING(2023)

引用 0|浏览4
暂无评分
摘要
Densely sampled light fields (LFs) are critical for their further applications, such as digital refocus and depth estimation. However, it is costly and time-consuming to capture them. LF reconstruction, which aims at reconstructing a densely sampled LF from a sparsely sampled one, has attracted extensive attention of researchers. Although existing methods have achieved significant progress, these methods synthesize novel views, either through depth estimation and image warping, which depend heavily on the accuracy of the depth maps and are prone to cause artifacts at occluded regions, or by stacking multi-layer convolutions to learn the inherent structure of the LF, which will result in blurring results due to limited receptive fields when processing scenes with large disparities. We propose a transformer-based neural network for LF reconstruction (termed as LFRTR). Specifically, two novel transformers are introduced, namely angular transformer and spatial transformer. The former can fully explore angular information and correlations among different views, whereas the latter can capture local and non-local spatial texture information within each view. Moreover, dense skip connections are employed to enhance information flow between different layers. Thanks to the inherent global modeling ability of self-attention, the proposed LFRTR can reconstruct high-quality densely sampled LF in complex scenarios, such as large disparity, occlusion, and reflection. Experimental results on both synthetic and real-world LF datasets show that the proposed LFRTR outperforms other state-of-the-art methods in terms of both visual and numerical evaluations. (C) 2023 SPIE and IS&T
更多
查看译文
关键词
light field,angular super-resolution,reconstruction,transformer
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要