Video Reconstruction by Spatio-Temporal Fusion of Blurred-Coded Image Pair
Anupama S
Prasan Shedligeri
Abhishek Pal
Kaushik Mitra
[Paper]
[GitHub]
Fully-Exposed Coded Fully exposed - coded (Ours)
Input
Output
24.35 dB / 0.968 28.80 dB / 0.950 34.07dB / 0.982 Ground Truth
Fully-exposed Coded Fully exposed - coded (Ours)
Input
Output
20.13dB / 0.862 33.75dB / 0.968 35.59dB / 0.978 Ground Truth

Abstract

Learning-based methods have enabled the recovery of a video sequence from a single motion-blurred image or a single coded exposure image. Recovering video from a single motion-blurred image is a very ill-posed problem and the recovered video usually has many artifacts. In addition to this, the direction of motion is lost and it results in motion ambiguity. However, it has the advantage of fully preserving the information in the static parts of the scene. The traditional coded exposure framework is better-posed but it only samples a fraction of the space-time volume, which is at best 50% of the space-time volume. Here, we propose to use the complementary information present in the fully-exposed (blurred) image along with the coded exposure image to recover a high fidelity video without any motion ambiguity. Our framework consists of a shared encoder followed by an attention module to selectively combine the spatial information from the fully-exposed image with the temporal information from the coded image, which is then super-resolved to recover a non-ambiguous high-quality video. The input to our algorithm is a fully-exposed and coded image pair. Such an acquisition system already exists in the form of a Coded-two-bucket (C2B) camera. We demonstrate that our proposed deep learning approach using blurred-coded image pair produces much better results than those from just a blurred image or just a coded image.


Talk


[Slides]

Key takeaways

  • We propose a learning-based framework for recovering video sequence from a pair of blurred-coded images.
  • The framework consists of an attenion-based feature fusion module which learns to fuse features from blurred image and the coded image.
  • We show qualitatively and quantitatively that higher-fidelity video can be recovered when using both blurred-coded images than from either blurred or coded image alone.

  • Code

    Our proposed framework first extracts a low-spatial resolution video sequence from blurred and coded images separately. A shared encoder network is used to exract features from the two video sequences. An attention-based fusion module based on a similarity measure is used to fuse features from the blurred and coded images. A deep U-Net is then used to extract the full spatial and temporal resolution video sequence from the fused features.

     [GitHub]


    Paper and Supplementary Material

    Anupama S., P. Shedligeri, Abhishek Pal, Kaushik Mitra
    Video Reconstruction by Spatio-temporal Fusion of Blurred-Coded Image Pair
    In ICPR, 2020.
    (hosted on ArXiv)


    [Supplementary Material] [Bibtex]


    Related Publications

  • Prasan Shedligeri, Anupama S & Kaushik Mitra. (2021) A Unified Framework for Compressive Video Recovery from Coded ExposureTechniques. Accepted at IEEE/CVF Winter Conference on Applications of Computer Vision, doi to be assigned [Preprint] [Slides] [Supplementary] [Code] [Webpage]
  • Prasan Shedligeri, Anupama S & Kaushik Mitra. (2021) CodedRecon: Video reconstruction for coded exposure imaging techniques. Accepted at Elsevier Journal of Software Impacts, https://doi.org/10.1016/j.simpa.2021.100064 [Paper] [Code]

  • Acknowledgements

    The authors would like to thank Sreyas Mohan and Subeesh Vasu for their helpful discussions.
    This template was originally made by Phillip Isola and Richard Zhang for a colorful ECCV project; the code can be found here.