Skip to content

Guided high dynamic range (HDR) imaging using single-photon avalanche diodes

License

Notifications You must be signed in to change notification settings

Way-Yuhao/Single-Photon-Guided-HDR

Repository files navigation

[WACV 2022] Single Photon Camera Guided HDR Imaging

Project webpage: https://www.yuhaoliu.net/spc-guided-hdr
Paper: https://openaccess.thecvf.com/content/WACV2022/html/Liu_Single-Photon_Camera_Guided_Extreme_Dynamic_Range_Imaging_WACV_2022_paper.html

1. Getting Started

Clone the repo:

git clone https://github.com/Way-Yuhao/Single-Photon-Guided-HDR.git

2. Requirements

python>=3.6
torch>=0.4.0
torchvision
torchsummary
tensorboardx
tqdm
natsort
numpy
pillow
scipy
scikit-image
sklearn

Install all dependent libraries:

pip install -r requirements.txt

Input file directory structure

input directory

CMOS
SPAD
ideal (optional)

Both CMOS and SPAD inputs are required. A ground truth (ideal) image can be added to compute the loss of network output.

CMOS & SPAD image requirements

  • To properly use our provided test.py testing script, it is required to know the constant CMOS saturation limit for all CMOS images. This value is used to normalize input data for the neural network.

3. Usage

Pretrained model

TODO: define pre-trained model

Inference

python test.py --input <input/dir> --out <output/dir> --weights <weight/path>.pth --saturation <saturation/float>

Parameters and their descriptions:

input: directory of input images.
out: path to output directory.
weights: path to the pretrained neural network weights.
cpu: toggle to use CPU only.
saturation: an int or float defining CMOS saturation limit.
gain: a float defining gain applied to SPAD images during data loading stage; for debugging only.
experimental: toggle if testing on experimental data / no ground truth provided.
plot: toggle if intend to use plt.show() to visualize outputs.

4. Citation

If you find this useful, please cite

@InProceedings{Liu_2022_WACV,
    author    = {Liu, Yuhao and Gutierrez-Barragan, Felipe and Ingle, Atul and Gupta, Mohit and Velten, Andreas},
    title     = {Single-Photon Camera Guided Extreme Dynamic Range Imaging},
    booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)},
    month     = {January},
    year      = {2022},
    pages     = {1575-1585}
}

5. Acknowledgment

This work was supported in part by the U.S. Department of Energy/NNSA (DE-NA0003921), National Science Foundation (CAREER 1846884 and 1943149), and UW-Madison’s Draper Technology Innovation Fund. The authors would like to thank the Computational Imaging Group at Rice University for providing conference travel funds for Yuhao Liu. U.S. DoE full legal disclaimer: https://www.osti.gov/stip/about/disclaimer.

We used a standard unet (https://github.com/bigmb/Unet-Segmentation-Pytorch-Nest-of-Unets) as a starting point for our network. Our latest version does not contain any code from the standard unet.

About

Guided high dynamic range (HDR) imaging using single-photon avalanche diodes

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages