The official implementation of ACMMM 2022 Award Candidates "Learnability Enhancement for Low-light Raw Denoising: Where Paired Real Data Meets Noise Modeling" 【Arxiv / ACM DL】
Presentation video for the Best Paper Session: YouTube / BiliBili
Interested readers are also referred to our official Note about this work in Zhihu (Chinese).
Please switch to the TPAMI branch to get the latest version of PMN
TPAMI branch is compatible with this branch on evaluation, but has little difference on training
(2023.08.30): 📰 We have written an official blog post (in Chinese) for the TPAMI version of the paper. I believe this blog will help you gain a deeper understanding of our work.
(2023.08.03): 🎉 Our paper was accepted by IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 2023~~
- We present a learnability enhancement strategy to reform paired real data according to noise modeling. Our strategy consists of two efficient techniques: shot noise augmentation (SNA) and dark shading correction (DSC).
- SNA improves the precision of data mapping by increasing the data volume. Benefiting from the increased data volume, the mapping can promote the denoised images with clear texture.
- DSC reduces the complexity of data mapping by reducing the noise complexity. Benefiting from the reduced noise complexity, the mapping can promote the denoised images with exact colors.
- Python >=3.6, PyTorch >= 1.6
- Requirements: opencv-python, rawpy, exifread, h5py, scipy
- Platforms: Ubuntu 16.04, cuda-10.1
- Our method can run on the CPU, but we recommend you run it on the GPU
Please download the ELD dataset and SID dataset first, which are necessary for validation (or training).
ELD (official project): download (11.46 GB)
SID (official project): download (25 GB)
- use
get_dataset_infos.py
to generate dataset infos
# Evaluate
python3 get_dataset_infos.py --dstname ELD --root_dir /data/ELD --mode SonyA7S2
python3 get_dataset_infos.py --dstname SID --root_dir /data/SID/Sony --mode evaltest
# Train
python3 get_dataset_infos.py --dstname SID --root_dir /data/SID/Sony --mode train
- evaluate
# If you don't want to save pictures, please add '--save_plot False'. This option will save your time and space.
# ELD & SID
python3 trainer_SID.py -f runfiles/Ours.yml --mode evaltest
# ELD only
python3 trainer_SID.py -f runfiles/Ours.yml --mode eval
# SID only
python3 trainer_SID.py -f runfiles/Ours.yml --mode test
- train
python3 trainer_SID.py -f runfiles/Ours.yml --mode train
The parameter sampling of SNA is implemented as the raw_wb_aug_torch
function in the file data_process/process.py
.
The complete process of SNA has the CPU version in the Mix_Dataset
class in data_process/real_datasets.py
and the GPU version in the preprocess
function in trainer_SID.py
.
Both dark shading calibration and noise calibration require massive dark frames. We provide the calibration results directly. The calibration results for dark shading are stored in the resources
folder.
The raw noise parameters at each ISO are stored in the get_camera_noisy_params_max
function in process.py
, which can be used to calibrate the noise parameters based on a noise model (P-G or ELD).
HINT: The calibration is based on a SonyA7S2 camera, which has the same sensor as the public datasets but not the same camera.
Dataset | Ratio | Index | P-G | ELD | SFRN | Paired | Ours |
---|---|---|---|---|---|---|---|
ELD | ×100 | PSNR | 42.05 | 45.45 | 46.02 | 44.47 | 46.50 |
SSIM | 0.872 | 0.975 | 0.977 | 0.968 | 0.985 | ||
×200 | PSNR | 38.18 | 43.43 | 44.10 | 41.97 | 44.51 | |
SSIM | 0.782 | 0.954 | 0.964 | 0.928 | 0.973 | ||
SID | ×100 | PSNR | 39.44 | 41.95 | 42.29 | 42.06 | 43.16 |
SSIM | 0.890 | 0.963 | 0.951 | 0.955 | 0.960 | ||
×250 | PSNR | 34.32 | 39.44 | 40.22 | 39.60 | 40.92 | |
SSIM | 0.768 | 0.931 | 0.938 | 0.938 | 0.947 | ||
×300 | PSNR | 30.66 | 36.36 | 36.87 | 36.85 | 37.77 | |
SSIM | 0.657 | 0.911 | 0.917 | 0.923 | 0.934 |
Note: The quantitative results on the SID dataset is different from the provided results in ELD(TPAMI) because only the central area is compared in ELD(TPAMI) on the SID dataset.
w/o brighteness alignment provided by ELD
Dataset | Ratio | Index | P-G | ELD | SFRN | Paired | Ours |
---|---|---|---|---|---|---|---|
ELD | ×100 | PSNR | 39.44 | 45.06 | 45.47 | 43.80 | 45.94 |
SSIM | 0.780 | 0.975 | 0.976 | 0.963 | 0.984 | ||
×200 | PSNR | 33.76 | 43.21 | 43.65 | 41.54 | 44.00 | |
SSIM | 0.609 | 0.954 | 0.962 | 0.918 | 0.968 | ||
SID | ×100 | PSNR | 37.50 | 41.21 | 41.38 | 41.39 | 42.65 |
SSIM | 0.856 | 0.952 | 0.949 | 0.954 | 0.959 | ||
×250 | PSNR | 31.67 | 38.54 | 39.48 | 38.90 | 40.39 | |
SSIM | 0.765 | 0.929 | 0.937 | 0.937 | 0.946 | ||
×300 | PSNR | 28.53 | 35.35 | 35.96 | 36.55 | 37.23 | |
SSIM | 0.667 | 0.908 | 0.915 | 0.922 | 0.933 |
If you find our code helpful in your research or work please cite our paper.
@inproceedings{feng2022learnability,
author = {Feng, Hansen and Wang, Lizhi and Wang, Yuzhi and Huang, Hua},
title = {Learnability Enhancement for Low-Light Raw Denoising: Where Paired Real Data Meets Noise Modeling},
booktitle = {Proceedings of the 30th ACM International Conference on Multimedia},
year = {2022},
pages = {1436–1444},
numpages = {9},
location = {Lisboa, Portugal},
series = {MM '22}
}
@ARTICLE{feng2023learnability,
author={Feng, Hansen and Wang, Lizhi and Wang, Yuzhi and Fan, Haoqiang and Huang, Hua},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Learnability Enhancement for Low-Light Raw Image Denoising: A Data Perspective},
year={2024},
volume={46},
number={1},
pages={370-387},
doi={10.1109/TPAMI.2023.3301502}
}
If you would like to get in-depth help from me, please feel free to contact me ([email protected] / [email protected]) with a brief self-introduction (including your name, affiliation, and position).