recognition tasks and leads to a degradation in
recognition performance.
4.6 Discussion
Our approach improves accuracy without retraining
the recognition model by incorporating exposure
recovery and noise reduction into the image
processing pipeline. This suggests that by applying
our proposed method, it became possible to retrieve
the overlooked features in Lee et al.'s pose estimation.
To explore the potential performance of our proposed
method, we conducted a grid search on the entire test
data of ExLPose. In this process, we searched for
optimal parameters for each input image and
processed the input images in the differentiable image
processing module. The processed images were then
input into Lee et al.'s recognition model. The results
are presented in Table 4. As evident from the results
of the preliminary experiment, the pose estimation
accuracy significantly improved across all subsets.
This suggests the potential to further enhance the
performance of the proposed method. The refinement
of the training method for the optimal parameter
predictor will be a future task.
5 CONCLUSIONS
We proposed an image-adaptive learnable module
that improves recognition performance in low-light
environments without retraining the pretrained
recognition model for pose estimation. Our proposed
method consists of a differentiable image processing
module and an optimal parameter predictor. The
Differentiable image processing module restores the
exposure and remove noise from low-light images to
recover the latent content of the images. The optimal
parameter predictor predicts the optimal
hyperparameters used in the modules by using a small
FCN. The entire framework was trained end-to-end,
and the optimal parameter predictor learned to predict
appropriate hyperparameters by referring only to the
loss of the subsequent pose estimation task in this
paper. The experimental results demonstrated that our
approach achieved a maximum recovery of up to
11.1% in the accuracy of pretrained pose estimation
models across different levels of low-light image data.
REFERENCES
Land, E. H. (1977). The retinex theory of color vision.
Scientific American, 237(6), 108-129.
Lore, K. G., Akintayo, A., & Sarkar, S. (2017). LLNet: A
deep autoencoder approach to natural low-light image
enhancement. Pattern Recognition, 61, 650-662.
Shen, L., Yue, Z., Feng, F., Chen, Q., Liu, S., & Ma, J.
(2017). Msr-net: Low-light image enhancement using
deep convolutional network. arXiv preprint
arXiv:1711.02488.
Lv, F., Lu, F., Wu, J., & Lim, C. (2018, September).
MBLLEN: Low-Light Image/Video Enhancement
Using CNNs. In BMVC (Vol. 220, No. 1, p. 4).
Wei, C., Wang, W., Yang, W., & Liu, J. (2018). Deep
retinex decomposition for low-light enhancement.
arXiv preprint arXiv:1808.04560.
Zhang, Y., Zhang, J., & Guo, X. (2019, October). Kindling
the darkness: A practical low-light image enhancer. In
Proceedings of the 27th ACM international conference
on multimedia (pp. 1632-1640).
Lee, S., Rim, J., Jeong, B., Kim, G., Woo, B., Lee, H., ... &
Kwak, S. (2023). Human pose estimation in extremely
low-light conditions. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern
Recognition (pp. 704-714).
Tomasi, C., & Manduchi, R. (1998, January). Bilateral
filtering for gray and color images. In Sixth
international conference on computer vision (IEEE Cat.
No. 98CH36271) (pp. 839-846). IEEE.
Liu, W., Ren, G., Yu, R., Guo, S., Zhu, J., & Zhang, L.
(2022, June). Image-adaptive YOLO for object
detection in adverse weather conditions. In Proceedings
of the AAAI Conference on Artificial Intelligence (Vol.
36, No. 2, pp. 1792-1800).
Mosleh, A., Sharma, A., Onzon, E., Mannan, F., Robidoux,
N., & Heide, F. (2020). Hardware-in-the-loop end-to-
end optimization of camera image processing pipelines.
In Proceedings of the IEEE/CVF Conference on
computer Vision and Pattern Recognition (pp. 7529-
7538).