留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于注意力残差网络的快照式多光谱相机图像重构

闫纲琦 梁宗林 宋延嵩 董科研 张博 刘天赐 张雷 王岩柏

闫纲琦, 梁宗林, 宋延嵩, 董科研, 张博, 刘天赐, 张雷, 王岩柏. 基于注意力残差网络的快照式多光谱相机图像重构[J]. 中国光学(中英文). doi: 10.37188/CO.2023-0196
引用本文: 闫纲琦, 梁宗林, 宋延嵩, 董科研, 张博, 刘天赐, 张雷, 王岩柏. 基于注意力残差网络的快照式多光谱相机图像重构[J]. 中国光学(中英文). doi: 10.37188/CO.2023-0196
YAN Gang-qi, LIANG Zong-lin, SONG Yan-song, DONG Ke-yan, ZHANG Bo, LIU Tian-ci, ZHANG LEI, WANG Yan-bo. Reconstruction of snapshot multispectral camera images based on an attention residual network[J]. Chinese Optics. doi: 10.37188/CO.2023-0196
Citation: YAN Gang-qi, LIANG Zong-lin, SONG Yan-song, DONG Ke-yan, ZHANG Bo, LIU Tian-ci, ZHANG LEI, WANG Yan-bo. Reconstruction of snapshot multispectral camera images based on an attention residual network[J]. Chinese Optics. doi: 10.37188/CO.2023-0196

基于注意力残差网络的快照式多光谱相机图像重构

doi: 10.37188/CO.2023-0196
基金项目: 国家重点研发计划(No. 2022YFB3902500);国家重点研发计划项目(No. 2021YFA0718804);国家自然科学基金青年基金(No. 62305032)
详细信息
    作者简介:

    闫纲琦(1996—),男,河北张家口人,长春理工大学硕士研究生,主要从事图像处理方面的研究。E-mail:y15032865361@163.com

    宋延嵩(1983—),男,吉林长春人,博士,研究员,博士生导师,2006年、2009年、2014年于长春理工大学分别获得学士、硕士、及博士学位,主要研究方向为空间激光通信技术。E-mail:songyansong2006@126.com

  • 中图分类号: TP394.1;TH691.9

Reconstruction of snapshot multispectral camera images based on an attention residual network

Funds: Supported by National key research and development program (No. 2022YFB3902500); National key research and development program (No. 2021YFA0718804); National Science Foundation of China (No. 62305032)
More Information
  • 摘要:

    随着光谱成像技术的飞速发展,使用多光谱滤光片阵列(Multispectral filter array,MSFA)采集多光谱图像的空间和光谱信息已经成为研究热点。如何利用低采样率且强频谱互相关性的原始数据进行重构成为了瓶颈和制约。因此基于一种含有全通波段的8波段4×4MSFA,提出了空谱联合的多分支注意力残差网络模型。首先使用多分支模型对各个波段插值后的图像特征进行学习。其次八个波段和全通波段的特征信息联合通过本文设计的空间通道注意力模型,应用多层卷积和卷积注意力模块并使用残差补偿的方式可以有效的弥补各波段的颜色差异和丰富边缘纹理相关特征信息。最后初步插值的全通波段和其余波段特征信息通过无需进行批量归一化的残差密集块对多光谱图像空间和光谱相关性进行特征学习以匹配各个波段的光谱信息。实验结果表明,本文对于在D65光源下测试图像的峰值信噪比、结构相似度和光谱角相似度分别优于最先进深度学习方法3.46%、0.27%和6%,并且此方法不仅减少了伪影还获得更多的纹理细节。

     

  • 图 1  配备MSFA的单传感器多光谱相机的原始图像采集与重构过程

    Figure 1.  Original image acquisition and reconstruction process of a single sensor multispectral camera equipped with MSFA

    图 2  多分支注意力残差网络模型框架

    Figure 2.  A multi-branch attention residual network model framework

    图 3  马赛克通道卷积块结构图

    Figure 3.  Mosaic channel convolutional block structure diagram

    图 4  空间通道注意模型结构图

    Figure 4.  Structural diagram of the spatial channel attention model

    图 5  D65光源下的测试图像在sRGB颜色空间中的去马赛克效果视觉对比

    Figure 5.  Visual comparison of the de-mosaicing effect of test images under the D65 light source in sRGB color space

    图 6  测试图像在679 nm处不同场景的去马赛克误差图视觉比较

    Figure 6.  Visual comparison of de-mosaicing error maps of test images in different scenarios at 679 nm

    图 7  测试图像CD不同波段的去马赛克误差图视觉比较

    Figure 7.  Visual comparison of de-mosaicing error maps in different bands of test image CD

    表  1  测试图像在A光源下的PSNR、SSIM和SAM值

    Table  1.   PSNR, SSIM, and SAM values of the test image under the A light source

    Methods PSNR ↑ SSIM ↑ SAM ↓
    WB PPID MGCC Ours WB PPID MGCC Ours WB PPID MGCC Ours
    balloons 39.99 43.23 45.52 45.93 0.9977 0.9988 0.9992 0.9993 4.892 3.693 3.409 3.329
    beads 28.35 31.03 32.75 33.57 0.9610 0.9768 0.9776 0.9858 10.087 8.002 7.571 6.182
    Egyptian 35.22 40.77 40.96 43.65 0.9908 0.9947 0.9953 0.9972 11.358 10.648 8.653 7.570
    feathers 32.38 36.32 37.03 38.48 0.9908 0.9962 0.9964 0.9976 8.351 6.782 6.036 5.960
    paints 32.39 36.29 35.79 38.63 0.9939 0.9973 0.9973 0.9984 6.557 5.178 4.891 4.500
    pompoms 36.37 38.28 40.54 41.02 0.9953 0.9968 0.9978 0.9987 4.395 3.524 3.003 2.916
    CD 36.54 38.54 39.01 39.75 0.9939 0.9953 0.9956 0.9969 4.211 4.043 3.956 4.145
    Character 29.27 34.91 38.91 39.51 0.9942 0.9981 0.9989 0.9992 6.103 3.811 3.119 3.049
    ChartRes 30.24 31.26 31.45 32.63 0.9931 0.9976 0.9984 0.9996 3.962 2.761 2.529 1.300
    Average 33.41 36.73 37.99 39.24 0.9900 0.9946 0.9951 0.9969 6.657 5.383 4.796 4.327
    下载: 导出CSV

    表  2  测试图像在D65光源下的PSNR、SSIM和SAM值

    Table  2.   PSNR, SSIM, and SAM values of the test image under the D65 light source

    MethodsPSNR ↑SSIM ↑SAM ↓
    WBPPIDMGCCOursWBPPIDMGCCOursWBPPIDMGCCOurs
    balloons40.7344.1646.9448.360.99810.99910.99950.99973.6293.1142.7002.539
    beads25.7531.2833.2134.060.96620.98260.98080.989810.3168.3246.9576.243
    Egyptian39.2942.542.5844.860.99290.98750.99650.997910.6428.0157.0386.919
    feathers32.6736.7237.2639.120.99160.99310.99660.99807.3775.8915.4124.843
    paints31.1836.0736.4138.900.99220.99740.98870.99866.4084.7883.9743.905
    pompoms36.8838.9041.1942.650.99620.99780.99850.99894.0323.2322.6932.753
    CD34.5438.2140.6141.530.99490.99520.99610.99673.2092.9422.7722.726
    Character28.8834.9039.6339.870.99360.99840.99940.99966.8674.1003.1082.907
    ChartRes29.3429.6532.5433.210.99510.99580.99690.99712.8402.2901.5611.324
    Average33.2536.9338.9340.280.99120.99410.99470.99746.1474.7444.0233.795
    下载: 导出CSV

    表  3  测试图像在F12光源下的PSNR、SSIM和SAM值

    Table  3.   PSNR, SSIM, and SAM values of the test image under the F12 light source

    MethodsPSNR ↑SSIM ↑SAM ↓
    WBPPIDMGCCOursWBPPIDMGCCOursWBPPIDMGCCOurs
    balloons36.6837.8338.6038.910.99370.99520.99610.99895.5405.6405.5375.955
    beads25.8427.6527.6729.920.93900.95660.96730.972014.95813.66913.53211.424
    Egyptian36.2337.4638.2339.470.98250.98570.98620.99136.1505.7404.7633.795
    feathers30.6132.7633.5134.700.98290.98870.98890.992811.52611.19411.25110.897
    paints28.0230.8931.9833.930.97610.98710.99120.99399.8019.4409.4159.113
    pompoms32.1732.8133.9233.970.98740.98890.99070.99096.7766.3556.1505.936
    CD35.5835.9936.2637.060.98690.98720.98980.99887.9346.6915.9576.622
    Character26.1128.8632.3332.640.98240.99080.99540.99627.7946.8816.8356.767
    ChartRes25.8126.3227.6129.330.97680.98140.98260.99163.8453.1572.8371.859
    Average30.7832.2833.3434.440.97860.98460.98750.99188.2587.6407.3646.929
    下载: 导出CSV

    表  4  测试图像在三种光源下的PSNR、SSIM和SAM与不同去马赛克方法的定量比较

    Table  4.   Quantitative comparison of PSNR, SSIM, and SAM of test images under three light sources and different de-mosaicing methods

    WBPPIDMGCCOurs
    PSNR29.3734.5635.6937.81
    SSIM0.89260.97050.99020.9936
    SAM8.5416.4735.6475.312
    下载: 导出CSV

    表  5  不同去马赛克方法运行时间比较(单位:ms)

    Table  5.   Comparison of running times of different de-mosaicing methods(Unit:ms)

    WBPPIDMGCCOurs
    CPU254.352134.5--
    GPU--2.652.12
    下载: 导出CSV

    表  6  不同网络架构的消融研究

    Table  6.   Research on the ablation of different network architectures

    多分支架构残差密集块PSNRSSIMSAM
    ΟΟ36.940.98697.263
    ΟΠ37.210.99076.726
    ΠΠ38.590.99386.218
    下载: 导出CSV

    表  7  不同注意力机制方案的消融研究

    Table  7.   Ablation study of different attention mechanism schemes

    注意力机制PSNRSSIMSAM
    RN36.240.98048.352
    CA36.790.98736.316
    MAM37.280.99276.047
    SCAM38.460.99515.247
    下载: 导出CSV
  • [1] JUNIOR J D D, BACKES A R, ESCARPINATI M C. Detection of control points for UAV-multispectral sensed data registration through the combining of feature descriptors[C]. Proceedings of the 14th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, SciTePress.
    [2] MANGAI U G, SAMANTA S, DAS S, et al. A hierarchical multi-classifier framework for landform segmentation using multi-spectral satellite images - a case study over the Indian subcontinent[C]. 2010 Fourth Pacific-Rim Symposium on Image and Video Technology, IEEE, 2010: 306-313.
    [3] LIU CH H, LIU W, LU X ZH, et al. Application of multispectral imaging to determine quality attributes and ripeness stage in strawberry fruit[J]. PLoS One, 2014, 9(2): e87818. doi: 10.1371/journal.pone.0087818
    [4] CHEN I T, LIN H Y. Detection, counting and maturity assessment of cherry tomatoes using multi-spectral images and machine learning techniques[C]. Proceedings of the 15th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, SciTePress, 2020: 759-766.
    [5] CHANG K, LI H X, TAN Y F, et al. A two-stage convolutional neural network for joint demosaicking and super-resolution[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(7): 4238-4254. doi: 10.1109/TCSVT.2021.3129201
    [6] SHINODA K, OGAWA S, YANAGI Y, et al. Multispectral filter array and demosaicking for pathological images[C]. 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), IEEE, 2015: 697-703.
    [7] ZENTENO O, TREUILLET S, LUCAS Y. 3D cylinder pose estimation by maximization of binary masks similarity: a simulation study for multispectral endoscopy image registration[C]. Proceedings of the 14th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, SciTePress, 2019: 857-864.
    [8] 李云辉. 压缩光谱成像系统中物理实现架构研究综述[J]. 中国光学(中英文),2022,15(5):929-945. doi: 10.37188/CO.2022-0104

    LI Y H. Review of physical implementation architecture in compressive spectral imaging system[J]. Chinese Optics, 2022, 15(5): 929-945. doi: 10.37188/CO.2022-0104
    [9] ZHUANG L N, NG M K, FU X Y, et al. Hy-demosaicing: hyperspectral blind reconstruction from spectral subsampling[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 1-15. doi: 10.1109/TGRS.2021.3102136
    [10] ZHANG J CH, CHEN J L, YU H W, et al. Polarization image demosaicking via nonlocal sparse tensor factorization[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 1-10. doi: 10.1109/TGRS.2021.3093903
    [11] RATHI V, GOYAL P. Multispectral image demosaicking based on novel spectrally localized average images[J]. IEEE Signal Processing Letters, 2022, 29: 449-453. doi: 10.1109/LSP.2021.3139581
    [12] RATHI V, GOYAL P. Generic multispectral demosaicking using spectral correlation between spectral bands and pseudo-panchromatic image[J]. Signal Processing:Image Communication, 2023, 110: 116893. doi: 10.1016/j.image.2022.116893
    [13] LIU SH M, ZHANG Y G, CHEN J, et al. A deep joint network for multispectral demosaicking based on pseudo-panchromatic images[J]. IEEE Journal of Selected Topics in Signal Processing, 2022, 16(4): 622-635. doi: 10.1109/JSTSP.2022.3172865
    [14] ZHANG Y, SUN W J, CHEN ZH ZH. Joint image demosaicking and denoising with mutual guidance of color channels[J]. Signal Processing, 2022, 200: 108674. doi: 10.1016/j.sigpro.2022.108674
    [15] AGGARWAL H K, MAJUMDAR A. Single-sensor multi-spectral image demosaicing algorithm using learned interpolation weights[C]. 2014 IEEE Geoscience and Remote Sensing Symposium, IEEE, 2014: 2011-2014.
    [16] CHINI M, CHIANCONE A, STRAMONDO S. Scale Object Selection (SOS) through a hierarchical segmentation by a multi-spectral per-pixel classification[J]. Pattern Recognition Letters, 2014, 49: 214-223. doi: 10.1016/j.patrec.2014.07.012
    [17] MIHOUBI S, LOSSON O, MATHON B, et al. Multispectral demosaicing using pseudo-panchromatic image[J]. IEEE Transactions on Computational Imaging, 2017, 3(4): 982-995. doi: 10.1109/TCI.2017.2691553
    [18] 齐海超, 宋延嵩, 张博, 等. 基于改进引导滤波器的多光谱去马赛克方法[J]. 中国光学(中英文),2023,16(5):1056-1065. doi: 10.37188/CO.2022-0231

    QI H CH, SONG Y S, ZHANG B, et al. Multispectral demosaicing method based on an improved guided filter[J]. Chinese Optics, 2023, 16(5): 1056-1065. doi: 10.37188/CO.2022-0231
    [19] CHEN L K, ZHAO Y Q, CHAN J C W, et al. Histograms of oriented mosaic gradients for snapshot spectral image description[J]. ISPRS Journal of Photogrammetry and Remote Sensing, 2022, 183: 79-93. doi: 10.1016/j.isprsjprs.2021.10.018
    [20] XIONG F CH, ZHOU J, QIAN Y T. Material based object tracking in hyperspectral videos: benchmark and algorithms[Z]. arXiv: 1812.04179, 2019.
    [21] WANG X H, CHEN J, WEI Q, et al. Hyperspectral image super-resolution via deep prior regularization with parameter estimation[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(4): 1708-1723. doi: 10.1109/TCSVT.2021.3078559
    [22] HU J, JIA X P, LI Y S, et al. Hyperspectral image super-resolution via intrafusion network[J]. IEEE Transactions on Geoscience and Remote Sensing, 2020, 58(10): 7459-7471. doi: 10.1109/TGRS.2020.2982940
    [23] KUMAR S P P, PETER K J, KINGSLY C S. De-noising and demosaicking of Bayer image using deep convolutional attention residual learning[J]. Multimedia Tools and Applications, 2023, 82(13): 20323-20342. doi: 10.1007/s11042-023-14334-z
    [24] KURNIAWAN E, PARK Y, LEE S. Noise-resistant demosaicing with deep image prior network and random RGBW color filter array[J]. Sensors, 2022, 22(5): 1767. doi: 10.3390/s22051767
    [25] SHINODA K, YOSHIBA S, HASEGAWA M. Deep demosaicking for multispectral filter arrays[Z]. arXiv: 1808.08021, 2018.
    [26] PAN ZH H, LI B P, BAO Y Z, et al. Deep panchromatic image guided residual interpolation for multispectral image demosaicking[C]. 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), IEEE, 2019: 1-5.
    [27] FENG K, ZHAO Y Q, CHAN J C W, et al. Mosaic convolution-attention network for demosaicing multispectral filter array images[J]. IEEE Transactions on Computational Imaging, 2021, 7: 864-878. doi: 10.1109/TCI.2021.3102052
    [28] DIJKSTRA K, VAN DE LOOSDRECHT J, SCHOMAKER L R B, et al. Hyperspectral demosaicking and crosstalk correction using deep learning[J]. Machine Vision and Applications, 2019, 30(1): 1-21. doi: 10.1007/s00138-018-0965-4
    [29] WANG ZH, BOVIK A C, SHEIKH H R, et al. Image quality assessment: from error visibility to structural similarity[J]. IEEE Transactions on Image Processing, 2004, 13(4): 600-612. doi: 10.1109/TIP.2003.819861
    [30] KRUSE F A, LEFKOFF A B, BOARDMAN J W, et al. The spectral image processing system (SIPS)-interactive visualization and analysis of imaging spectrometer data[J]. AIP Conference Proceedings, 1993, 283(1): 192-201.
    [31] WANG X T, YU K, WU SH X, et al. ESRGAN: Enhanced super-resolution generative adversarial networks[C]. European Conference on Computer Vision, Springer, 2018: 63-79.
    [32] ZHAO H, GALLO O, FROSIO I, et al. Loss functions for image restoration with neural networks[J]. IEEE Transactions on Computational Imaging, 2017, 3(1): 47-57. doi: 10.1109/TCI.2016.2644865
    [33] KIM J, LEE J K, LEE K M. Deeply-recursive convolutional network for image super-resolution[C]. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2016.
    [34] KINGMA D P, BA L J. Adam: A method for stochastic optimization[C]. 3rd International Conference on Learning Representations (ICLR), ICLR, 2017.
    [35] WOO S, PARK J, LEE J Y, et al. CBAM: convolutional block attention module[M]. FERRARI V, HEBERT M, SMINCHISESCU C, et al. Computer Vision – ECCV 2018. Cham: Springer, 2018: 3-19.
    [36] YASUMA F, MITSUNAGA T, ISO D, et al. Generalized assorted pixel camera: postcapture control of resolution, dynamic range, and spectrum[J]. IEEE Transactions on Image Processing, 2010, 19(9): 2241-2253. doi: 10.1109/TIP.2010.2046811
    [37] MONNO Y, KIKUCHI S, TANAKA M, et al. A practical one-shot multispectral imaging system using a single image sensor[J]. IEEE Transactions on Image Processing, 2015, 24(10): 3048-3059. doi: 10.1109/TIP.2015.2436342
    [38] HU J, SHEN L, ALBANIE S, et al. Squeeze-and-excitation networks[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020, 42(8): 2011-2023. doi: 10.1109/TPAMI.2019.2913372
  • 加载中
图(7) / 表(7)
计量
  • 文章访问数:  59
  • HTML全文浏览量:  32
  • PDF下载量:  6
  • 被引次数: 0
出版历程
  • 网络出版日期:  2024-02-01

目录

    /

    返回文章
    返回

    重要通知

    2024年2月16日科睿唯安通过Blog宣布,2024年将要发布的JCR2023中,229个自然科学和社会科学学科将SCI/SSCI和ESCI期刊一起进行排名!《中国光学(中英文)》作为ESCI期刊将与全球SCI期刊共同排名!