2.845

2023影响因子

(CJCR)

  • 中文核心
  • EI
  • 中国科技核心
  • Scopus
  • CSCD
  • 英国科学文摘

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于图像特定分类器的弱监督语义分割

郭子麟 吴东岳 高常鑫 桑农

郭子麟, 吴东岳, 高常鑫, 桑农. 基于图像特定分类器的弱监督语义分割. 自动化学报, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c240636
引用本文: 郭子麟, 吴东岳, 高常鑫, 桑农. 基于图像特定分类器的弱监督语义分割. 自动化学报, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c240636
Guo Zi-Lin, Wu Dong-Yue, Gao Chang-Xin, Sang Nong. Image-specific classifiers for weakly supervised semantic segmentation. Acta Automatica Sinica, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c240636
Citation: Guo Zi-Lin, Wu Dong-Yue, Gao Chang-Xin, Sang Nong. Image-specific classifiers for weakly supervised semantic segmentation. Acta Automatica Sinica, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c240636

基于图像特定分类器的弱监督语义分割

doi: 10.16383/j.aas.c240636 cstr: 32138.14.j.aas.c240636
基金项目: 国家自然科学基金(62176097, 61433007), 中央高校基本科研业务费(2019kfyXKJC024), 计算智能与智能控制111计划(B18024)资助
详细信息
    作者简介:

    郭子麟:华中科技大学人工智能与自动化学院博士研究生. 主要研究方向为语义分割. E-mail: zilin_guo@hust.edu.cn

    吴东岳:华中科技大学人工智能与自动化学院博士研究生. 主要研究方向为语义分割, 模型剪枝. E-mail: dongyue_wu@hust.edu.cn

    高常鑫:华中科技大学人工智能与自动化学院教授. 主要研究方向为模式识别, 视频分析. E-mail: cgao@hust.edu.cn

    桑农:华中科技大学人工智能与自动化学院教授. 主要研究方向为低质图像增强, 图像/视频语义分割, 行为检测与识别, 行人检索. 本文通信作者. E-mail: nsang@hust.edu.cn

Image-Specific Classifiers for Weakly Supervised Semantic Segmentation

Funds: Supported by National Natural Science Foundation of China (62176097, 61433007), Fundamental Research Funds for the Central Universities (2019kfyXKJC024), and 111 Project on Computational Intelligence and Intelligent Control (B18024)
More Information
    Author Bio:

    GUO Zi-Lin Ph.D. candidate at the School of Artificial Intelligence and Automation, Huazhong University of Science and Technology. His main research interest is semantic segmentation

    WU Dong-Yue Ph.D. candidate at the School of Artificial Intelligence and Automation, Huazhong University of Science and Technology. His research interest covers semantic segmentation and model pruning

    GAO Chang-Xin Professor at the School of Artificial Intelligence and Automation, Huazhong University of Science and Technology. His research interest covers pattern recognition and video analysis

    SANG Nong Professor at the School of Artificial Intelligence and Automation, Huazhong University of Science and Technology. His research interest covers low quality image enhancement, image/video semantic segmentation, action detection and recognition, and person retrieval. Corresponding author of this paper

  • 摘要: 基于图像级标签的弱监督语义分割(Weakly supervised semantic segmentation, WSSS)算法因极低的标注成本引起学界广泛关注. 该领域的算法利用分类网络产生的类激活图(Class activation maps, CAMs)实现从图像级标签到像素级标签的转化. 然而类激活图往往只关注于图像中最显著的区域, 致使基于类激活图产生的伪标签与真实标注存在较大差距, 主要包括前景未被有效激活的欠激活问题以及前景间预测混淆的错误激活问题. 欠激活源于数据集类内差异过大, 致使单一分类器不足以准确识别同一类别的所有像素, 错误激活则是数据集类间差异过小, 导致分类器不能有效区分不同类别的像素. 本文考虑到同一类别像素在图像内的差异小于在数据集中的差异, 设计基于类中心的图像特定分类器(Image-specific classifier, ISC), 以提升对同类像素的识别能力,从而改善欠激活, 同时考虑到类中心是类别在特征空间的代表, 设计类中心约束函数 (Class center constrained loss function, Lcccl), 通过扩大类中心间的差距从而间接地疏远不同类别的特征分布, 以缓解错误激活现象. 图像特定分类器可以插入其他弱监督语义分割网络, 替代分类网络的分类器, 以产生更高质量的类激活图. 实验结果表明, 本文所提出的方案在两个基准数据集上均具有良好的表现, 证实了该方案的有效性.
  • 图  1  类激活图与特征、全局分类器以及图像特定分类器在t-SNE下的可视化结果

    Fig.  1  The visualization results of CAMs, features, global Classifier, and ISC under t-SNE

    图  2  基于图像特定分类器的网络框架.

    Fig.  2  The network framework based on image-specific classifier.

    图  3  类中心生成器示意图.

    Fig.  3  The schematic diagram of the class center generator.

    图  4  不同的$ \tau $对模型性能的影响

    Fig.  4  The impact of different $ \tau $ values on the model's performance

    图  5  不同方法产生的伪标签的可视化

    Fig.  5  The visualization of the pseudo mask under different methods

    图  6  DeepLabV2在PASCAL VOC 2012验证集上的预测结果与真实标签对比

    Fig.  6  Comparison of the segmentation results with the ground truth on the PASCAL VOC 2012 validation set

    表  1  本方案中各组件性能对比

    Table  1  Comparison of the performance about each component in this approach

    R-50ISC$ L_{cccl} $监督优化mIoU(%)
    CRFIRNCRFIRN
    $ \checkmark$48.5
    $ \checkmark$$ \checkmark$52.4
    $ \checkmark$$ \checkmark$66.5
    $ \checkmark$$ \checkmark$$ \checkmark$54.3
    $ \checkmark$$ \checkmark$$ \checkmark$$ \checkmark$54.7
    $ \checkmark$$ \checkmark$$ \checkmark$59.5
    $ \checkmark$$ \checkmark$$ \checkmark$$ \checkmark$61.5
    $ \checkmark$$ \checkmark$$ \checkmark$$ \checkmark$68.7
    $ \checkmark$$ \checkmark$$ \checkmark$$ \checkmark$$ \checkmark$69.5
    下载: 导出CSV

    表  2  不同相似度度量算法对比

    Table  2  Comparison of different similarity measurement algorithms

    方法 B-Net 距离 角度 卷积
    mIoU(%) ResNet50 51.3 50.5 54.3
    下载: 导出CSV

    表  3  基于本文方法产生的类激活图以及伪标签与其他方法在PASCAL VOC 2012数据集上的质量对比 ($ \dagger $和$ * $分别表示骨架网络基于ImageNet21K预训练以及使用了显著性图, Ours表示同时使用ISC和$ L_{cccl} $)

    Table  3  Comparison about CAMs and pseudo labels generated by our method with those from other methods on the PASCAL VOC 2012 dataset ($ \dagger $ and $ * $ denotes the backbone is pretrained on ImageNet 21k and using saliency map, respectively, and Ours denotes the simultaneous use of ISC and $ L_{cccl} $)

    方法 CAM(%) 伪标签(%)
    SEAM[15] 55.4 63.6
    IRN[23] 48.3 66.5
    AdvCAM[43] 55.6 68.9
    MCTformer[26] 61.7 69.1
    FPR[44] 63.8 68.5
    SEAM+OCR[45] 67.8 68.4
    EPS*[24] 69.5 71.6
    VIT-PCM†[46] 67.7 71.4
    ToCo†[27] 73.6
    EPS*+PPC[14] 70.5 73.3
    KTSE[47] 67.0 73.8
    CLIPES[41] 70.8 75.0
    CLIPES+CPAL[42] 71.9 75.8
    IRN+Ours 61.5 69.5
    SEAM+Ours 65.5 70.7
    EPS*+Ours 71.9 74.4
    CLIPES+SEAM+Ours 69.6 76.1
    CLIPES+EPS*+Ours 73.5 76.8
    下载: 导出CSV

    表  4  与其他先进方法在PASCAL VOC 2012数据集上的分割性能对比 ($ I $, $ S $和$ L $分别表示图像级标签、显著性图和大语言模型)

    Table  4  Comparison of the segmentation performance with the state-of-the-art methods on the PASCAL VOC 2012 dataset ($ I $, $ S $ and $ L $ denotes image-level label, the saliency map and large language model, respectively)

    方法 监督 验证集(%) 测试集(%)
    IRN[23]I63.564.8
    SEAM[15]I64.565.7
    SEAM+OCR[45]I67.868.4
    IRN+ReCAM[37]I68.768.5
    IRN+LPCAM[20]I68.668.7
    CLIMS[19]I+L70.470.0
    CLIPES[41]I+L71.171.4
    MCTformer[26]I71.971.6
    KTSE[47]I73.072.9
    CTI[49]I74.173.8
    CLIPES+CPAL[42]I+L74.574.7
    AuxSegNet[48]I+S69.068.6
    Yazhou Yao[50]I+S70.470.2
    EPS[24]I+S70.970.8
    SANCE[51]I+S72.072.9
    RCA[52]I+S72.272.8
    EPS+PPC[14]I+S72.673.6
    SEAM+OursI68.970.0
    IRN+OursI69.570.0
    EPS+OursI+S73.173.5
    CLIPES+SEAM+OursI+L74.174.1
    CLIPES+EPS+OursI+S+L74.474.8
    下载: 导出CSV

    表  5  与其他先进方法在MS COCO 2014数据集上的分割性能对比 ($ I $和$ S $分别表示图像级标签和显著性图)

    Table  5  Comparison of the segmentation performance with the state-of-the-art methods on the MS COCO 2014 dataset ($ I $,$ S $ denotes image-level label and the saliency map, respectively)

    方法 骨架网络 监督 验证集(%)
    PSA[9]ResNet38I29.5
    SEAM[15]ResNet38I31.9
    SEAM+OCR[45]ResNet38I33.2
    CDA[53]ResNet38I33.2
    URN[31]ResNet101I40.7
    IRN[23]ResNet101I42.0
    IRN+ReCAM[37]ResNet101I42.9
    IRN+LPCAM[20]ResNet101I44.5
    RIB[54]ResNet101I44.5
    BECO[32]ResNet101I45.1
    EPS[24]ResNet101I+S35.7
    RCA[52]ResNet101I+S36.8
    L2G[55]ResNet101I+S44.2
    IRN+OursResNet101I44.6
    BECO+OursResNet101I45.4
    下载: 导出CSV

    表  6  与其他可插入性方法的计算开销对比

    Table  6  Comparison of computational overhead with other pluggable methods

    方法 FPS FlOPS Params Memory
    ResNet50[33] 108.2 32.6 G 23.5 M 2.02 GB
    ISC 100.2 32.8 G 111.6 M 2.36 GB
    AdvCAM[43] 0.5 423.8 G 23.5 M 2.23 GB
    ReCAM[37] 106.1 32.6 G 23.6 M 2.02 GB
    LPCAM[20] 35.3 32.7 G 24.5 M 2.37 GB
    下载: 导出CSV

    表  7  不同形式的类中心约束函数对实验结果的影响, 其中基线代表ResNet50+ISC

    Table  7  The impact of different forms of class center constrained loss function on experimental results, where the baseline represents ResNet50+ISC

    方法 基线 $ L_{cccl} $ $ L_{info} $
    $ {\boldsymbol{w}} $ $ {{\boldsymbol{w}}_{{\boldsymbol{aux}}}} $
    mIoU(%) 59.5 61.5 61.0 60.4
    下载: 导出CSV
  • [1] 徐鹏斌, 瞿安国, 王坤峰, 李大字. 全景分割研究综述. 自动化学报, 2021, 47(3): 549−568

    Xu Peng-Bin, Qu An-Guo, Wang Kun-Feng, Li Da-Zi. A survey of panoptic segmentation methods. Acta Automatica Sinica, 2021, 47(3): 549−568
    [2] 张琳, 陆耀, 卢丽华, 周天飞, 史青宣. 一种改进的视频分割网络及其全局信息优化方法. 自动化学报, 2022, 48(3): 787−796

    Zhang Lin, Lu Yao, Lu Li-Hua, Zhou Tian-Fei, Shi Qing-Xuan. An improved video segmentation network and its global information optimization method. Acta Automatica Sinica, 2022, 48(3): 787−796
    [3] 胡斌, 何克忠. 计算机视觉在室外移动机器人中的应用. 自动化学报, 2006, 32(5): 774−784

    Hu Bin, He Ke-Zhong. Applications of computer vision to outdoor mobile robot. Acta Automatica Sinica, 2006, 32(5): 774−784
    [4] Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston, USA: IEEE, 2015. 3431−3440
    [5] Zheng S, Lu J, Zhao H, Zhu X, Luo Z, Wang Y, et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 6881−6890
    [6] Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, et al. The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE, 2016. 3213−3223
    [7] Lee J, Yi J, Shin C, Yoon S. BBAM: Bounding box attribution map for weakly supervised semantic and instance segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 2643−2652
    [8] Lin D, Dai J, Jia J, He K, Sun J. ScribbleSup: Scribble-supervised convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE, 2016. 3159−3167
    [9] Ahn J, Kwak S. Learning pixel-level semantic affinity with image-level supervision for weakly supervised semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE, 2018. 4981−4990
    [10] Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE, 2016. 2921−2929
    [11] Hou Q, Jiang P, Wei Y, Cheng M. Self-erasing network for integral object attention. In: Proceedings of the Advances in Neural Information Processing Systems. Montreal, Canada: Curran Associates, 2018. 549−559
    [12] Kweon H, Yoon S, Kim H, Park D, Yoon K. Unlocking the potential of ordinary classifier: class-specific adversarial erasing framework for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Virtual Event: IEEE, 2021. 6994−7003
    [13] Chen L, Wu W, Fu C, Han X, Zhang Y. Weakly supervised semantic segmentation with boundary exploration. In: Proceedings of the European Conference on Computer Vision. Virtual Event: Springer, 2020. 347−362
    [14] Du Y, Fu Z, Liu Q, Wang Y. Weakly supervised semantic segmentation by pixel-to-prototype contrast. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 4320−4329
    [15] Wang Y, Zhang J, Kan M, Shan S, Chen X. Self-supervised equivariant attention mechanism for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2020. 12275−12284
    [16] Everingham M, Eslami S M, Van Gool L, Williams C K I, Winn J, Zisserman A. The PASCAL visual object classes challenge: a retrospective. International Journal of Computer Vision, 2015, 111(1): 98−136 doi: 10.1007/s11263-014-0733-5
    [17] Lin T Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al. Microsoft COCO: common objects in context. In: Proceedings of the European Conference on Computer Vision. Zurich, Switzerland: Springer, 2014. 740−755
    [18] Kim B, Han S, Kim J. Discriminative region suppression for weakly-supervised semantic segmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence. Virtual Event: AAAI, 2021. 1754−1761
    [19] Chen Q, Yang L, Lai J, Xie X. Self-supervised image-specific prototype exploration for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 4288−4298
    [20] Chen Z, Sun Q. Extracting class activation maps from non-discriminative features as well. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023. 3135−3144
    [21] Xie J, Hou X, Lai J, Ye K, Shen L. CLIMS: cross language image matching for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 4483−4492
    [22] Krähenbühl P, Koltun V. Efficient inference in fully connected crfs with gaussian edge potentials. In: Proceedings of the Advances in Neural Information Processing Systems. Granada, Spain: Curran Associates, 2011. 109−117
    [23] Ahn J, Cho S, Kwak S. Weakly supervised learning of instance segmentation with inter-pixel relations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Long Beach, USA: IEEE, 2019. 2209−2218
    [24] Lee S, Lee M, Lee J, Shim H. Railroad is not a train: Saliency as pseudo-pixel supervision for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 5495−5505
    [25] Wu T, Huang J, Gao G, Wei X, Wei X, Luo X, et al. Embedded discriminative attention mechanism for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 16765−16774
    [26] Xu L, Ouyang W, Bennamoun M, Boussaid F, Xu D. Multi-class token transformer for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 4310−4319
    [27] Ru L, Zheng H, Zhan Y, Du B. Token contrast for weakly-supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023. 3093−3102
    [28] Chen L, Papandreou G, Iasonas K, Murphy K, Yuille A. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 40(4): 834−848
    [29] Zhao H, Shi J, Qi X, Wang X, Jia J. Pyramid scene parsing network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE, 2017. 2881−2890
    [30] Li Y, Kuang Z, Liu L, Chen Y, Zhang W. Pseudo-mask matters in weakly-supervised semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Virtual Event: IEEE, 2021. 6964−6973
    [31] Li Y, Duan Y, Kuang Z, Chen Y, Zhang W, Li X. Uncertainty estimation via response scaling for pseudo-mask noise mitigation in weakly-supervised semantic segmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence. Virtual Event: AAAI, 2022. 1447−1455
    [32] Rong S, Tu B, Wang Z, Li J. Boundary-enhanced co-training for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023. 19574−19584
    [33] He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE, 2016. 770−778
    [34] Zhang F, Chen Y, Li Z, Hong Z, Liu J, Ma F, et al. ACFNet: Attentional class feature network for semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul, South Korea: IEEE, 2019. 6798−6807
    [35] Yuan Y, Chen X, Chen X, Wang J. Object-contextual representations for semantic segmentation. In: Proceedings of the European Conference on Computer Vision. Virtual Event: Springer, 2020. 173−190
    [36] Radford A, Kim J, Hallacy C, Ramesh A, Goh G, Agarwal S, et al. Learning transferable visual models from natural language supervision. In: Proceedings of the International Conference on Machine Learning. Virtual Event: PMLR, 2021. 8748−8763
    [37] Chen Z, Wang T, Wu X, Hua X, Zhang H, Sun Q. Class re-activation maps for weakly-supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 969−978
    [38] Hariharan B, Arbeláez P, Bourdev L, Maji S, Malik J. Semantic contours from inverse detectors. In: Proceedings of the 2011 International Conference on Computer Vision. Barcelona, Spain: IEEE, 2011. 991−998
    [39] Wu Z, Shen C, Van Den Hengel A. Wider or deeper: revisiting the resnet model for visual recognition. Pattern Recognition, DOI: 10.16383/j.aas.c210352
    [40] Deng J, Dong W, Socher R, Li J, Li K, Li F. Imagenet: A large-scale hierarchical image database. In: Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. Florida, USA: IEEE, 2009. 248−255
    [41] Lin Y, Chen M, Wang W, Wu B, Li K, Lin B, et al. Clip is also an efficient segmenter: a text-driven approach for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023. 15305−15314
    [42] Tang F, Xu Z, Qu Z, Feng W, Jiang X, Ge Z. Hunting attributes: context prototype-aware learning for wWeakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2024. 3324−3334
    [43] Lee J, Kim E, Yoon S. Anti-adversarially manipulated attributions for weakly and semi-supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 4071−4080
    [44] Chen L, Lei C, Li R, Li S, Zhang Z, Zhang L. FPR: False positive rectification for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Paris, France: IEEE, 2023. 1108−1118
    [45] Cheng Z, Qiao P, Li K, Li S, Wei P, Ji X, et al. Out-of-candidate rectification for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023. 23673−23684
    [46] Rossetti S, Zappia D, Sanzari M, Schaerf M, Pirri F. Max pooling with vision transformers reconciles class and shape in weakly supervised semantic segmentation. In: Proceedings of the European Conference on Computer Vision. Tel Aviv, Israel. Springer, 2022. 446−463
    [47] Chen T, Jiang X, Pei G, Sun Z, Wang Y, Yao Y. Knowledge transfer with simulated inter-image erasing for weakly supervised semantic segmentation. In: Proceedings of the European Conference on Computer Vision. Milan, Italy: Springer, 2025. 441−458
    [48] Xu L, Ouyang W, Bennamoun M, Boussaid F, Sohe F, Xu D. Leveraging auxiliary tasks with affinity learning for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Virtual Event: IEEE, 2021. 6984−6993
    [49] Yoon S, Kwon H, Kim H, Yoon K. Class tokens infusion for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2024. 3595−3605
    [50] Yao Y, Chen T, Xie G, Zhang C, Shen F, Wu Q, et al. Non-salient region object mining for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual Event: IEEE, 2021. 2623−2632
    [51] Li J, Fan J, Zhang Z. Towards noiseless object contours for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 16856−16865
    [52] Zhou T, Zhang M, Zhao F, Li J. Regional semantic contrast and aggregation for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 4299−4309
    [53] Su Y, Sun R, Lin G, Wu Q. Context decoupling augmentation for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. Virtual Event: IEEE, 2021. 7004−7014
    [54] Lee J, Choi J, Mok J, Yoon S. Reducing information bottleneck for weakly supervised semantic segmentation. In: Proceedings of the Advances in Neural Information Processing Systems. Virtual Event: Curran Associates, 2021. 27408−27421
    [55] Jiang P, Yang Y, Hou Q, Wei Y. L2g: A simple local-to-global knowledge transfer framework for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans, USA: IEEE, 2022. 16886−16896
    [56] Oord A, Li Y, Vinyals O. Representation learning with contrastive predictive coding. arXiv: 1807. 03748, 2018
  • 加载中
计量
  • 文章访问数:  32
  • HTML全文浏览量:  19
  • 被引次数: 0
出版历程
  • 收稿日期:  2024-10-23
  • 录用日期:  2025-01-17
  • 网络出版日期:  2025-04-13

目录

    /

    返回文章
    返回