skip to main content
10.1145/3580305.3599316acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Free Access

Doubly Robust AUC Optimization against Noisy and Adversarial Samples

Authors Info & Claims
Published:04 August 2023Publication History

ABSTRACT

Area under the ROC curve (AUC) is an important and widely used metric in machine learning especially for imbalanced datasets. In current practical learning problems, not only adversarial samples but also noisy samples seriously threaten the performance of learning models. Nowadays, there have been a lot of research works proposed to defend the adversarial samples and noisy samples separately. Unfortunately, to the best of our knowledge, none of them with AUC optimization can secure against the two kinds of harmful samples simultaneously. To fill this gap and also address the challenge, in this paper, we propose a novel doubly robust dAUC optimization (DRAUC) algorithm. Specifically, we first exploit the deep integration of self-paced learning and adversarial training under the framework of AUC optimization, and provide a statistical upper bound to the AUC adversarial risk. Inspired by the statistical upper bound, we propose our optimization objective followed by an efficient alternatively stochastic descent algorithm, which can effectively improve the performance of learning models by guarding against adversarial samples and noisy samples. Experimental results on several standard datasets demonstrate that our DRAUC algorithm has better noise robustness and adversarial robustness than the state-of-the-art algorithms.

Skip Supplemental Material Section

Supplemental Material

rtfp1185-2min-promo.mp4

mp4

36.4 MB

References

  1. Talha Burak Alakus and Ibrahim Turkoglu. 2020. Comparison of deep learning approaches to predict COVID-19 infection. Chaos, Solitons & Fractals , Vol. 140 (2020), 110120.Google ScholarGoogle ScholarCross RefCross Ref
  2. Saeid Asgari Taghanaki, Kumar Abhishek, Joseph Paul Cohen, Julien Cohen-Adad, and Ghassan Hamarneh. 2021. Deep semantic segmentation of natural and medical images: a review. Artificial Intelligence Review , Vol. 54, 1 (2021), 137--178.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Wieland Brendel, Jonas Rauber, Alexey Kurakin, Nicolas Papernot, Behar Veliqi, Sharada P Mohanty, Florian Laurent, Marcel Salathé, Matthias Bethge, Yaodong Yu, et al. 2020. Adversarial vision challenge. In The NeurIPS'18 Competition. Springer, 129--153.Google ScholarGoogle Scholar
  4. Lei Cai, Jingyang Gao, and Di Zhao. 2020. A review of the application of deep learning in medical image classification and segmentation. Annals of translational medicine , Vol. 8, 11 (2020).Google ScholarGoogle ScholarCross RefCross Ref
  5. Qi-Zhi Cai, Chang Liu, and Dawn Song. 2018. Curriculum adversarial training. In Proceedings of the 27th International Joint Conference on Artificial Intelligence. 3740--3747.Google ScholarGoogle ScholarCross RefCross Ref
  6. Yair Carmon, Aditi Raghunathan, Ludwig Schmidt, John C Duchi, and Percy S Liang. 2019. Unlabeled data improves adversarial robustness. Advances in Neural Information Processing Systems , Vol. 32 (2019).Google ScholarGoogle Scholar
  7. Gilad Cohen, Guillermo Sapiro, and Raja Giryes. 2020. Detecting adversarial samples using influence functions and nearest neighbors. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 14453--14462.Google ScholarGoogle ScholarCross RefCross Ref
  8. Francesco Croce and Matthias Hein. 2020. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International conference on machine learning. PMLR, 2206--2216.Google ScholarGoogle Scholar
  9. Zhiyuan Dang, Xiang Li, Bin Gu, Cheng Deng, and Heng Huang. 2020. Large-scale nonlinear auc maximization via triply stochastic gradients. IEEE Transactions on Pattern Analysis and Machine Intelligence (2020).Google ScholarGoogle ScholarCross RefCross Ref
  10. Beno^it Frénay and Michel Verleysen. 2013. Classification in the presence of label noise: a survey. IEEE transactions on neural networks and learning systems, Vol. 25, 5 (2013), 845--869.Google ScholarGoogle Scholar
  11. Ellen H Fukuda and LM Grana Drummond. 2011. On the convergence of the projected gradient method for vector optimization. Optimization, Vol. 60, 8--9 (2011), 1009--1021.Google ScholarGoogle ScholarCross RefCross Ref
  12. Aritra Ghosh, Himanshu Kumar, and PS Sastry. 2017. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 31.Google ScholarGoogle ScholarCross RefCross Ref
  13. Tieliang Gong, Qian Zhao, Deyu Meng, and Zongben Xu. 2016. Why curriculum learning & self-paced learning work in big/noisy data: A theoretical perspective. Big Data & Information Analytics , Vol. 1, 1 (2016), 111.Google ScholarGoogle ScholarCross RefCross Ref
  14. Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 (2014).Google ScholarGoogle Scholar
  15. Bin Gu, Zhouyuan Huo, and Heng Huang. 2019. Scalable and Efficient Pairwise Learning to Achieve Statistical Accuracy. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 3697--3704.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Bin Gu, Zhou Zhai, Xiang Li, and Heng Huang. 2021. Finding age path of self-paced learning. In 2021 IEEE International Conference on Data Mining (ICDM). IEEE, 151--160.Google ScholarGoogle ScholarCross RefCross Ref
  17. Bin Gu, Chenkang Zhang, Huan Xiong, and Heng Huang. 2022. Balanced Self-Paced Learning for AUC Maximization. (2022).Google ScholarGoogle Scholar
  18. Chuan Guo, Mayank Rana, Moustapha Cisse, and Laurens Van Der Maaten. 2017. Countering adversarial images using input transformations. arXiv preprint arXiv:1711.00117 (2017).Google ScholarGoogle Scholar
  19. Karimollah Hajian-Tilaki. 2013. Receiver operating characteristic (ROC) curve analysis for medical diagnostic test evaluation. Caspian journal of internal medicine , Vol. 4, 2 (2013), 627.Google ScholarGoogle Scholar
  20. John Hancock and Taghi M Khoshgoftaar. 2020. Medicare fraud detection using catboost. In 2020 IEEE 21st international conference on information reuse and integration for data science (IRI). IEEE, 97--103.Google ScholarGoogle ScholarCross RefCross Ref
  21. Rongyao Hu, Xiaofeng Zhu, Yonghua Zhu, and Jiangzhang Gan. 2020. Robust SVM with adaptive graph learning. World Wide Web , Vol. 23 (2020), 1945--1968.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Mengda Huang, Yang Liu, Xiang Ao, Kuan Li, Jianfeng Chi, Jinghua Feng, Hao Yang, and Qing He. 2022. AUC-oriented Graph Neural Network for Fraud Detection. In Proceedings of the ACM Web Conference 2022. 1311--1321.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Alfredo N Iusem. 2003. On the convergence properties of the projected gradient method for convex optimization. Computational & Applied Mathematics , Vol. 22, 1 (2003), 37--52.Google ScholarGoogle ScholarCross RefCross Ref
  24. Tomoharu Iwata, Akinori Fujino, and Naonori Ueda. 2020. Semi-supervised learning for maximizing the partial AUC. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 4239--4246.Google ScholarGoogle ScholarCross RefCross Ref
  25. Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2018. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In International Conference on Machine Learning. PMLR, 2304--2313.Google ScholarGoogle Scholar
  26. Jukka-Pekka Kauppi, Melih Kandemir, Veli-Matti Saarinen, Lotta Hirvenkari, Lauri Parkkonen, Arto Klami, Riitta Hari, and Samuel Kaski. 2015. Towards brain-activity-controlled information retrieval: Decoding image relevance from MEG signals. NeuroImage , Vol. 112 (2015), 288--298.Google ScholarGoogle ScholarCross RefCross Ref
  27. Pascal Klink, Hany Abdulsamad, Boris Belousov, and Jan Peters. 2020. Self-paced contextual reinforcement learning. In Conference on Robot Learning. PMLR, 513--529.Google ScholarGoogle Scholar
  28. Alex Krizhevsky, Geoffrey Hinton, et al. 2009. Learning multiple layers of features from tiny images. (2009).Google ScholarGoogle Scholar
  29. M Kumar, Benjamin Packer, and Daphne Koller. 2010. Self-paced learning for latent variable models. Advances in neural information processing systems , Vol. 23 (2010).Google ScholarGoogle Scholar
  30. Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-based learning applied to document recognition. Proc. IEEE, Vol. 86, 11 (1998), 2278--2324.Google ScholarGoogle ScholarCross RefCross Ref
  31. Chao Li, Shangqian Gao, Cheng Deng, De Xie, and Wei Liu. 2019. Cross-modal learning with adversarial samples. Advances in Neural Information Processing Systems , Vol. 32 (2019).Google ScholarGoogle Scholar
  32. Fangzhou Liao, Ming Liang, Yinpeng Dong, Tianyu Pang, Xiaolin Hu, and Jun Zhu. 2018. Defense against adversarial attacks using high-level representation guided denoiser. In Proceedings of the IEEE conference on computer vision and pattern recognition. 1778--1787.Google ScholarGoogle ScholarCross RefCross Ref
  33. Guanxiong Liu, Issa Khalil, and Abdallah Khreishah. 2021. Using single-step adversarial training to defend iterative adversarial examples. In Proceedings of the Eleventh ACM Conference on Data and Application Security and Privacy. 17--27.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Shiqi Liu, Zilu Ma, and Deyu Meng. 2018. Understanding self-paced learning under co ve conjugacy theory. arXiv preprint arXiv:1805.08096 (2018).Google ScholarGoogle Scholar
  35. Fan Ma, Deyu Meng, Qi Xie, Zina Li, and Xuanyi Dong. 2017. Self-paced co-training. In International Conference on Machine Learning. PMLR, 2275--2284.Google ScholarGoogle Scholar
  36. Xingjun Ma, Bo Li, Yisen Wang, Sarah M Erfani, Sudanthi Wijewickrema, Grant Schoenebeck, Dawn Song, Michael E Houle, and James Bailey. 2018. Characterizing adversarial subspaces using local intrinsic dimensionality. In 6th International Conference on Learning Representations, ICLR 2019.Google ScholarGoogle Scholar
  37. Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2018. Towards Deep Learning Models Resistant to Adversarial Attacks. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  38. Patrick E McKnight and Julius Najab. 2010. Mann-Whitney U Test. The Corsini encyclopedia of psychology (2010), 1--1.Google ScholarGoogle Scholar
  39. Sarang Narkhede. 2018. Understanding auc-roc curve. Towards Data Science, Vol. 26, 1 (2018), 220--227.Google ScholarGoogle Scholar
  40. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. 2011. Reading digits in natural images with unsupervised feature learning. (2011).Google ScholarGoogle Scholar
  41. Tianyu Pang, Kun Xu, Chao Du, Ning Chen, and Jun Zhu. 2019. Improving adversarial robustness via promoting ensemble diversity. In International Conference on Machine Learning. PMLR, 4970--4979.Google ScholarGoogle Scholar
  42. Tianyu Pang, Xiao Yang, Yinpeng Dong, Hang Su, and Jun Zhu. 2020. Bag of Tricks for Adversarial Training. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  43. Nicolas Papernot, Patrick McDaniel, Xi Wu, Somesh Jha, and Ananthram Swami. 2016. Distillation as a defense to adversarial perturbations against deep neural networks. In 2016 IEEE symposium on security and privacy (SP). IEEE, 582--597.Google ScholarGoogle ScholarCross RefCross Ref
  44. Leslie Rice, Eric Wong, and Zico Kolter. 2020. Overfitting in adversarially robust deep learning. In International Conference on Machine Learning. PMLR, 8093--8104.Google ScholarGoogle Scholar
  45. Ali Shafahi, Mahyar Najibi, Mohammad Amin Ghiasi, Zheng Xu, John Dickerson, Christoph Studer, Larry S Davis, Gavin Taylor, and Tom Goldstein. 2019. Adversarial training for free! Advances in Neural Information Processing Systems , Vol. 32 (2019).Google ScholarGoogle Scholar
  46. Florian Tramèr, Alexey Kurakin, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. 2017. Ensemble adversarial training: Attacks and defenses. arXiv preprint arXiv:1705.07204 (2017).Google ScholarGoogle Scholar
  47. Arash Vahdat. 2017. Toward robustness against label noise in training deep discriminative neural networks. Advances in Neural Information Processing Systems , Vol. 30 (2017).Google ScholarGoogle Scholar
  48. Ben Van Calster, Vanya Van Belle, George Condous, Tom Bourne, Dirk Timmerman, and Sabine Van Huffel. 2008. Multi-class AUC metrics and weighted alternatives. In 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence). IEEE, 1390--1396.Google ScholarGoogle ScholarCross RefCross Ref
  49. Yu Wan, Baosong Yang, Derek F Wong, Yikai Zhou, Lidia S Chao, Haibo Zhang, and Boxing Chen. 2020. Self-Paced Learning for Neural Machine Translation. arXiv preprint arXiv:2010.04505 (2020).Google ScholarGoogle Scholar
  50. Huaxia Wang and Chun-Nam Yu. 2018. A Direct Approach to Robust Deep Learning Using Adversarial Networks. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  51. Shijun Wang, Diana Li, Nicholas Petrick, Berkman Sahiner, Marius George Linguraru, and Ronald M Summers. 2015. Optimizing area under the ROC curve using semi-supervised learning. Pattern recognition, Vol. 48, 1 (2015), 276--287.Google ScholarGoogle Scholar
  52. Xinshao Wang, Yang Hua, Elyor Kodirov, and Neil M Robertson. 2019a. IMAE for Noise-Robust Learning: Mean Absolute Error Does Not Treat Examples Equally and Gradient Magnitude's Variance Matters. arXiv preprint arXiv:1903.12141 (2019).Google ScholarGoogle Scholar
  53. Yisen Wang, Xingjun Ma, James Bailey, Jinfeng Yi, Bowen Zhou, and Quanquan Gu. 2019b. On the convergence and robustness of adversarial training. In ICML 2019: Proceedings of the 36th International Conference on Machine Learning. PMLR, 11426--11438.Google ScholarGoogle Scholar
  54. Yisen Wang, Xingjun Ma, James Bailey, Jinfeng Yi, Bowen Zhou, and Quanquan Gu. 2021. On the convergence and robustness of adversarial training. arXiv preprint arXiv:2112.08304 (2021).Google ScholarGoogle Scholar
  55. Yisen Wang, Xingjun Ma, Zaiyi Chen, Yuan Luo, Jinfeng Yi, and James Bailey. 2019c. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 322--330.Google ScholarGoogle ScholarCross RefCross Ref
  56. Zhengyang Wang, Meng Liu, Youzhi Luo, Zhao Xu, Yaochen Xie, Limei Wang, Lei Cai, Qi Qi, Zhuoning Yuan, Tianbao Yang, et al. 2022. Advanced graph and sequence neural networks for molecular property prediction and drug discovery. Bioinformatics, Vol. 38, 9 (2022), 2579--2586.Google ScholarGoogle ScholarCross RefCross Ref
  57. Eric Wong, Leslie Rice, and J Zico Kolter. 2020. Fast is better than free: Revisiting adversarial training. arXiv preprint arXiv:2001.03994 (2020).Google ScholarGoogle Scholar
  58. Huimin Wu, Zhengmian Hu, and Bin Gu. 2021. Fast and scalable adversarial training of kernel SVM via doubly stochastic gradients. In Proceedings of the AAAI conference on artificial intelligence, Vol. 35. 10329--10337.Google ScholarGoogle ScholarCross RefCross Ref
  59. Huimin Wu, William Vazelhes, and Bin Gu. 2022. Efficient Semi-Supervised Adversarial Training without Guessing Labels. In 2022 IEEE International Conference on Data Mining (ICDM). IEEE, 538--547.Google ScholarGoogle ScholarCross RefCross Ref
  60. Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. 2015. Learning from massive noisy labeled data for image classification. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2691--2699.Google ScholarGoogle Scholar
  61. Cihang Xie, Mingxing Tan, Boqing Gong, Alan Yuille, and Quoc V Le. 2020. Smooth adversarial training. arXiv preprint arXiv:2006.14536 (2020).Google ScholarGoogle Scholar
  62. Ziran Xiong, Wanli Shi, and Bin Gu. 2022. End-to-End Semi-Supervised Ordinal Regression AUC Maximization with Convolutional Kernel Networks. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2140--2150.Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. Weilin Xu, David Evans, and Yanjun Qi. 2017. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155 (2017).Google ScholarGoogle Scholar
  64. Yilun Xu, Peng Cao, Yuqing Kong, and Yizhou Wang. 2019. L_dmi: A novel information-theoretic loss function for training deep nets robust to label noise. Advances in neural information processing systems , Vol. 32 (2019).Google ScholarGoogle Scholar
  65. Yiming Ying, Longyin Wen, and Siwei Lyu. 2016. Stochastic online auc maximization. Advances in neural information processing systems , Vol. 29 (2016).Google ScholarGoogle Scholar
  66. Zhuoning Yuan, Yan Yan, Milan Sonka, and Tianbao Yang. 2021. Large-scale robust deep auc maximization: A new surrogate loss and empirical studies on medical image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 3040--3049.Google ScholarGoogle ScholarCross RefCross Ref
  67. Runtian Zhai, Tianle Cai, Di He, Chen Dan, Kun He, John Hopcroft, and Liwei Wang. 2019. Adversarially robust generalization just requires more unlabeled data. arXiv preprint arXiv:1906.00555 (2019).Google ScholarGoogle Scholar
  68. Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. 2021. Understanding deep learning (still) requires rethinking generalization. Commun. ACM, Vol. 64, 3 (2021), 107--115.Google ScholarGoogle ScholarDigital LibraryDigital Library
  69. Dinghuai Zhang, Tianyuan Zhang, Yiping Lu, Zhanxing Zhu, and Bin Dong. 2019b. You only propagate once: Accelerating adversarial training via maximal principle. Advances in Neural Information Processing Systems , Vol. 32 (2019).Google ScholarGoogle Scholar
  70. Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric Xing, Laurent El Ghaoui, and Michael Jordan. 2019a. Theoretically principled trade-off between robustness and accuracy. In International Conference on Machine Learning. PMLR, 7472--7482.Google ScholarGoogle Scholar
  71. Jingfeng Zhang, Xilie Xu, Bo Han, Gang Niu, Lizhen Cui, Masashi Sugiyama, and Mohan Kankanhalli. 2020a. Attacks which do not kill training make adversarial learning stronger. In International Conference on Machine Learning. PMLR, 11278--11287.Google ScholarGoogle Scholar
  72. Jingfeng Zhang, Xilie Xu, Bo Han, Gang Niu, Lizhen Cui, Masashi Sugiyama, and Mohan Kankanhalli. 2020b. Attacks which do not kill training make adversarial learning stronger. In International conference on machine learning. PMLR, 11278--11287.Google ScholarGoogle Scholar
  73. Zhilu Zhang and Mert Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems , Vol. 31 (2018).Google ScholarGoogle Scholar
  74. Zhengxia Zou, Zhenwei Shi, Yuhong Guo, and Jieping Ye. 2019. Object detection in 20 years: A survey. arXiv preprint arXiv:1905.05055 (2019).Google ScholarGoogle Scholar
  75. Keneilwe Zuva and Tranos Zuva. 2012. Evaluation of information retrieval systems. AIRCC's International Journal of Computer Science and Information Technology, Vol. 4, 3 (2012), 35--43. ioGoogle ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Doubly Robust AUC Optimization against Noisy and Adversarial Samples

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          KDD '23: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
          August 2023
          5996 pages
          ISBN:9798400701030
          DOI:10.1145/3580305

          Copyright © 2023 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 4 August 2023

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate1,133of8,635submissions,13%

          Upcoming Conference

        • Article Metrics

          • Downloads (Last 12 months)196
          • Downloads (Last 6 weeks)19

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader