• 中国精品科技期刊
  • CCF推荐A类中文期刊
  • 计算领域高质量科技期刊T1类
Advanced Search
Yin Yuyu, Wu Guangqiang, Li Youhuizi, Wang Xinyu, Gao Honghao. A Machine Unlearning Method via Feature Constraint and Adaptive Loss Balance[J]. Journal of Computer Research and Development, 2024, 61(10): 2649-2661. DOI: 10.7544/issn1000-1239.202440476
Citation: Yin Yuyu, Wu Guangqiang, Li Youhuizi, Wang Xinyu, Gao Honghao. A Machine Unlearning Method via Feature Constraint and Adaptive Loss Balance[J]. Journal of Computer Research and Development, 2024, 61(10): 2649-2661. DOI: 10.7544/issn1000-1239.202440476

A Machine Unlearning Method via Feature Constraint and Adaptive Loss Balance

Funds: This work was supported by the National Natural Science Foundation of China (62272140) , the Natural Science Foundation of Zhejiang Province (LY22F020018), and the “Pioneer”and“Leading Goose” Research and Development Program of Zhejiang Province (2024C01166).
More Information
  • Author Bio:

    Yin Yuyu: born in 1980. PhD, professor. Member of CCF. His main research interests include service computing, edge computing, and business process management

    Wu Guangqiang: born in 1999. Master. His main research interests include machine unlearning, deep learning, and incremental learning

    Li Youhuizi: born in 1989. PhD, associate professor. Member of CCF. Her main research interests include edge computing, privacy security, and mobile edge computing

    Wang Xinyu: born in 2000. Master. His main research interests include edge computing, deep learning, and natural language processing

    Gao Honghao: born in 1985. PhD, associate professor. Member of CCF. His main research interests include software formal verification, service collaborative computing, wireless networks and industrial Internet of things, and intelligent medical image processing

  • Received Date: June 02, 2024
  • Revised Date: July 18, 2024
  • Available Online: September 13, 2024
  • With the accelerated advancement of digitization, data elements have become the core driving force for the operation of modern society. However, at the same time, data security issues have become increasingly prominent, with frequent occurrences of data breaches and privacy violations, causing serious losses to individuals, organizations, and even countries. Against this backdrop, the security of data elements has become the focus of attention from all sectors of society, and the issue of data privacy protection in deep learning models has also attracted widespread attention. Among them, machine unlearning, as a key technology for protecting user’s privacy, aims to enable models to remove the influence of specific data while maintaining generalization performance for remaining data, providing an effective solution for protecting the security of data elements in deep learning models. Existing machine unlearning methods are mainly divided into two categories: exact unlearning and approximate unlearning. However, exact unlearning methods need to intervene in the original training process of the models, while approximate unlearning methods find it difficult to strike a balance between unlearning performance and model generalization ability. To address these issues, we propose an approximate unlearning framework based on feature constraints and adaptive loss balancing. We adopt a “forgetting-recovering” machine unlearning framework. First, for the “forgetting” process, in order to mimic the feature outputs of retrained models for the forgetting samples, we use a randomly initialized model that has not been trained on the forgetting samples to guide the feature outputs of the unlearning model, constraining forgetting at the feature level to avoid easily obtaining forgetting data information from the model. Then, a small amount of data is used for fine-tuning to “recover” the generalization performance of models on the remaining data. At the same time, we regard the above machine unlearning framework as a multi-task optimization problem and introduce adaptive loss balance to automatically balance the “forgetting” and “recovering” tasks, preventing the model from “over-forgetting” or “over-recovering”, so that the “forgetting” and “recovering” tasks can be trained relatively balanced and steadily. Extensive experiments on 3 image classification datasets show that our method can effectively forget the forgetting data and achieve optimal performance in multiple metrics.

  • [1]
    Pulido-Gaytan L B, Tchernykh A, Cortés-Mendoza J M, et al. A survey on privacy-preserving machine learning with fully homomorphic encryption[C]//Proc of Latin American High Performance Computing Conf. Cham: Springer International Publishing, 2020: 115−129
    [2]
    Liu Yi, Xu Lei, Yuan Xingliang, et al. The right to be forgotten in federated learning: An efficient realization with rapid retraining[C]//Proc of IEEE Conf on Computer Communications (IEEE INFOCOM 2022). Piscataway, NJ: IEEE, 2022: 1749−1758
    [3]
    Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition[C/OL]//Proc of the Int Conf on Learning Representations. 2015 [2024-05-28]. https://arxiv.org/abs/1409.1556
    [4]
    Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets[J]. Advances in Neural Information Processing Systems, 2014, 27: 2672−2680
    [5]
    Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[C]//Advances in Neural Information Processing Systems, Cambridge, MA: MIT Press, 2017: 5998−6008
    [6]
    Fredrikson M, Jha S, Ristenpart T. Model inversion attacks that exploit confidence information and basic countermeasures[C]//Proc of the 22nd ACM SIGSAC Conf on Computer and Communications Security. New York: ACM, 2015: 1322−1333
    [7]
    Voigt P, Von dem Bussche A. The EU General Data Protection Regulation (GDPR)[EB/OL]. https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:32016R0679
    [8]
    Bourtoule L, Chandrasekaran V, Choquette-Choo C A, et al. Machine unlearning[C]//Proc of 2021 IEEE Symp on Security and Privacy (SP). Piscataway, NJ: IEEE, 2021: 141−159
    [9]
    Yan Haonan, Li Xiaoguang, Guo Ziyao, et al. ARCANE: An efficient architecture for exact machine unlearning[C]//Proc of the 31st Int Joint Conf on Artificial Intelligence. Menlo Park, CA: AAAI, 2022: 4006−4013
    [10]
    Dukler Y, Bowman B, Achille A, et al. SAFE: Machine unlearning with shard graphs[C]//Proc of the IEEE/CVF Int Conf on Computer Vision. Piscataway, NJ: IEEE, 2023: 17108−17118
    [11]
    Golatkar A, Achille A, Soatto S. Eternal sunshine of the spotless net: Selective forgetting in deep networks[C]//Proc of the IEEE/CVF Conf on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE, 2020: 9304−9312
    [12]
    Huang H, Ma X, Erfani S M, et al. Unlearnable examples: Making personal data unexploitable[C/OL]//Proc of the Int Conf on Learning Representations. 2021 [2024-01-17]. https://arxiv.org/abs/2101.04898
    [13]
    Tarun A K, Chundawat V S, Mandal M, et al. Fast yet effective machine unlearning[J]. IEEE Transactions on Neural Networks and Learning Systems, 2023. Doi: 10.1109/TNNLS.2023.3266233, 2023
    [14]
    Kim J, Woo S S. Efficient two-stage model retraining for machine unlearning[C]//Proc of the IEEE/CVF Conf on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE, 2022: 4361−4369
    [15]
    Chundawat V S, Tarun A K, Mandal M, et al. Can bad teaching induce forgetting? unlearning in deep networks using an incompetent teacher[C]//Proc of the AAAI Conf on Artificial Intelligence. Palo Alto, CA: AAAI, 2023, 37(6): 7210−7217
    [16]
    Chen Min, Gao Weizhuo, Liu Gaoyang, et al. Boundary unlearning: Rapid forgetting of deep networks via shifting the decision boundary[C]//Proc of the IEEE/CVF Conf on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE, 2023: 7766−7775
    [17]
    Vandenhende S, Georgoulis S, Van Gansbeke W, et al. Multi-task learning for dense prediction tasks: A survey[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, 44(7): 3614−3633
    [18]
    Zhao Shiji, Yu Jie, Sun Zhenlong, et al. Enhanced accuracy and robustness via multi-teacher adversarial distillation[C]//Proc of European Conf on Computer Vision. Cham: Springer, 2022: 585−602
    [19]
    Kirkpatrick J, Pascanu R, Rabinowitz N, et al. Overcoming catastrophic forgetting in neural networks[J]. Proceedings of the National Academy of Sciences, 2017, 114(13): 3521−3526 doi: 10.1073/pnas.1611835114
    [20]
    Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network[J]. Computer Science, 2015, 14(7): 38−39
    [21]
    Zhang Xulong, Wang Jianzong, Cheng Ning, et al. Machine unlearning methodology based on stochastic teacher network[C]//Proc of Int Conf on Advanced Data Mining and Applications. Cham: Springer, 2023: 250−261
    [22]
    郭虎升,张洋,王文剑. 面向不同类型概念漂移的两阶段自适应集成学习方法[J]. 计算机研究与发展,2024,61(7):1799−1811 doi: 10.7544/issn1000-1239.202330452

    Guo Husheng, Zhang Yang, Wang Wenjian. Two-stage adaptive ensemble learning method for different types of concept drift[J]. Journal of Computer Research and Development, 2024, 61(7): 1799−1811(in Chinese) doi: 10.7544/issn1000-1239.202330452
    [23]
    Foster J, Schoepf S, Brintrup A. Fast machine unlearning without retraining through selective synaptic dampening[C]//Proc of the AAAI Conf on Artificial Intelligence. Palo Alto, CA: AAAI, 2024, 38(11): 12043−12051
    [24]
    Aich A. Elastic weight consolidation (EWC): Nuts and bolts[J]. arXiv preprint, arXiv: 2105.04093, 2021
    [25]
    Cotogni M, Bonato J, Sabetta L, et al. DUCK: Distance-based unlearning via centroid kinematics[J]. arXiv preprint, arXiv: 2312.02052, 2023
    [26]
    Paszke A, Gross S, Massa F, et al. Pytorch: An imperative style, high-performance deep learning library[C]//Advances in Neural Information Processing Systems Cambridge, MA: MIT, 2019: 8024−8035
    [27]
    Krizhevsky A, Hinton G. Learning multiple layers of features from tiny images[R]. Toronto: University of Toronto, 2009
    [28]
    LeCun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998, 86(11): 2278−2324 doi: 10.1109/5.726791
    [29]
    Springenberg J T, Dosovitskiy A, Brox T, et al. Striving for simplicity: The all convolutional net[C/OL]//Proc of the Int Conf on Learning Representations. 2015 [2024-02-02]. https://arxiv.org/abs/1412.6806
    [30]
    He Kaiming, Zhang Xiangyu, Ren Shaoqing, et al. Deep residual learning for image recognition[C]//Proc of the IEEE Conf on Computer Vision and Pattern Recognition. Piscataway, NJ: IEEE, 2016: 770−778
    [31]
    Shokri R, Stronati M, Song C, et al. Membership inference attacks against machine learning models[C]//Proc of 2017 IEEE Symp on Security and Privacy (SP). Piscataway, NJ: IEEE, 2017: 3−18
    [32]
    李文斌,熊亚锟,范祉辰,等. 持续学习的研究进展与趋势[J]. 计算机研究与发展,2024,61(6):1476−1496 doi: 10.7544/issn1000-1239.202220820

    Li Wenbin, Xiong Yakun, Fan Zhichen, et al. Advances and trends of continual learning[J]. Journal of Computer Research and Development, 2024, 61(6): 1476−1496 (in Chinese) doi: 10.7544/issn1000-1239.202220820
    [33]
    Goodfellow I J, Shlens J, Szegedy C. Explaining and harnessing adversarial examples[C/OL]//Proc of the Int Conf on Learning Representations. 2015 [2024-02-02]. https://arxiv.org/abs/ 1412.6572
    [34]
    Selvaraju R R, Cogswell M, Das A, et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization[C]//Proc of the IEEE Int Conf on Computer Vision. Piscataway, NJ: IEEE, 2017: 618−626
  • Related Articles

    [1]Cao Yiran, Zhu Youwen, He Xingyu, Zhang Yue. Utility-Optimized Local Differential Privacy Set-Valued Data Frequency Estimation Mechanism[J]. Journal of Computer Research and Development, 2022, 59(10): 2261-2274. DOI: 10.7544/issn1000-1239.20220504
    [2]Hong Jinxin, Wu Yingjie, Cai Jianping, Sun Lan. Differentially Private High-Dimensional Binary Data Publication via Attribute Segmentation[J]. Journal of Computer Research and Development, 2022, 59(1): 182-196. DOI: 10.7544/issn1000-1239.20200701
    [3]Wu Wanqing, Zhao Yongxin, Wang Qiao, Di Chaofan. A Safe Storage and Release Method of Trajectory Data Satisfying Differential Privacy[J]. Journal of Computer Research and Development, 2021, 58(11): 2430-2443. DOI: 10.7544/issn1000-1239.2021.20210589
    [4]Zhang Yuxuan, Wei Jianghong, Li Ji, Liu Wenfen, Hu Xuexian. Graph Degree Histogram Publication Method with Node-Differential Privacy[J]. Journal of Computer Research and Development, 2019, 56(3): 508-520. DOI: 10.7544/issn1000-1239.2019.20170886
    [5]Zhu Weijun, You Qingguang, Yang Weidong, Zhou Qinglei. Trajectory Privacy Preserving Based on Statistical Differential Privacy[J]. Journal of Computer Research and Development, 2017, 54(12): 2825-2832. DOI: 10.7544/issn1000-1239.2017.20160647
    [6]Wu Yingjie, Zhang Liqun, Kang Jian, Wang Yilei. An Algorithm for Differential Privacy Streaming Data Adaptive Publication[J]. Journal of Computer Research and Development, 2017, 54(12): 2805-2817. DOI: 10.7544/issn1000-1239.2017.20160555
    [7]Wang Liang, Wang Weiping, Meng Dan. Privacy Preserving Data Publishing via Weighted Bayesian Networks[J]. Journal of Computer Research and Development, 2016, 53(10): 2343-2353. DOI: 10.7544/issn1000-1239.2016.20160465
    [8]Lu Guoqing, Zhang Xiaojian, Ding Liping, Li Yanfeng, Liao Xin. Frequent Sequential Pattern Mining under Differential Privacy[J]. Journal of Computer Research and Development, 2015, 52(12): 2789-2801. DOI: 10.7544/issn1000-1239.2015.20140516
    [9]Ouyang Jia, Yin Jian, Liu Shaopeng, Liu Yubao. An Effective Differential Privacy Transaction Data Publication Strategy[J]. Journal of Computer Research and Development, 2014, 51(10): 2195-2205. DOI: 10.7544/issn1000-1239.2014.20130824
    [10]Ni Weiwei, Chen Geng, Chong Zhihong, Wu Yingjie. Privacy-Preserving Data Publication for Clustering[J]. Journal of Computer Research and Development, 2012, 49(5): 1095-1104.
  • Cited by

    Periodical cited type(5)

    1. 张涵,于航,周继威,白云开,赵路坦. 面向隐私计算的可信执行环境综述. 计算机应用. 2025(02): 467-481 .
    2. 付裕,林璟锵,冯登国. 虚拟化与密码技术应用:现状与未来. 密码学报(中英文). 2024(01): 3-21 .
    3. 徐传康,李忠月,刘天宇,种统洪,杨发雪. 基于可信执行环境的汽车域控系统安全研究. 汽车实用技术. 2024(15): 18-25+73 .
    4. 徐文嘉,岑孟杰,陈亮. 隐私保护下单细胞RNA测序数据细胞分类研究. 医学信息学杂志. 2024(10): 86-89 .
    5. 孙钰,熊高剑,刘潇,李燕. 基于可信执行环境的安全推理研究进展. 信息网络安全. 2024(12): 1799-1818 .

    Other cited types(4)

Catalog

    Article views (126) PDF downloads (66) Cited by(9)

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return