UnlearnCanvas: A Stylized Image Dataset to Benchmark Machine Unlearning for Diffusion Models |
arXiv |
2024 |
[Code] |
Model-Agnostic |
On the Trade-Off between Actionable Explanations and the Right to be Forgotten |
arXiv |
2024 |
- |
Model-Agnostic |
Post-Training Attribute Unlearning in Recommender Systems |
arXiv |
2024 |
- |
Model-Agnostic |
Making Users Indistinguishable: Attribute-wise Unlearning in Recommender Systems |
arXiv |
2024 |
- |
Model-Agnostic |
CovarNav: Machine Unlearning via Model Inversion and Covariance Navigation |
arXiv |
2024 |
- |
Model-Agnostic |
Partially Blinded Unlearning: Class Unlearning for Deep Networks a Bayesian Perspective |
arXiv |
2024 |
- |
Model-Agnostic |
Unlearning Backdoor Threats: Enhancing Backdoor Defense in Multimodal Contrastive Learning via Local Token Unlearning |
arXiv |
2024 |
- |
Model-Agnostic |
∇τ: Gradient-based and Task-Agnostic machine Unlearning |
arXiv |
2024 |
- |
Model-Agnostic |
Towards Independence Criterion in Machine Unlearning of Features and Labels |
arXiv |
2024 |
- |
Model-Agnostic |
Challenging Forgets: Unveiling the Worst-Case Forget Sets in Machine Unlearning |
arXiv |
2024 |
[Code] |
Model-Agnostic |
Corrective Machine Unlearning |
arXiv |
2024 |
- |
Model-Agnostic |
Fair Machine Unlearning: Data Removal while Mitigating Disparities |
- |
2024 |
[Code] |
Model-Agnostic |
Label-Agnostic Forgetting: A Supervision-Free Unlearning in Deep Models |
arXiv |
2024 |
[Code] |
Model-Agnostic |
CaMU: Disentangling Causal Effects in Deep Model Unlearning |
arXiv |
2024 |
[Code] |
Model-Agnostic |
SalUn: Empowering Machine Unlearning via Gradient-based Weight Saliency in Both Image Classification and Generation |
ICLR |
2024 |
[Code] |
Model-Agnostic |
Fast Machine Unlearning Without Retraining Through Selective Synaptic Dampening |
AAAI |
2024 |
[Code] |
Model-Agnostic |
Learning to Unlearn: Instance-wise Unlearning for Pre-trained Classifiers |
AAAI |
2024 |
- |
Model-Agnostic |
Parameter-tuning-free data entry error unlearning with adaptive selective synaptic dampening |
arXiv |
2024 |
[Code] |
Model-Agnostic |
Zero-Shot Machine Unlearning at Scale via Lipschitz Regularization |
arXiv |
2024 |
[Code] |
Model-Agnostic |
Layer Attack Unlearning: Fast and Accurate Machine Unlearning via Layer Level Attack and Knowledge Distillation |
arXiv |
2023 |
- |
Model-Agnostic |
Towards bridging the gaps between the right to explanation and the right to be forgotten |
- |
2023 |
- |
Model-Agnostic |
Unlearn What You Want to Forget: Efficient Unlearning for LLMs |
EMNLP |
2023 |
[Code] |
Model-Agnostic |
Fast Model DeBias with Machine Unlearning |
NIPS |
2023 |
[Code] |
Model-Agnostic |
DUCK: Distance-based Unlearning via Centroid Kinematics |
arXiv |
2023 |
[Code] |
Model-Agnostic |
Open Knowledge Base Canonicalization with Multi-task Unlearning |
arXiv |
2023 |
- |
Model-Agnostic |
Unlearning via Sparse Representations |
arXiv |
2023 |
- |
Model-Agnostic |
SecureCut: Federated Gradient Boosting Decision Trees with Efficient Machine Unlearning |
arXiv |
2023 |
- |
Model-Agnostic |
Hidden Poison: Machine Unlearning Enables Camouflaged Poisoning Attacks |
NeurIPS |
2023 |
- |
Model-Agnostic |
Model Sparsity Can Simplify Machine Unlearning |
NeurIPS |
2023 |
[Code] |
Model-Agnostic |
Fast Model Debias with Machine Unlearning |
arXiv |
2023 |
- |
Model-Agnostic |
FRAMU: Attention-based Machine Unlearning using Federated Reinforcement Learning |
arXiv |
2023 |
- |
Model-Agnostic |
Tight Bounds for Machine Unlearning via Differential Privacy |
arXiv |
2023 |
- |
Model-Agnostic |
Machine Unlearning Methodology base on Stochastic Teacher Network |
arXiv |
2023 |
- |
Model-Agnostic |
Fast Machine Unlearning Without Retraining Through Selective Synaptic Dampening |
arXiv |
2023 |
[Code] |
Model-Agnostic |
From Adaptive Query Release to Machine Unlearning |
arXiv |
2023 |
- |
Model-Agnostic |
Towards Adversarial Evaluations for Inexact Machine Unlearning |
arXiv |
2023 |
[Code] |
Model-Agnostic |
KGA: A General Machine Unlearning Framework Based on Knowledge Gap Alignment |
arXiv |
2023 |
[Code] |
Model-Agnostic |
On the Trade-Off between Actionable Explanations and the Right to be Forgotten |
arXiv |
2023 |
- |
Model-Agnostic |
Towards Unbounded Machine Unlearning |
arXiv |
2023 |
[Code] |
Model-Agnostic |
Netflix and Forget: Efficient and Exact Machine Unlearning from Bi-linear Recommendations |
arXiv |
2023 |
- |
Model-Agnostic |
To Be Forgotten or To Be Fair: Unveiling Fairness Implications of Machine Unlearning Methods |
arXiv |
2023 |
[Code] |
Model-Agnostic |
Sequential Informed Federated Unlearning: Efficient and Provable Client Unlearning in Federated Optimization |
arXiv |
2022 |
- |
Model-Agnostic |
Certified Data Removal in Sum-Product Networks |
ICKG |
2022 |
[Code] |
Model-Agnostic |
Learning with Recoverable Forgetting |
ECCV |
2022 |
- |
Model-Agnostic |
Continual Learning and Private Unlearning |
CoLLAs |
2022 |
[Code] |
Model-Agnostic |
Verifiable and Provably Secure Machine Unlearning |
arXiv |
2022 |
[Code] |
Model-Agnostic |
VeriFi: Towards Verifiable Federated Unlearning |
arXiv |
2022 |
- |
Model-Agnostic |
FedRecover: Recovering from Poisoning Attacks in Federated Learning using Historical Information |
S&P |
2022 |
- |
Model-Agnostic |
Fast Yet Effective Machine Unlearning |
arXiv |
2022 |
- |
Model-Agnostic |
Membership Inference via Backdooring |
IJCAI |
2022 |
[Code] |
Model-Agnostic |
Forget Unlearning: Towards True Data-Deletion in Machine Learning |
ICLR |
2022 |
- |
Model-Agnostic |
Zero-Shot Machine Unlearning |
arXiv |
2022 |
- |
Model-Agnostic |
Efficient Attribute Unlearning: Towards Selective Removal of Input Attributes from Feature Representations |
arXiv |
2022 |
- |
Model-Agnostic |
Few-Shot Unlearning |
ICLR |
2022 |
- |
Model-Agnostic |
Federated Unlearning: How to Efficiently Erase a Client in FL? |
UpML Workshop |
2022 |
- |
Model-Agnostic |
Machine Unlearning Method Based On Projection Residual |
DSAA |
2022 |
- |
Model-Agnostic |
Hard to Forget: Poisoning Attacks on Certified Machine Unlearning |
AAAI |
2022 |
[Code] |
Model-Agnostic |
Athena: Probabilistic Verification of Machine Unlearning |
PoPETs |
2022 |
- |
Model-Agnostic |
FP2-MIA: A Membership Inference Attack Free of Posterior Probability in Machine Unlearning |
ProvSec |
2022 |
- |
Model-Agnostic |
Deletion Inference, Reconstruction, and Compliance in Machine (Un)Learning |
PETS |
2022 |
- |
Model-Agnostic |
Prompt Certified Machine Unlearning with Randomized Gradient Smoothing and Quantization |
NeurIPS |
2022 |
- |
Model-Agnostic |
The Right to be Forgotten in Federated Learning: An Efficient Realization with Rapid Retraining |
INFOCOM |
2022 |
[Code] |
Model-Agnostic |
Backdoor Defense with Machine Unlearning |
INFOCOM |
2022 |
- |
Model-Agnostic |
Markov Chain Monte Carlo-Based Machine Unlearning: Unlearning What Needs to be Forgotten |
ASIA CCS |
2022 |
- |
Model-Agnostic |
Federated Unlearning for On-Device Recommendation |
arXiv |
2022 |
- |
Model-Agnostic |
Can Bad Teaching Induce Forgetting? Unlearning in Deep Networks using an Incompetent Teacher |
arXiv |
2022 |
- |
Model-Agnostic |
Efficient Two-Stage Model Retraining for Machine Unlearning |
CVPR Workshop |
2022 |
- |
Model-Agnostic |
Learn to Forget: Machine Unlearning Via Neuron Masking |
IEEE |
2021 |
- |
Model-Agnostic |
Adaptive Machine Unlearning |
NeurIPS |
2021 |
[Code] |
Model-Agnostic |
Descent-to-Delete: Gradient-Based Methods for Machine Unlearning |
ALT |
2021 |
- |
Model-Agnostic |
Remember What You Want to Forget: Algorithms for Machine Unlearning |
NeurIPS |
2021 |
- |
Model-Agnostic |
FedEraser: Enabling Efficient Client-Level Data Removal from Federated Learning Models |
IWQoS |
2021 |
- |
Model-Agnostic |
Federated Unlearning |
IWQoS |
2021 |
[Code] |
Model-Agnostic |
Machine Unlearning via Algorithmic Stability |
COLT |
2021 |
- |
Model-Agnostic |
EMA: Auditing Data Removal from Trained Models |
MICCAI |
2021 |
[Code] |
Model-Agnostic |
Knowledge-Adaptation Priors |
NeurIPS |
2021 |
[Code] |
Model-Agnostic |
PrIU: A Provenance-Based Approach for Incrementally Updating Regression Models |
NeurIPS |
2020 |
- |
Model-Agnostic |
Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks |
CVPR |
2020 |
- |
Model-Agnostic |
Learn to Forget: User-Level Memorization Elimination in Federated Learning |
arXiv |
2020 |
- |
Model-Agnostic |
Certified Data Removal from Machine Learning Models |
ICML |
2020 |
- |
Model-Agnostic |
Class Clown: Data Redaction in Machine Unlearning at Enterprise Scale |
arXiv |
2020 |
- |
Model-Agnostic |
A Novel Online Incremental and Decremental Learning Algorithm Based on Variable Support Vector Machine |
Cluster Computing |
2019 |
- |
Model-Agnostic |
Making AI Forget You: Data Deletion in Machine Learning |
NeurIPS |
2019 |
- |
Model-Agnostic |
Lifelong Anomaly Detection Through Unlearning |
CCS |
2019 |
- |
Model-Agnostic |
Learning Not to Learn: Training Deep Neural Networks With Biased Data |
CVPR |
2019 |
- |
Model-Agnostic |
Efficient Repair of Polluted Machine Learning Systems via Causal Unlearning |
ASIACCS |
2018 |
[Code] |
Model-Agnostic |
Understanding Black-box Predictions via Influence Functions |
ICML |
2017 |
[Code] |
Model-Agnostic |
Towards Making Systems Forget with Machine Unlearning |
S&P |
2015 |
- |
Model-Agnostic |
Towards Making Systems Forget with Machine Unlearning |
S&P |
2015 |
- |
Model-Agnostic |
Incremental and decremental training for linear classification |
KDD |
2014 |
[Code] |
Model-Agnostic |
Multiple Incremental Decremental Learning of Support Vector Machines |
NIPS |
2009 |
- |
Model-Agnostic |
Incremental and Decremental Learning for Linear Support Vector Machines |
ICANN |
2007 |
- |
Model-Agnostic |
Decremental Learning Algorithms for Nonlinear Langrangian and Least Squares Support Vector Machines |
OSB |
2007 |
- |
Model-Agnostic |
Multicategory Incremental Proximal Support Vector Classifiers |
KES |
2003 |
- |
Model-Agnostic |
Incremental and Decremental Proximal Support Vector Classification using Decay Coefficients |
DaWak |
2003 |
- |
Model-Agnostic |
Incremental and Decremental Support Vector Machine Learning |
NeurIPS |
2000 |
- |
Model-Agnostic |
Machine Unlearning for Image-to-Image Generative Models |
arXiv |
2024 |
[Code] |
Model-Intrinsic |
Towards Efficient and Effective Unlearning of Large Language Models for Recommendation |
arXiv |
2024 |
[Code] |
Model-Intrinsic |
Dissecting Language Models: Machine Unlearning via Selective Pruning |
arXiv |
2024 |
- |
Model-Intrinsic |
Decentralized Federated Unlearning on Blockchain |
arXiv |
2024 |
- |
Model-Intrinsic |
Unlink to Unlearn: Simplifying Edge Unlearning in GNNs |
WWW |
2024 |
[Code] |
Model-Intrinsic |
Preserving Privacy Through Dememorization: An Unlearning Technique For Mitigating Memorization Risks In Language Models |
EMNLP |
2024 |
- |
Model-Intrinsic |
Towards Effective and General Graph Unlearning via Mutual Evolution |
AAAI |
2024 |
- |
Model-Intrinsic |
Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Operation |
AAAI |
2024 |
[Code] |
Model-Intrinsic |
FAST: Feature Aware Similarity Thresholding for Weak Unlearning in Black-Box Generative Models |
arXiv |
2023 |
[Code] |
Model-Intrinsic |
Fast-NTK: Parameter-Efficient Unlearning for Large-Scale Models |
arXiv |
2023 |
- |
Model-Intrinsic |
Certified Minimax Unlearning with Generalization Rates and Deletion Capacity |
NeurIPS |
2023 |
- |
Model-Intrinsic |
FairSISA: Ensemble Post-Processing to Improve Fairness of Unlearning in LLMs |
NeurIPS |
2023 |
- |
Model-Intrinsic |
Multimodal Machine Unlearning |
arXiv |
2023 |
- |
Model-Intrinsic |
Adapt then Unlearn: Exploiting Parameter Space Semantics for Unlearning in Generative Adversarial Networks |
arXiv |
2023 |
- |
Model-Intrinsic |
SAFE: Machine Unlearning With Shard Graphs |
ICCV |
2023 |
- |
Model-Intrinsic |
MUter: Machine Unlearning on Adversarially Trained Models |
ICCV |
2023 |
- |
Model-Intrinsic |
Heterogeneous Federated Knowledge Graph Embedding Learning and Unlearning |
WWW |
2023 |
[Code] |
Model-Intrinsic |
One-Shot Machine Unlearning with Mnemonic Code |
arXiv |
2023 |
- |
Model-Intrinsic |
Inductive Graph Unlearning |
USENIX |
2023 |
[Code] |
Model-Intrinsic |
ERM-KTP: Knowledge-level Machine Unlearning via Knowledge Transfer |
CVPR |
2023 |
[Code] |
Model-Intrinsic |
GNNDelete: A General Strategy for Unlearning in Graph Neural Networks |
ICLR |
2023 |
[Code] |
Model-Intrinsic |
Unfolded Self-Reconstruction LSH: Towards Machine Unlearning in Approximate Nearest Neighbour Search |
arXiv |
2023 |
[Code] |
Model-Intrinsic |
Efficiently Forgetting What You Have Learned in Graph Representation Learning via Projection |
AISTATS |
2023 |
[Code] |
Model-Intrinsic |
Unrolling SGD: Understanding Factors Influencing Machine Unlearning |
EuroS&P |
2022 |
[Code] |
Model-Intrinsic |
Graph Unlearning |
CCS |
2022 |
[Code] |
Model-Intrinsic |
Certified Graph Unlearning |
GLFrontiers Workshop |
2022 |
[Code] |
Model-Intrinsic |
Skin Deep Unlearning: Artefact and Instrument Debiasing in the Context of Melanoma Classification |
ICML |
2022 |
[Code] |
Model-Intrinsic |
Near-Optimal Task Selection for Meta-Learning with Mutual Information and Online Variational Bayesian Unlearning |
AISTATS |
2022 |
- |
Model-Intrinsic |
Unlearning Protected User Attributes in Recommendations with Adversarial Training |
SIGIR |
2022 |
[Code] |
Model-Intrinsic |
Recommendation Unlearning |
TheWebConf |
2022 |
[Code] |
Model-Intrinsic |
Knowledge Neurons in Pretrained Transformers |
ACL |
2022 |
[Code] |
Model-Intrinsic |
Memory-Based Model Editing at Scale |
MLR |
2022 |
[Code] |
Model-Intrinsic |
Forgetting Fast in Recommender Systems |
arXiv |
2022 |
- |
Model-Intrinsic |
Unlearning Nonlinear Graph Classifiers in the Limited Training Data Regime |
arXiv |
2022 |
- |
Model-Intrinsic |
Deep Regression Unlearning |
arXiv |
2022 |
- |
Model-Intrinsic |
Quark: Controllable Text Generation with Reinforced Unlearning |
arXiv |
2022 |
[Code] |
Model-Intrinsic |
Forget-SVGD: Particle-Based Bayesian Federated Unlearning |
DSL Workshop |
2022 |
- |
Model-Intrinsic |
Machine Unlearning of Federated Clusters |
arXiv |
2022 |
- |
Model-Intrinsic |
Machine Unlearning for Image Retrieval: A Generative Scrubbing Approach |
MM |
2022 |
- |
Model-Intrinsic |
Machine Unlearning: Linear Filtration for Logit-based Classifiers |
Machine Learning |
2022 |
- |
Model-Intrinsic |
Deep Unlearning via Randomized Conditionally Independent Hessians |
CVPR |
2022 |
[Code] |
Model-Intrinsic |
Challenges and Pitfalls of Bayesian Unlearning |
UPML Workshop |
2022 |
- |
Model-Intrinsic |
Federated Unlearning via Class-Discriminative Pruning |
WWW |
2022 |
- |
Model-Intrinsic |
Active forgetting via influence estimation for neural networks |
Int. J. Intel. Systems |
2022 |
- |
Model-Intrinsic |
Variational Bayesian unlearning |
NeurIPS |
2022 |
- |
Model-Intrinsic |
Revisiting Machine Learning Training Process for Enhanced Data Privacy |
IC3 |
2021 |
- |
Model-Intrinsic |
Knowledge Removal in Sampling-based Bayesian Inference |
ICLR |
2021 |
[Code] |
Model-Intrinsic |
Mixed-Privacy Forgetting in Deep Networks |
CVPR |
2021 |
- |
Model-Intrinsic |
HedgeCut: Maintaining Randomised Trees for Low-Latency Machine Unlearning |
SIGMOD |
2021 |
[Code] |
Model-Intrinsic |
A Unified PAC-Bayesian Framework for Machine Unlearning via Information Risk Minimization |
MLSP |
2021 |
- |
Model-Intrinsic |
DeepObliviate: A Powerful Charm for Erasing Data Residual Memory in Deep Neural Networks |
arXiv |
2021 |
- |
Model-Intrinsic |
Approximate Data Deletion from Machine Learning Models: Algorithms and Evaluations |
AISTATS |
2021 |
[Code] |
Model-Intrinsic |
Bayesian Inference Forgetting |
arXiv |
2021 |
[Code] |
Model-Intrinsic |
Approximate Data Deletion from Machine Learning Models |
AISTATS |
2021 |
[Code] |
Model-Intrinsic |
Online Forgetting Process for Linear Regression Models |
AISTATS |
2021 |
- |
Model-Intrinsic |
RevFRF: Enabling Cross-domain Random Forest Training with Revocable Federated Learning |
IEEE |
2021 |
- |
Model-Intrinsic |
Coded Machine Unlearning |
IEEE Access |
2021 |
- |
Model-Intrinsic |
Machine Unlearning for Random Forests |
ICML |
2021 |
- |
Model-Intrinsic |
Bayesian Variational Federated Learning and Unlearning in Decentralized Networks |
SPAWC |
2021 |
- |
Model-Intrinsic |
Forgetting Outside the Box: Scrubbing Deep Networks of Information Accessible from Input-Output Observations |
ECCV |
2020 |
- |
Model-Intrinsic |
Influence Functions in Deep Learning Are Fragile |
arXiv |
2020 |
- |
Model-Intrinsic |
Deep Autoencoding Topic Model With Scalable Hybrid Bayesian Inference |
IEEE |
2020 |
- |
Model-Intrinsic |
Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks |
CVPR |
2020 |
- |
Model-Intrinsic |
Uncertainty in Neural Networks: Approximately Bayesian Ensembling |
AISTATS |
2020 |
[Code] |
Model-Intrinsic |
Certified Data Removal from Machine Learning Models |
ICML |
2020 |
- |
Model-Intrinsic |
DeltaGrad: Rapid retraining of machine learning models |
ICML |
2020 |
[Code] |
Model-Intrinsic |
Making AI Forget You: Data Deletion in Machine Learning |
NeurIPS |
2019 |
- |
Model-Intrinsic |
“Amnesia” – Towards Machine Learning Models That Can Forget User Data Very Fast |
AIDB Workshop |
2019 |
[Code] |
Model-Intrinsic |
A Novel Online Incremental and Decremental Learning Algorithm Based on Variable Support Vector Machine |
Cluster Computing |
2019 |
- |
Model-Intrinsic |
Neural Text Degeneration With Unlikelihood Training |
arXiv |
2019 |
[Code] |
Model-Intrinsic |
Bayesian Neural Networks with Weight Sharing Using Dirichlet Processes |
IEEE |
2018 |
[Code] |
Model-Intrinsic |
Towards Machine Unlearning Benchmarks: Forgetting the Personal Identities in Facial Recognition Systems |
arXiv |
2023 |
[Code] |
Data-Driven |
Hidden Poison: Machine Unlearning Enables Camouflaged Poisoning Attacks |
NeurIPS-TSRML |
2022 |
[Code] |
Data-Driven |
Forget Unlearning: Towards True Data Deletion in Machine Learning |
ICLR |
2022 |
- |
Data-Driven |
ARCANE: An Efficient Architecture for Exact Machine Unlearning |
IJCAI |
2022 |
- |
Data-Driven |
PUMA: Performance Unchanged Model Augmentation for Training Data Removal |
AAAI |
2022 |
- |
Data-Driven |
Certifiable Unlearning Pipelines for Logistic Regression: An Experimental Study |
MAKE |
2022 |
[Code] |
Data-Driven |
Zero-Shot Machine Unlearning |
arXiv |
2022 |
- |
Data-Driven |
GRAPHEDITOR: An Efficient Graph Representation Learning and Unlearning Approach |
- |
2022 |
[Code] |
Data-Driven |
Fast Model Update for IoT Traffic Anomaly Detection with Machine Unlearning |
IEEE IoT-J |
2022 |
- |
Data-Driven |
Learning to Refit for Convex Learning Problems |
arXiv |
2021 |
- |
Data-Driven |
Fast Yet Effective Machine Unlearning |
arXiv |
2021 |
- |
Data-Driven |
Learning with Selective Forgetting |
IJCAI |
2021 |
- |
Data-Driven |
SSSE: Efficiently Erasing Samples from Trained Machine Learning Models |
NeurIPS-PRIML |
2021 |
- |
Data-Driven |
How Does Data Augmentation Affect Privacy in Machine Learning? |
AAAI |
2021 |
[Code] |
Data-Driven |
Coded Machine Unlearning |
IEEE |
2021 |
- |
Data-Driven |
Machine Unlearning |
IEEE |
2021 |
[Code] |
Data-Driven |
How Does Data Augmentation Affect Privacy in Machine Learning? |
AAAI |
2021 |
[Code] |
Data-Driven |
Amnesiac Machine Learning |
AAAI |
2021 |
[Code] |
Data-Driven |
Unlearnable Examples: Making Personal Data Unexploitable |
ICLR |
2021 |
[Code] |
Data-Driven |
Descent-to-Delete: Gradient-Based Methods for Machine Unlearning |
ALT |
2021 |
- |
Data-Driven |
Fawkes: Protecting Privacy against Unauthorized Deep Learning Models |
USENIX Sec. Sym. |
2020 |
[Code] |
Data-Driven |
PrIU: A Provenance-Based Approach for Incrementally Updating Regression Models |
SIGMOD |
2020 |
- |
Data-Driven |
DeltaGrad: Rapid retraining of machine learning models |
ICML |
2020 |
[Code] |
Data-Driven |