可信深度学习Trustworthy Deep Learning相关论文

这些论文涵盖了人工智能领域的关键议题,包括可信AI的原则和监管,联邦学习的安全与隐私,对抗性学习中的后门攻击和权重攻击,以及模型在分布变化下的泛化能力。此外,还讨论了数据增强对自然分布偏移的鲁棒性,隐私保护的数据合成方法,以及针对深度学习模型的解释性和反向工程。同时,强调了在信号处理、机器学习和深度学习中的改进方法对于故障诊断和时间序列分析的重要性。
摘要由CSDN通过智能技术生成

Survey

An Overview of Catastrophic AI Risks. [paper]

Connecting the Dots in Trustworthy Artificial Intelligence: From AI Principles, Ethics, and Key Requirements to Responsible AI Systems and Regulation. [paper]

A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy. [paper]

Adversarial Machine Learning: A Systematic Survey of Backdoor Attack, Weight Attack and Adversarial Example. [paper]

Out-of-Distribution Generalization

Simple and Fast Group Robustness by Automatic Feature Reweighting. [paper]

Optimal Transport Model Distributional Robustness. [paper]

Explore and Exploit the Diverse Knowledge in Model Zoo for Domain Generalization. [paper]

Exact Generalization Guarantees for (Regularized) Wasserstein Distributionally Robust Models. [paper]

Rethinking the Evaluation Protocol of Domain Generalization. [paper]

Dynamic Regularized Sharpness Aware Minimization in Federated Learning: Approaching Global Consistency and Smooth Landscape. [paper]

On the nonlinear correlation of ML performance between data subpopulations. [paper]

An Adaptive Algorithm for Learning with Unknown Distribution Drift. [paper]

PGrad: Learning Principal Gradients For Domain Generalization. [paper]

Benchmarking Low-Shot Robustness to Natural Distribution Shifts. [paper]

eweighted Mixup for Subpopulation Shift. [paper]

ERM++: An Improved Baseline for Domain Generalization. [paper]

Domain Generalization via Nuclear Norm Regularization. [paper]

ManyDG: Many-domain Generalization for Healthcare Applications. [paper]

DEJA VU: Continual Model Generalization For Unseen Domains. [paper]

Alignment with human representations supports robust few-shot learning. [paper]

Free Lunch for Domain Adversarial Training: Environment Label Smoothing. [paper]

Effective Robustness against Natural Distribution Shifts for Models with Different Training Data. [paper]

Leveraging Domain Relations for Domain Generalization. [paper]

Evasion Attacks and Defenses

Jailbroken: How Does LLM Safety Training Fails. [paper]

REaaS: Enabling Adversarially Robust Downstream Classifiers via Robust Encoder as a Service. [paper]

On adversarial robustness and the use of Wasserstein ascent-descent dynamics to enforce it. [paper]

On the Robustness of AlphaFold: A COVID-19 Case Study. [paper]

Data Augmentation Alone Can Improve Adversarial Training. [paper]

Improving the Accuracy-Robustness Trade-off of Classifiers via Adaptive Smoothing. [paper]

Uncovering Adversarial Risks of Test-Time Adaptation. [paper]

Benchmarking Robustness to Adversarial Image Obfuscations. [paper]

Are Defenses for Graph Neural Networks Robust? [paper]

On the Robustness of Randomized Ensembles to Adversarial Perturbations. [paper]

Defensive ML: Defending Architectural Side-channels with Adversarial Obfuscation. [paper]

Exploring and Exploiting Decision Boundary Dynamics for Adversarial Robustness. [paper]

Poisoning Attacks and Defenses

Poisoning Language Models During Instruction Tuning. [paper]

Backdoor Attacks Against Dataset Distillation. [paper]

Run-Off Election: Improved Provable Defense against Data Poisoning Attacks. [paper]

Temporal Robustness against Data Poisoning. [paper]

Poisoning Web-Scale Training Datasets is Practical. [paper]

CleanCLIP: Mitigating Data Poisoning Attacks in Multimodal Contrastive Learning. [paper]

TrojDiff: Trojan Attacks on Diffusion Models with Diverse Targets. [paper]

Privacy

SoK: Privacy-Preserving Data Synthesis. [paper]

Ticketed Learning-Unlearning Schemes. [paper]

Forgettable Federated Linear Learning with Certified Data Removal. [paper]

Privacy Auditing with One (1) Training Run. [paper]

DPMLBench: Holistic Evaluation of Differentially Private Machine Learning. [paper]

On User-Level Private Convex Optimization. [paper]

Re-thinking Model Inversion Attacks Against Deep Neural Networks. [paper]

A Recipe for Watermarking Diffusion Models. [paper]

CUDA: Convolution-based Unlearnable Datasets. [paper]

Why Is Public Pretraining Necessary for Private Model Training? [paper]

Personalized Privacy Auditing and Optimization at Test Time. [paper]

Interpretability

Towards Trustworthy Explanation: On Causal Rationalization. [paper]

Don't trust your eyes: on the (un)reliability of feature visualizations. [paper]

Probabilistic Concept Bottleneck Models. [paper]

Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence. [paper]

eXplainable Artificial Intelligence on Medical Images: A Survey. [paper]

 

 

擅长现代信号处理(改进小波分析系列,改进变分模态分解,改进经验小波变换,改进辛几何模态分解等等),改进机器学习,改进深度学习,机械故障诊断,改进时间序列分析(金融信号,心电信号,振动信号等)

 

 

评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

哥廷根数学学派

码字不易,且行且珍惜

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值