Skip to content
#

machine-learning-security

Here are 13 public repositories matching this topic...

Production-grade demonstration of AI safety mechanisms: Out-of-Distribution detection and adversarial robustness testing for NLP models using DistilBERT, PyTorch, and TextAttack.

  • Updated Nov 20, 2025
  • Python

Inspired by dynamic taint tracking, PoisonSpot uses fine-grained training provenance tracker that: (1) tags & traces the impact of every single training sample on model updates, (2) probabilistically scores suspect samples based on their linage of impact on model weights, and (3) separates the clean from the poisonous before retraining a model.

  • Updated Oct 27, 2025
  • Python

Improve this page

Add a description, image, and links to the machine-learning-security topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the machine-learning-security topic, visit your repo's landing page and select "manage topics."

Learn more