Skip to content
conflictMEDIUM2026-05-06 19:36 UTC

Building Mithridatium: Detecting Hidden Backdoors in ML Models

As pretrained AI models become more common, one growing concern is whether those models can actually be trusted. A model may appear completely normal during testing, but behave maliciously when exposed to a hidden trigger. These attacks are known as backdoor or poisoning attacks, and they represent

ADVERTISEMENT
⚡ STAY AHEAD

Events like this, convergence-verified across 689 sources, land in your inbox every Sunday. Free.

GET THE SUNDAY BRIEFING →

RELATED · conflict