Building Mithridatium: Detecting Hidden Backdoors in ML Models
As pretrained AI models become more common, one growing concern is whether those models can actually be trusted. A model may appear completely normal during testing, but behave maliciously when exposed to a hidden trigger. These attacks are known as backdoor or poisoning attacks, and they represent
ORIGINAL SOURCE →via Dev.to
ADVERTISEMENT
⚡ STAY AHEAD
Events like this, convergence-verified across 689 sources, land in your inbox every Sunday. Free.
GET THE SUNDAY BRIEFING →RELATED · conflict
- [CONFLICT] Intermodal Asia
- [CONFLICT] UNDRR Regional Office for Arab States
- [CONFLICT] Digital security in war and conflict: challenges for civil society and tools for resilience
- [CONFLICT] Securing the Untrusted Agentic Development Layer
- [CONFLICT] Ordu'nun Kumru Belediyesi'nden çocuk teşvik program! 10 çocuk yapana sıfır otomobil
- [CONFLICT] Hazine ve Maliye Bakanı Mehmet Şimşek: "Eşel mobil olmasaydı mazot 89 TL'ydi"