Proactively identify vulnerabilities through red teaming to produce safe, secure, and reliable models.
Deploy generative AI applications and agents in a safe, secure, and scalable way with guardrails.
Learn more about Amazon’s new models, and ActiveFence’s contribution in their Technical Report
As AI models become more powerful, ensuring their safety, trustworthiness, and resilience is more critical than ever. Mentioned in Amazon’s report, the Nova family of models, which power text, image, and video generation, agentic workflows, and long-context understanding, underwent extensive red teaming to identify and mitigate vulnerabilities before deployment. As a key partner in this process, ActiveFence’s AI safety and security red team played a pivotal role in strengthening Nova’s defenses against real-world threats, helping Amazon quickly build safer AI models that meet the highest industry standards.
Learn more about Amazon’s new models, and ActiveFence’s contribution in their Technical Report and Model Card.
When red teaming AI models, Amazon and ActiveFence identify a range of potential risks that could impact safety, security, and ethical deployment. These risks include:
Addressing these risks is crucial to building AI systems that are responsible, safe, and aligned with ethical standards.
Taxonomy and count of attack techniques used in red teaming exercises
ActiveFence’s red teaming process is designed to proactively stress-test AI systems by simulating adversarial attacks and edge-case scenarios. As an external AI safety partner, we provide real-world adversarial testing, ensuring Nova models can handle harmful content detection, misinformation mitigation, and manipulation attempts effectively and quickly.
Critically, ActiveFence shortened red teaming timelines from weeks to mere hours.
By using a combination of automatic and manual red teaming processes, we dramatically shorten red teaming timelines, helping Amazon ensure their Nova models are safe, secure – and timely
Taking it a step beyond merely identifying the vulnerability, we provide Amazon with the data required to mitigate it.
With this combined approach, we enable our clients, like Amazon and Cohere to mitigate vulnerabilities and risks within hours or days, a process which would previously take 6-8 weeks. By focusing on speedy outputs, we support our clients in launching the world’s most powerful models confidently – and on time.
As AI continues to evolve, safety and security must remain a top priority. Through our collaboration with Amazon, ActiveFence has helped push the boundaries of what’s possible in AI safety and security operations. Our expertise in content moderation, adversarial testing, and AI security ensures that companies like Amazon can deploy cutting-edge models on time and with confidence.
By working together, we are setting a new standard for Responsible AI — one that protects users, businesses, and the broader digital ecosystem.
Read more about ActiveFence’s impact on Amazon’s new Nova models in their Technical Report.
Learn more about ActiveFence’s AI safety and security solutions
For Trust & Safety leaders, ensuring team performance is a key part of keeping users safe. Here are 12 questions that will help evaluate your team's readiness.
ActiveFence announces its acquisition of Rewire, an innovator of online safety AI, to expand our automated detection of online harm.
ActiveFence provides cutting-edge AI Content Safety solutions, specifically designed for LLM-powered applications. By integrating with NVIDIA NeMo Guardrails, we’re making AI safety more accessible to businesses of all sizes.