GTIG AI Threat Tracker: Distilling, Experimenting, and (continued) Integrating AI for Adversarial Use

Applications of AI


Build AI safely and responsibly

We believe our approach to AI must be bold and responsible. This means developing AI in a way that maximizes positive benefits to society while addressing challenges. With our guidance, AI principlesGoogle designs its AI systems with robust security measures and strong safety guardrails, and continually tests and improves the security and safety of its models.

our policy guidelines and prohibited to use policy We prioritize the safety and responsible use of Google’s generative AI tools. Google Policy formulation process This includes identifying emerging trends, thinking end-to-end, and designing with safety in mind. We continually enhance our product safety measures to provide extensive protection to users around the world.

At Google, Leverage threat intelligence to destroy enemy strategy. We investigate abuse of our products, services, users, and platforms, including malicious cyber activity by government-sponsored actors, and work with law enforcement as appropriate. Additionally, learnings from responding to malicious activity are fed back into product development to improve the safety and security of AI models. These changes can be made at both the classifier and model level and are essential to keeping defenses agile and preventing further exploitation.

Google DeepMind will also develop generative AI threat models to identify potential vulnerabilities and create new assessment and training techniques to address exploits. In conjunction with this study, Google DeepMind shared how it is actively deploying defenses to AI systems, along with measurement and monitoring tools, including a robust assessment framework that can automatically red-team AI vulnerabilities against indirect prompt injection attacks.

Our AI development and Trust & Safety teams work closely with our threat intelligence, security, and modeling teams to stop exploits.

The potential for AI, especially generative AI, is enormous. As innovation advances, the industry needs security standards to build and deploy AI responsibly. Therefore, we introduced Secure AI Framework (SAIF)a conceptual framework for securing AI systems. We shared comprehensive content Toolkit for developers and Resources and guidance Design, build, and evaluate AI models responsibly. We also shared the following best practices: Implementing safeguards, Assessing model safety, red team To test and secure your AI systems, and our comprehensive Rapid injection approach.

It is important to work closely with industry partners to build strong protection for all users. To this end, we are fortunate to have strong collaborative partnerships with many researchers and are grateful for the efforts of these researchers and others in the community to support Red Team and improve our defenses.

Google also continues to invest in AI research. AI is built responsiblyand leveraging its potential to discover risks automatically. I introduced it last year, but big sleepis an AI agent developed by Google DeepMind and Google Project Zero that proactively searches for and discovers unknown security vulnerabilities in software. Since then, Big Sleep has discovered the first real-world security vulnerabilities and helped identify vulnerabilities that could be imminently exploited by threat actors, while GTIG was able to proactively block them. We are also experimenting with AI to not only find vulnerabilities but also patch them. Recently introduced code menderis an experimental AI-powered agent that uses the advanced reasoning capabilities of Gemini models to automatically fix critical code vulnerabilities.

Indicators of Compromise (IOC)

To assist the broader community in exploring and identifying the activities outlined in this blog post, we have made the IOC available for free. GTI Collection For registered users.

About the author

The Google Threat Intelligence Group is focused on identifying, analyzing, mitigating, and eliminating all types of cyber threats to Alphabet, our users, and our customers. Our work includes countering threats from government-sponsored attackers, targeted zero-day exploits, coordinated information operations (IO), and serious cybercrime networks. We use intelligence to strengthen Google’s defenses and protect our users and customers.



Source link