Government Threats Using Generative AI
Government Threats Using Generative AI
Misuse of
Generative AI
2025 January
Table of Contents
Foreword 1 1
Executive Summary 22
AI-Focused Threats 44
Findings 66
Findings 22
22
Large language models (LLMs) open new possibilities for defenders, from sifting through
complex telemetry to secure coding, vulnerability discovery, and streamlining operations.
However, some of these same AI capabilities are also available to attackers, leading to
understandable anxieties about the potential for AI to be misused for malicious purposes.
Much of the current discourse around cyber threat actors' misuse of AI is confined to theoretical
research. While these studies demonstrate the potential for malicious exploitation of AI, they
don't necessarily reflect the reality of how AI is currently being used by threat actors in the wild.
To bridge this gap, we are sharing a comprehensive analysis of how threat actors interacted with
Google's AI-powered assistant, Gemini. Our analysis was grounded by the expertise of Google's
Threat Intelligence Group (GTIG), which combines decades of experience tracking threat actors
on the front lines and protecting Google, our users, and our customers from government-backed
attackers, targeted 0-day exploits, coordinated information operations (IO), and serious cyber
crime networks.
We believe the private sector, governments, educational institutions, and other stakeholders
must work together to maximize AI's benefits while also reducing the risks of abuse. At Google,
we are committed to developing responsible AI guided by our principles, and we share resources
and best practices to enable responsible AI development across the industry. We continuously
improve our AI models to make them less susceptible to misuse, and we apply our intelligence
to improve Google's defenses and protect users from cyber threat activity. We also proactively
disrupt malicious activity to protect our users and help make the internet safer. We share our
findings with the security community to raise awareness and enable stronger protections for all.
1
Executive Summary
Google Threat Intelligence Group (GTIG)
is committed to tracking and protecting
against cyber threat activity.
We relentlessly defend Google, our users, and our customers by building the most complete
threat picture to disrupt adversaries. As part of that effort, we investigate activity associated
with threat actors to protect against malicious activity, including the misuse of generative
AI or LLMs.
This report shares our findings on government-backed threat actor use of the Gemini web
application. The report encompasses new findings across advanced persistent threat (APT)
and coordinated information operations (IO) actors tracked by GTIG. By using a mix of analyst
review and LLM-assisted analysis, we investigated prompts by APT and IO threat actors who
attempted to misuse Gemini.
GTIG takes a holistic, intelligence-driven approach to detecting and disrupting threat activity,
and our understanding of government-backed threat actors and their campaigns provides
the needed context to identify threat enabling activity. We use a wide variety of technical
signals to track government-backed threat actors and their infrastructure, and we are able
to correlate those signals with activity on our platforms to protect Google and our users.
By tracking this activity, we’re able to leverage our insights to counter threats across Google
platforms, including disrupting the activity of threat actors who have misused Gemini. We also
actively share our insights with the public to raise awareness and enable stronger protections
across the wider ecosystem.
Our analysis of government-backed threat actor use of Gemini focused on understanding how
threat actors are using AI in their operations and if any of this activity represents novel or unique
AI-enabled attack or abuse techniques. Our findings, which are consistent with those of our
industry peers, reveal that while AI can be a useful tool for threat actors, it is not yet the game-
changer it is sometimes portrayed to be. While we do see threat actors using generative AI to
perform common tasks like troubleshooting, research, and content generation, we do not see
indications of them developing novel capabilities.
2
Our key findings include:
> We did not observe any original or persistent attempts by threat actors to use prompt attacks or other
machine learning (ML)-focused threats as outlined in the Secure AI Framework (SAIF) risk taxonomy.
Rather than engineering tailored prompts, threat actors used more basic measures or publicly available
jailbreak prompts in unsuccessful attempts to bypass Gemini's safety controls.
> Threat actors are experimenting with Gemini to enable their operations, finding productivity gains
but not yet developing novel capabilities. At present, they primarily use AI for research, troubleshooting
code, and creating and localizing content.
> APT actors used Gemini to support several phases of the attack lifecycle, including researching
potential infrastructure and free hosting providers, reconnaissance on target organizations,
research into vulnerabilities, payload development, and assistance with malicious scripting and evasion
techniques. Iranian APT actors were the heaviest users of Gemini, using it for a wide range of purposes.
Of note, we observed limited use of Gemini by Russian APT actors during the period of analysis.
> IO actors used Gemini for research; content generation including developing personas and messaging;
translation and localization; and to find ways to increase their reach. Again, Iranian IO actors were the
heaviest users of Gemini, accounting for three quarters of all use by IO actors. We also observed Chinese
and Russian IO actors using Gemini primarily for general research and content creation.
> Gemini's safety and security measures restricted content that would enhance adversary capabilities
as observed in this dataset. Gemini provided assistance with common tasks like creating content,
summarizing, explaining complex concepts, and even simple coding tasks. Assisting with more elaborate
or explicitly malicious tasks generated safety responses from Gemini.
> Threat actors attempted unsuccessfully to use Gemini to enable abuse of Google products, including
researching techniques for Gmail phishing, stealing data, coding a Chrome infostealer, and bypassing
Google's account verification methods.
Rather than enabling disruptive change, generative AI allows threat actors to move faster
and at higher volume. For skilled actors, generative AI tools provide a helpful framework,
similar to the use of Metasploit or Cobalt Strike in cyber threat activity. For less skilled actors,
they also provide a learning and productivity tool, enabling them to more quickly develop tools
and incorporate existing techniques. However, current LLMs on their own are unlikely to enable
breakthrough capabilities for threat actors. We note that the AI landscape is in constant flux,
with new AI models and agentic systems emerging daily. As this evolution unfolds, GTIG
anticipates the threat landscape to evolve in stride as threat actors adopt new AI technologies
in their operations.
3
AI-Focused Threats
Attackers can use LLMs in two ways. One way is attempting to leverage LLMs
to accelerate their campaigns (e.g., by generating code for malware or content
for phishing emails). The overwhelming majority of activity we observed falls
into this category. The second way attackers can use LLMs is to instruct
a model or AI agent to take a malicious action (e.g., finding sensitive user data
and exfiltrating it). These risks are outlined in Google's Secure AI Framework
(SAIF) risk taxonomy.
Jailbreak attempts:
Basic and based on publicly available prompts
We observed a handful of cases of low-effort experimentation using publicly available jailbreak
prompts in unsuccessful attempts to bypass Gemini's safety controls. Threat actors copied
and pasted publicly available prompts and appended small variations in the final instruction
(e.g., basic instructions to create ransomware or malware). Gemini responded with safety
fallback responses and declined to follow the threat actor's instructions.
In one example of a failed jailbreak attempt, an APT actor copied publicly available prompts
into Gemini and appended basic instructions to perform coding tasks. These tasks included
encoding text from a file and writing it to an executable and writing Python code for a
distributed denial-of-service (DDoS) tool. In the former case, Gemini provided Python code
to convert Base64 to hex, but provided a safety filtered response when the user entered
a follow-up prompt that requested the same code as a VBScript.
What is an AI jailbreak?
4
The same group used a different publicly available jailbreak prompt to request Python code for
DDoS. Gemini provided a safety filtered response stating that it could not assist, and the threat
actor abandoned the session and did not attempt further interaction.
5
Findings:
Government-backed threat actors misusing Gemini
At a glance:
Government-backed threat actors
Government-backed attackers attempted to use Gemini for coding and scripting tasks,
gathering information about potential targets, researching publicly known vulnerabilities,
and enabling post-compromise activities, such as defense evasion in a target environment.
> Iran: Iranian APT actors were the heaviest users of Gemini, using it for a wide range of purposes,
including research on defense organizations, vulnerability research, and creating content for campaigns.
APT42 focused on crafting phishing campaigns, conducting reconnaissance on defense experts and
organizations, and generating content with cybersecurity themes.
> China: Chinese APT actors used Gemini to conduct reconnaissance, for scripting and development,
to troubleshoot code, and to research how to obtain deeper access to target networks. They focused
on topics such as lateral movement, privilege escalation, data exfiltration, and detection evasion.
> North Korea: North Korean APT actors used Gemini to support several phases of the attack lifecycle,
including researching potential infrastructure and free hosting providers, reconnaissance on target
organizations, payload development, and assistance with malicious scripting and evasion techniques.
They also used Gemini to research topics of strategic interest to the North Korean government,
such as the South Korean military and cryptocurrency. Of note, North Korean actors also used Gemini
to draft cover letters and research jobs—activities that would likely support North Korea's efforts
to place clandestine IT workers at Western companies.
> Russia: With Russian APT actors, we observed limited use of Gemini during the period of analysis. Their
Gemini use focused on coding tasks, including converting publicly available malware into another coding
language and adding encryption functions to existing code.
Google analyzed Gemini activity associated with known APT actors and identified APT
groups from more than 20 countries that used Gemini. The highest volume of usage was
from Iran and China. APT actors used Gemini to support several phases of the attack lifecycle,
including researching potential infrastructure and free hosting providers, reconnaissance on
target organizations, research into vulnerabilities, payload development, and assistance with
malicious scripting and evasion techniques. The top use cases by APT actors focused on:
> Assistance with coding tasks, including troubleshooting, tool and script development,
and converting or rewriting existing code
> Vulnerability research focused on publicly reported vulnerabilities and specific CVEs
> General research on various technologies, translations, and technical explanations
> Reconnaissance about likely targets, including details about specific organizations
> Enabling post-compromise activity, such as seeking advice on techniques to evade detection,
escalate privileges, or conduct internal reconnaissance in a target environment
6
ATTACK TOPICS OF
LIFECYCLE GEMINI USAGE
Recon - China
> Research on US military, US-based IT service providers
> Understand public database of US intelligence personnel
> Research on target network ranges; determine domain names of targets
Installation > Sign an Outlook Visual Studio Tools for Office (VSTO) plug-in and deploy
it silently to all computers
> Add a self-signed certificate to Active Directory
> Research Mimikatz for Windows 11
> Research Chrome extensions that provide parental controls and monitoring
Command and Control (C2) > Generate code to remotely access Windows Event Log
> Active Directory management commands
> JSON Web Token (JWT) security and routing rules in Ruby on Rails
> Character encoding issues in smbclient
> Command to check IPs of admins on the domain controller
Actions on Objectives > Automate workflows with Selenium (e.g., logging into compromised account)
> Generate a PHP script to extract emails from Gmail into electronic mail file
(EML) files
> Upload large files to OneDrive
> Solution to TLS 1.3 visibility challenges
7
Iranian government-backed actors
Iranian government-backed actors accounted for the largest Gemini use linked to APT
actors. Across this cohort, we observed a broad scope of research and use cases, including
to enable reconnaissance on targets, for research into publicly reported vulnerabilities,
to request translation and technical explanations, and to create content for possible use in
future campaigns. Their use reflected strategic Iranian interests including research focused
on defense organizations and experts, defense systems, foreign governments, individual
dissidents, the Israel-Hamas conflict, and social issues in Iran
We observed
APT actors use
Gemini to support
all phases of the
attack lifecycle.
8
At a glance:
Iranian APT actors using Gemini
9
Crafting phishing
campaigns
30% Over 30% of Iranian APT actors' Gemini use was
linked to APT42, whose Gemini activity reflected
the group's focus on crafting successful phishing
campaigns. We observed the group using Gemini
to conduct reconnaissance into individual policy
and defense experts, as well as organizations
of interest for the group.
Vulnerability research
The majority of APT42's efforts focused on
research into publicly known vulnerabilities,
such as a request to generate a list of critical
vulnerabilities from 2023. They also focused
on vulnerabilities in specific products such as
Mikrotik, Apereo, and Atlassian.
10
Research into military
and weapons systems
APT42 also appears to have used Gemini's
translation and explanation functions to better
understand publicly available information on
defense systems. Their efforts included general
research into the Israel-Hamas conflict, as well
as strategic trends in China's defense industry.
The threat actor also used Gemini for technical
explanations about US-made aerospace systems.
11
People's Republic of China (PRC)
government-backed actors
Government-backed actors linked to the People's Republic of China (PRC) attempted to
use Gemini to enable reconnaissance on targets, for scripting and development, to request
translation and explanation of technical concepts, and attempting to enable deeper access
to a network following initial compromise. PRC threat actors' usage resembled an IT admin
seeking to streamline, troubleshoot, or automate their tasks. In a malicious context, however,
this activity could be used to enable lateral movement, privilege escalation, data exfiltration,
and detection evasion.
Multiple PRC-backed
groups sought
Gemini's assistance
conducting research
and reconnaissance
on likely targets.
12
At a glance:
People’s Republic of China APT actors using Gemini
13
In another example, PRC-backed APT actors
Enabling deeper access asked Gemini for assistance with Active
in a target network Directory management commands and
requested help troubleshooting impacket,
PRC-backed APT actors also used Gemini to a Python-based tool for working with network
work through scripting and development tasks, protocols. While impacket is commonly used
many of which appeared intended to enable for benign purposes, the context of the threat
deeper access in a target network after threat actor made it clear that the actor was using
actors obtained initial access. For example, the tool for malicious purposes.
one PRC-backed group asked Gemini for
assistance figuring out how to sign a plugin
for Microsoft Outlook and silently deploy it
to all computers. The same actor also asked
Explaining tools,
Gemini to generate code to remotely access concepts, and code
Windows Event Log; sought instructions on
how to add a self-signed certificate to Active PRC actors utilized Gemini to learn
Directory; and asked Gemini for a command about specific tools and technologies and
to identify the IP addresses of administrators develop solutions to technical challenges.
on the domain controller. Other actors used For example, a PRC APT actor used Gemini
Gemini for help troubleshooting Chinese to break down how to use the graph database
character encoding issues in smbclient Nebula Graph. In another instance, the same
and how to record passwords on the actor used Gemini to offer possible solutions
VMware vCenter. to TLS 1.3 visibility challenges. Another PRC-
backed APT group sought to understand
a malicious PHP script.
14
Vulnerability research
and reverse engineering
In one case, a PRC-backed APT actor
attempted unsuccessfully to get Gemini's
help reverse engineering the endpoint
detection and response (EDR) tool Carbon
Black. The same threat actor copied
disassembled Python bytecode into Gemini
to convert the bytecode into Python code.
It's not clear what their objective was.
Unsuccessful attempts
to elicit internal system
information from Gemini
In one case, the PRC-backed APT actor APT41
attempted unsuccessfully to use Gemini to
learn about Gemini's underlying infrastructure
and systems. The actor asked Gemini to share
details such as its IP address, kernel version,
and network configuration. Gemini responded
but did not disclose sensitive information.
In a helpful tone, the responses provided
publicly available details that would be widely
known about the topic, while also indicating
that the requested information is kept secret
to prevent unauthorized access.
15
North Korean government-backed actors
North Korean APT actors used Gemini to support several phases of the attack lifecycle,
including researching potential infrastructure and free hosting providers, reconnaissance
on target organizations, payload development, and assistance with malicious scripting
and evasion techniques. They also used Gemini to research topics of strategic interest to
the North Korean government, such as South Korean nuclear technology and cryptocurrency.
We also observed that North Korean actors were using LLMs in likely attempts to enable
North Korea's efforts to place clandestine IT workers at Western companies.
16
At a glance:
North Korean APT actors using Gemini
17
Clandestine IT
worker threat
North Korean APT actors used Gemini to draft While normally employment-related research
cover letters and research jobs—activities would be typical for any job seeker, we assess
that would likely support efforts by North the usage is likely related to North Korea's
Korean nationals to use fake identities and ongoing efforts to place clandestine workers
obtain freelance and full-time jobs at foreign in freelance gigs or full-time jobs at Western
companies while concealing their true firms. The scheme, which involves thousands
identities and locations. One North Korea- of North Korean workers and has affected
backed group utilized Gemini to draft cover hundreds of US-based companies, uses IT
letters and proposals for job descriptions, workers with false identities to complete
researched average salaries for specific freelance work and send wages back
jobs, and asked about jobs on LinkedIn. to the North Korean regime.
The group also used Gemini for information
about overseas employee exchanges. Many
of the topics would be common for anyone
researching and applying for jobs.
Outside of their use of Gemini, North Korean cyber threat actors have shown a long-
standing interest in AI tools. They likely use AI applications to augment malicious
operations and improve efficiency and capabilities, and for producing content to
support their campaigns, such as phishing lures and profile photos for fake personas.
We assess with high confidence that North Korean cyber threat actors will continue
to demonstrate an interest in these emerging technologies for the foreseeable future.
DPRK IT Workers
We have observed DPRK IT Workers leverage accounts on assistive writing tools,
Monica (monica.im) and Ahrefs (ahrefs.com), which could potentially aid the group's
work despite a lack of language fluency. Additionally, the group has maintained
accounts on Data Annotation Tech, a company hiring individuals to train AI models.
Notably, a profile photo used by a suspected IT worker bore a noticeable resemblance
to multiple different images on the internet, suggesting that a manipulation tool was
used to generate the threat actor's profile photo.
APT43
Google Threat Intelligence Group (GTIG) has detected evidence of APT43 actors
accessing multiple publicly available LLM tools; however, the intended purpose
is not clear. Based on the capabilities of these platforms and historical APT43
activities, it is possible these applications could be used in the creation of rapport-
building emails, lure content, and malicious PowerShell and scripting efforts.
> GTIG has detected APT43 actors reference publicly available AI chatbot tools alongside
the topic "북핵 해결" (translation: "North Korean nuclear issue solution"), indicating the group
is using AI applications to conduct technical research as well as open-source analysis on
South Korean foreign and military affairs and nuclear issues.
> GTIG has identified APT43 actors accessing multiple publicly available AI image
generation tools, including tools used for image manipulation and creating realistic-
looking human portraits.
18
Target research Help with scripting,
and reconnaissance payload development,
North Korean actors also engaged with defense evasion
Gemini with several questions that appeared
focused on conducting initial research and North Korean actors also tried to use Gemini
reconnaissance into prospective targets. They to assist with development and scripting tasks.
also researched organizations and industries One North Korea-backed group attempted to
that are typical targets for North Korean actors, use Gemini to help develop webcam recording
including the US and South Korean militaries code in C++. Gemini provided multiple
and defense contractors. One North Korean versions of code, and repeated efforts by the
APT group asked Gemini for information about actor potentially suggested their frustration by
companies and organizations across a variety Gemini's answers. The same group also asked
of industry sectors and regions. Some of this Gemini to generate a robots.txt file to block
Gemini usage related directly to organizations crawlers and an .htaccess file to redirect all
that the same group had attempted to target in URLs except CSS extensions.
phishing and malware campaigns that Google
One North Korean APT actor used Gemini
previously detected and disrupted.
for assistance developing code for sandbox
In addition to research into companies, evasion. For example, the threat actor utilized
North Korean APT actors researched nuclear Gemini to write code in C++ to detect VM
technology and power plants in South Korea, environments and Hyper-V virtual machines.
such as site locations, recent news articles, Gemini provided responses with short
and the security status of the plants. Gemini code snippets to perform simple sandbox
responded with widely available, public checks. The same group also sought
information and facts that would be easily help troubleshooting Java errors when
discoverable in an online search. implementing AES encryption, and separately
asked Gemini if it is possible to acquire
a system password on Windows 11 using
Mimikatz.
19
Russian government-backed actors
During the period of analysis, we observed limited use of Gemini by Russia-backed APT
actors. Of this limited use, the majority of usage appeared benign, rather than threat-enabling.
The reasons for this low engagement are unclear. It is possible Russian actors avoided Gemini
out of operational security considerations, staying off Western-controlled platforms to avoid
monitoring of their activities. They may be using AI tools produced by Russian firms or locally
hosting LLMs, which would ensure full control of their infrastructure. Alternatively, they may
have favored other Western LLMs.
One Russian government-backed group used Gemini to request help with a handful of tasks,
including help rewriting publicly available malware into another language, adding encryption
functionality to code, and explanations for how a specific block of publicly available malicious
code functions.
20
At a glance:
Russian APT actors using Gemini
21
Findings:
Information operations (IO) actors misusing Gemini
At a glance:
Information operations actors
> Iran: Iranian IO actors used Gemini for a wide range of tasks, accounting for three quarters of all IO
prompts. They used Gemini for content creation and manipulation, including generating articles, rewriting
text with a specific tone, and optimizing it for better reach. Their activity also focused on translation and
localization, adapting content for different audiences, and for general research into news, current events,
and political issues.
> China: Pro-China IO actors used Gemini primarily for general research on various topics, including
a variety of topics of strategic interest to the Chinese government. The most prolific IO actor we track,
DRAGONBRIDGE, was responsible for the majority of this activity. They also used Gemini to research
current events and politics, and in a few cases, they used Gemini to generate articles or content
on specific topics.
> Russia: Russian IO actors used Gemini primarily for general research, content creation, and translation.
For example, their use involved assistance drafting content, rewriting article titles, and planning social
media campaigns. Some activity demonstrated an interest in developing AI capabilities, asking
for information on tools for creating online AI chatbots, developer tools for interacting with LLMs,
and options for textual content analysis.
IO actors used Gemini for research, content generation including developing personas and
messaging, translation and localization, and to find ways to increase their reach. Common use
cases include general research into news and current events as well as specific research into
individuals and organizations. In addition to creating content for campaigns, including personas
and content, the actors researched increasing the efficacy of campaigns, including automating
distribution, using search engine optimization (SEO) to optimize the reach of campaigns, and
increasing operational security. As with government-backed groups, IO actors also used Gemini
for translation and localization and for understanding the meanings or context of content.
22
Iran-linked information operations (IO) actors
Iran-based information operations (IO) groups used Gemini for a wide range of tasks, including
general research, translation and localization, content creation and manipulation, and generating
content with a specific bias or tone. We also observed Iran-based IO actors engage with Gemini
about news events and ask Gemini to provide details on economic and political issues in Iran,
the US, the Middle East, and Europe.
In line with their practice of mixing original and borrowed content, Iranian IO actors translated
existing material, including news-like articles. They then used Gemini to explain the context
and meaning of particular phrases within the given text.
Iran-based IO actors also used Gemini to localize the content, seeking human-like translation
and asking Gemini for help with tasks like making the text sound like a native English speaker.
They used Gemini to manipulate text (e.g., asking for help rewriting existing text on immigration
and crime in a specific style or tone).
Iran's activity also included research about improving the reach of their campaigns.
For example, they attempted to generate SEO-optimized content, likely in an effort to reach
a larger audience. Some actors also used Gemini to suggest strategies for increasing
engagement on social media.
23
At a glance:
Iran-linked IO actors using Gemini
24
PRC-linked information operations (IO) actors
IO actors linked to the People's Republic of China (PRC) used Gemini primarily for general
research on a wide variety of topics. The most prolific IO actor we track, the pro-China group
DRAGONBRIDGE, was responsible for approximately three quarters of this activity. Of their
activity, the majority use was general research about a wide variety of topics, ranging from
details about the features of various social media platforms to questions about various topics
of strategic interest to the PRC government. Actors researched information on current events
and politics in other regions, with a focus on the US and Taiwan. They also showed interest in
assessing the impact and risk of certain events. In a handful of cases, DRAGONBRIDGE used
Gemini to generate articles or content on specific topics.
DRAGONBRIDGE has experimented with other generative AI tools to create synthetic content
in support of their IO campaigns. As early as 2022, the group used a commercial AI service
in videos on YouTube to depict AI-generated news presenters. Their use of AI-generated video
continued through 2024 but has not resulted in significantly higher engagement from real
viewers. Google detected and terminated the channels distributing this content immediately
upon discovery. DRAGONBRIDGE's use of AI-generated videos or images has not resulted in
significantly higher engagement from real viewers.
25
At a glance:
PRC-linked IO actors using Gemini
26
Russia-linked information operations (IO) actors
Russian IO actors used Gemini for general research, content creation, and translation.
Half of this activity was associated with the Russian IO actor we track as KRYMSKYBRIDGE,
which is linked to a Russian consulting firm that works with the Russian government.
Approximately 40% of activity was linked to actors associated with Russian state sponsored
entities formerly controlled by the late Russian oligarch Yevgeny Prigozhin. We also observed
usage by actors tracked publicly as Doppelganger.
The majority of Russian IO actor usage was related to general research tasks, ranging from
the Russia-Ukraine war to details about various tools and online services. Russian IO actors
also used Gemini for content creation, rewriting article titles and planning social media
campaigns. Translation to and from Russian was also a common task.
Russian IO actors focused on the generative AI landscape, which may indicate an interest
in developing native capabilities in AI on infrastructure they control. They researched tools
that can be used to create an online AI chatbot and developer tools for interacting with LLMs.
One Russian IO actor used Gemini to suggest options for textual content analysis.
Pro-Russia IO actors have used AI in their influence campaigns in the past. In 2024, the actor
known as CopyCop likely used LLMs to generate content, and some stories on their sites
included metadata indicating an LLM was prompted to rewrite articles from genuine news
sources with a particular political perspective or tone. CopyCop's inauthentic news sites pose
as US- and Europe-based news outlets and post Kremlin-aligned views on Western policy,
the war in Ukraine, and domestic politics in the US and Europe.
27
At a glance:
Russia-linked IO actors using Gemini
28
Building AI safely
and responsibly
We believe our approach to AI must be both AI to identify potential vulnerabilities,
bold and responsible. To us, that means and creates new evaluation and training
developing AI in a way that maximizes the techniques to address misuse caused by
positive benefits to society while addressing them. In conjunction with this research,
the challenges. Guided by our AI Principles, DeepMind has shared how they're actively
Google designs AI systems with robust deploying defenses within AI systems along
security measures and strong safety with measurement and monitoring tools, one
guardrails, and we continuously test the of which is a robust evaluation framework
security and safety of our models to improve used to automatically red team an AI system's
them. Our policy guidelines and prohibited vulnerability to indirect prompt injection
use policies prioritize safety and responsible attacks. Our AI development and Trust &
use of Google's generative AI tools. Google's Safety teams also work closely with our
policy development process includes threat intelligence, security, and modelling
identifying emerging trends, thinking end-to- teams to stem misuse.
end, and designing for safety. We continuously
enhance safeguards in our products to offer The potential of AI, especially generative AI,
scaled protections to users across the globe. is immense. As innovation moves forward,
the industry needs security standards for
At Google, we leverage threat intelligence to building and deploying AI responsibly. That's
disrupt adversary operations. We investigate why we introduced the Secure AI Framework
abuse of our products, services, users and (SAIF), a conceptual framework to secure
platforms, including malicious cyber activities AI systems. We've shared a comprehensive
by government-backed threat actors, and toolkit for developers with resources and
work with law enforcement when appropriate. guidance for designing, building, and
Moreover, our learnings from countering evaluating AI models responsibly. We've
malicious activities are fed back into our also shared best practices for implementing
product development to improve safety and safeguards, evaluating model safety, and
security for our AI models. Google DeepMind red teaming to test and secure AI systems.
also develops threat models for generative
29
About the Authors
Google Threat Intelligence Group brings
together the Mandiant Intelligence and
Threat Analysis Group (TAG) teams, and
focuses on identifying, analyzing, mitigating,
and eliminating entire classes of cyber
threats against Alphabet, our users, and our
customers. Our work includes countering
threats from government-backed attackers,
targeted 0-day exploits, coordinated
information operations (IO), and serious
cyber crime networks. We apply our
intelligence to improve Google's defenses
and protect our users and customers.
30