Amazon Web Services (AWS) announces the availability of Amazon EC2 I7ie instances in the AWS Europe (Ireland) region. Designed for large storage I/O intensive workloads, these new instances are powered by 5th generation Intel Xeon Scalable processors with an all-core turbo frequency of 3.2 GHz, offering up to 40% better compute performance and 20% better price performance over existing I3en instances.
I7ie instances offer up to 120TB local NVMe storage density—the highest available in the cloud for storage optimized instances—and deliver up to twice as many vCPUs and memory compared to prior generation instances. Powered by 3rd generation AWS Nitro SSDs, these instances achieve up to 65% better real-time storage performance, up to 50% lower storage I/O latency, and 65% lower storage I/O latency variability compared to I3en instances. Additionally, the 16KB torn write prevention feature, enables customers to eliminate performance bottlenecks for database workloads.
I7ie instances are high-density storage-optimized instances, for workloads that demand rapid local storage with high random read/write performance and consistently low latency for accessing large data sets. These versatile instances are offered in eleven different sizes including 2 metal sizes, providing flexibility to match customers computational needs. They deliver up to 100 Gbps of network performance bandwidth, and 60 Gbps of dedicated bandwidth for Amazon Elastic Block Store (EBS), ensuring fast and efficient data transfer for the most demanding applications.
Starting today, Amazon Elastic Compute Cloud (Amazon EC2) M7i-flex instances powered by custom 4th Gen Intel Xeon Scalable processors (code-named Sapphire Rapids) are available in AWS Asia Pacific (Melbourne) Region. These custom processors, available only on AWS, offer up to 15% better performance over comparable x86-based Intel processors utilized by other cloud providers.
M7i-flex instances are the easiest way for you to get price-performance benefits for a majority of general-purpose workloads. They deliver up to 19% better price-performance compared to M6i. M7i-flex instances offer the most common sizes, from large to 16xlarge, and are a great first choice for applications that don’t fully utilize all compute resources such as web and application servers, virtual-desktops, batch-processing, and microservices. In addition, these instances support the new Intel Advanced Matrix Extensions (AMX) that accelerate matrix multiplication operations for applications such as CPU-based ML. For workloads that need larger instance sizes (up to 192 vCPUs and 768 GiB memory) or continuous high CPU usage, you can leverage M7i instances.
To learn more, visit Amazon EC2 M7i-flex instance page.
We’re at an inflection point right now, where every industry and entire societies are witnessing sweeping change, with AI as the driving force. This isn’t just about incremental improvements, it’s about total transformation. The public sector is already experiencing sweeping change with the introduction of AI, and that pace will only intensify. This is the promise of AI, and it’s here and now. At our recent Google Cloud Next ‘25 we showcased our latest innovations and reinforced our commitment to bringing the latest and best technologies to help public sector agencies meet their missions.
Key public sector announcements at Next
It was an exciting week at Next ‘25 with hundreds of product and customer announcements from Google Cloud. Here are key AI, security, and productivity announcements that can help the public sector deliver improved services, enhance decision-making and operate with greater efficiency.
Advancements in Google Distributed Cloud that let customers bring Gemini models on premises. This compliments our GDC air-gapped product, now authorized for U.S. Government Secret and Top Secret levels, and on which Gemini is available, provides the highest levels of security and compliance. This enables public sector agencies to have greater flexibility in how and where they access the latest Google AI innovations.
Support for a full suite of generative media models and Gemini 2.5 – Our most intelligent model yet, Gemini 2.5 is designed for the agentic era and now available in Vertex AI platform. This builds on our recent announcement that Vertex AI Search and Generative AI (with Gemini) achieve FedRAMP High authorization,providing agencies with a secure platform and the latest AI innovations and capabilities.
Simplifying security with the launch of Google Unified Security– We are offering customers a security solution powered by AI that brings together our best-in-class security products for threat intelligence, security operations, cloud security, and secure enterprise browsing, along with Mandiant expertise to provide a unified view and improved threat detection across complex infrastructures.
Transforming agency productivity and unlocking significant savings – We are offering Google Workspace, our FedRAMP High authorized communication and collaboration platform, at a significant discount of 71% off for U.S. federal government agencies. This offering in combination with Gemini in Workspace being authorized at the FedRAMP High level gives unprecedented access to cutting edge AI services for U.S. government workers.
Helping customers meet their mission
All of this incredible technology – and more – came to life on stage and across the showfloor at our Google Public Sector Hub, where we showcased our solutions for security, defense, transportation, productivity & automation, education, citizen services, health & human services, and Google Distributed Cloud (GDC). In case you missed our live demos on Medicaid redetermination, unemployment insurance claims, transportation coordination, and research grant sourcing, contact us to schedule a virtual demo or discuss a pilot. To get hands on with the technology register for an upcoming Google Cloud Days training for the public sector here.
We are proud to work with customers across the public sector, as they apply the latest Google innovations and technologies to achieve real mission-value impact. Ai2 and Google Cloud announced a partnership with Google Cloud to make its portfolio of open AI models available in Vertex AI Model Garden. The collaboration will help set a new standard for openness that leverages Google Cloud’s infrastructure resources and AI development platform with Ai2’s open models that will advance AI research and offer enterprise-quality deployment for the public sector. This builds on our announcement that Ai2 and Google Cloud will commit $20M to advance AI-powered research for the Cancer AI Alliance. You can catch the highlights from my conversation at Next with Ali Farhadi, CEO of Ai2 here.
CEO perspectives: A new era of AI-powered research and innovation
All of this incredible innovation with our customers is further enabled by our ecosystem of partners who help us scale our impact across the public sector. At Google Cloud Next, Accenture Federal Services and Google Public Sector announced the launch of a joint Managed Extended Detection and Response (MxDR) solution. The new MxDR for government solution integrates Google Security Operations (SecOps) platform with Accenture Federal’s deep federal cybersecurity expertise. This solution uses security-specific generative artificial intelligence (Gen AI) to significantly enhance threat detection and response, and the overall security posture for federal agencies.
Lastly, Lockheed Martin and Google Public Sector also announced a collaboration to advance generative AI for national security. Integrating Google’s advanced generative artificial intelligence into Lockheed Martin’s AI Factory ecosystem will enhance Lockheed Martin’s ability to train, deploy, and sustain high-performance AI models and accelerate AI-driven capabilities in critical national security, aerospace, and scientific applications.
A new era of innovation and growth
AI presents a unique opportunity to enter a new era of innovation and economic growth, enabling the public sector to get more out of limited resources to improve public services and infrastructure, make public systems more secure, and better meet the needs of their constituents. Harnessing the power of AI can help governments become agile and more secure, and serve citizens better. At Google Public Sector, we’re passionate about applying the latest cloud, AI and security innovations to help you meet your mission.
Subscribe to our Google Public Sector Newsletter to stay informed and stay ahead with the latest updates, announcements, events and more.
Amazon Q Developer in the AWS Management Console and Amazon Q Developer in the IDE is now GA in the Europe (Frankfurt) Region.
Pro tier customers can now use and configure Amazon Q Developer in the AWS Management Console and Amazon Q Developer in the IDE to store data in the Europe (Frankfurt) Region and perform inference in European Union (EU) Regions giving them more choice over where their data resides and transits. Amazon Q Developer Administrators can configure their user settings so that data is stored in Europe (Frankfurt) Region and inference is performed in EU geographies using cross-region inference (CRIS) to reduce latency and optimize availability. If you are requesting to contact AWS Support your data will be processed in the US East (N. Virginia) region.
Amazon Q Developer in is generally available, and you can use it in the following AWS Regions: US East (N. Virginia), and Europe (Frankfurt).
Today, Amazon Simple Email Service (SES) launched support for logging email sending events through AWS CloudTrail. Customers can maintain a record of email send actions performed using the SES APIs, including actions taken by a user, role, or an AWS service in SES.
Previously, customers could use SES event destinations to route sending event notifications to custom data stores they created and managed themselves. This required custom solutions for data storage and data indexing, including development costs and operational oversight costs. Now, customers can configure event logging to AWS CloudTrail without any custom solution development. Customers can search for events, view the events, and download lists of events for processing in their private workflows. This gives customers a turn-key solution for event history management.
SES supports AWS CloudTrail data events for sending events in all AWS Regions where SES is available.
Today, Amazon Q Business is launching a feature to reduce hallucinations in chat responses. Hallucinations are confident responses made by generative AI applications that are not justified by its underlying data. The new feature enables customers to mitigate hallucinations in real-time during chat conversations.
Large Language Models (LLMs) underlying generative AI applications have reduced the extent of hallucination in their responses, but it is possible that these models could hallucinate. Hallucination mitigation is therefore needed to generate reliable and trustworthy responses. The Q Business hallucination mitigationfeature helps ensure more accurate retrieval augmented generation (RAG) responses from data connected to the application. This data could either come from connected data sources, or from files uploaded during chat. During chat, Q Business evaluates a response for hallucinations. If a hallucination is detected with high confidence, it corrects the inconsistencies in its response real-time during chat and generates a new, edited message.
The feature for Amazon Q Business is available in all regions where Q Business is available. Customers can opt into using this feature by enabling it through API or through the Amazon Q console. For more details, refer to the documentation. For more information about Amazon Q Business and its features, please visit the Amazon Q product page.
AWS Lambda@Edge now supports AWS Lambda’s advanced logging controls to improve how function logs are captured, processed, and consumed at the edge. This enhancement provides you with more control over your logging data, making it easier to monitor application behavior and quickly resolve issues.
The new advanced logging controls for Lambda@Edge give you three flexible ways to manage and analyze your logs. New JSON structured logs make it easier to search, filter, and analyze large volumes of log entries without using custom logging libraries. Log level granularity controls can switch log levels instantly, allowing you to filter for specific types of logs like errors or debug information when investigating issues. Custom CloudWatch log group selection lets you choose which Amazon CloudWatch log group Lambda@Edge sends logs to, making it easier to aggregate and manage logs at scale.
To get started, you can specify advanced logging controls for your Lambda functions using Lambda APIs, Lambda console, AWS CLI, AWS Serverless Application Model (SAM), and AWS CloudFormation. To learn more, visit the Lambda Developer Guide, and the CloudFront Developer Guide.
Last week at Google Cloud Next 25, we announced Cloud WAN, a fully managed, reliable, and secure solution for enterprise wide area network (WAN) architectures that’s built on Google’s planet-scale network. Today, we begin a series of deep dives into the products that power Cloud WAN, starting with NCC Gateway, a newregionally managed spoke of Network Connectivity Center (NCC) that integrates cloud-native security services, starting with third-party security service edge (SSE) solutions.
Securing the modern hybrid workforce is complex, driven by the surge of SaaS and remote work. In fact, many enterprises still employ disparate security stacks for on-premises and remote users. For on-prem deployments — especially for branches and campuses — a common approach is to use a colocation-based architecture, in which regional branches aggregate traffic in a colo using SD-WAN headends or VPN concentrators, and firewalls secure user traffic. However, remote users often connect via SSE, resulting in inconsistent security enforcement policies for remote and on-prem users.
To state the obvious, managing separate solutions for on-prem and remote user access to public and private applications can be challenging for security administrators.
For on-prem users and applications:
There’s no good scalable and cost-efficient way to send aggregated traffic from a colocation facility to SSE, so organizations continue to use firewalls to secure access to public and private applications. This results in complex configurations, lengthy onboarding processes, and costly infrastructure upgrades.
Firewalls in colocation facilities need to be sized for peak capacity and high availability, increasing total cost of ownership (TCO).
Remote users:
Disjointed security approaches across SSE and colocation firewalls creates inconsistencies between remote and on-prem users’ security postures.
Using VPN tunnels or application connectors for remote access to cloud resources introduces considerable overhead. This stems from the performance limitations of these connections and the operational complexity of managing numerous tunnels, resulting in higher latency for remote users.
The cloud-first era demands simpler, cloud-delivered security, without the complexity of traditional on-prem routing. Critically, businesses need a single, cloud-native security approach that delivers consistent controls and policies across every application and user, spanning cloud, on-prem, and SaaS environments. Cloud WAN using NCC Gateway is the first major cloud solution to offer managed integration of security service edge (SSE) for users accessing private and public applications. With integration to SSE solutions like Palo Alto Networks Prisma Access and Broadcom Cloud SWG, NCC Gateway offers enterprises a streamlined approach to securing their distributed workforce and applications using the provider of their choice.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 to try Google Cloud networking’), (‘body’, <wagtail.rich_text.RichText object at 0x3ec2cadb7100>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectpath=/products?#networking’), (‘image’, None)])]>
What is NCC Gateway?
For organizations managing complex hybrid and multi-cloud environments, Google Cloud’s Network Connectivity Center has long provided a simplified, unified management experience powered by Google’s global infrastructure. Now, we’re thrilled to announce an evolution that takes security to the next level with NCC Gateway.
Imagine a unified security solution that protects all your users, regardless of their location or how they connect — whether it’s through Cloud Interconnect, SD-WAN, Cloud VPN, or even the public internet. With NCC Gateway’s managed integration of third-party SSE, secure access to your private, public, and Google Cloud APIs is now a reality across your entire distributed infrastructure.
NCC Gateway does all this by eliminating the complexities of traditional IPSec tunnel management and traffic steering, enabling rapid onboarding of branch locations and helping to optimize performance for high-bandwidth applications. This ensures that user traffic is securely routed through the chosen SSE stack, while maintaining privacy and integrity within Google Cloud’s private network to minimize latency and enhance the overall user experience.
Key use cases
Here are three key use cases where NCC Gateway simplifies your network security and boosts performance:
1. Streamlined, high-bandwidth on-ramp for branch users
NCC Gateway provides a high-performance on-ramp for branch users connecting over 10 or 100 Gbps Cloud Interconnect, a substantial improvement over single-gigabit IPsec tunnels. This ensures dedicated, high-throughput connectivity, for optimal application performance. Then, following SSE inspection, traffic is efficiently routed to public applications on the internet, to private applications over the Google backbone, or to applications in other clouds via Cross-Cloud Interconnect.
2. High-performance, private off-ramp to private applications for remote users
NCC Gateway natively integrates third-party SSE stacks within Google Cloud’s private backbone. This removes the need for internet-based encryption, while maintaining privacy and integrity with higher performance. For applications running in other clouds, customers can leverage private connectivity with dedicated bandwidth through Cross-Cloud Interconnect backed with an SLA.
3. Protected application access to the internet
NCC Gateway provides a unified secure internet gateway for users and applications that are on-prem or in other clouds, while offering streamlined multi-gigabit onboarding with minimal configuration, eliminating complex tunnel management, and enabling rapid, secure deployment. For internet-bound SaaS traffic, Google’s Premium Tier network sends data to the best peering location, for optimized and secure access.
Key benefits of Cloud WAN with NCC Gateway
The addition of NCC Gateway to Cloud WAN brings a number of advantages:
Unified security posture: Enhance your security posture by consolidating your security stack and minimizing the attack surface. NCC Gateway enforces consistent ingress and egress security through Cloud WAN, providing a uniform security experience for all users, regardless of location or device, with your preferred SSE provider.
Improved application experience: Deliver a superior user experience with lower latency for both SaaS and private applications, powered by our premium backbone and native encryption. Cloud WAN provides up to 40% improved performance compared to the public internet.1
Lower costs: Achieve significant cost savings by streamlining multi-cloud connectivity and adopting a consumption-based model. Cloud WAN provides up to a 40% savings in total cost of ownership (TCO) over a customer-managed WAN solution.2
What our partners are saying
This is what our SSE partners had to say about NCC Gateway integration.
Palo Alto Networks:
“The integration of Prisma SASE with Cloud WAN unlocks new possibilities for customers, offering a high-bandwidth on-ramp to Prisma Access from large branches and campuses, while providing a high-performance private off-ramp to optimize secure access to private applications in Google Cloud or any other clouds.” – Anupam Upadhyaya, Vice President, Product Management, Palo Alto Networks
Broadcom:
“In an age where enterprises demand cutting-edge security at line speed, we’re proud to partner with Google Cloud to deliver a game-changing solution: Symantec Security Service Edge (SSE) natively integrated into Cloud WAN via Symantec Cloud SWG Express Connect. This gives our joint customers a secure express lane to critical data and seamless access to world-class AI capabilities. It’s another major step in our mission to bring enterprise-grade security to all.” – Jason Rolleston, General Manager, Enterprise Security Group, Broadcom
Learn more
Be among the first to experience the power of NCC Gateway, which will be available in preview in Q2 ‘25. You can learn more about Cloud WAN on the Cross-Cloud Network solution page.
1. During testing, network latency was more than 40% lower when traffic to a target traveled over the Cross-Cloud Network compared to when traffic to the same target traveled across the public internet. 2. Architecture includes SD-WAN and 3rd party firewalls, and compares a customer-managed WAN using multi-site colocation facilities to a WAN managed and hosted by Google Cloud.
Last week at Google Cloud Next 25, we announced Cloud WAN, a fully managed, reliable, and secure solution for enterprise wide area network (WAN) architectures that’s built on Google’s planet-scale network. Today, we begin a series of deep dives into the products that power Cloud WAN, starting with NCC Gateway, a newregionally managed spoke of Network Connectivity Center (NCC) that integrates cloud-native security services, starting with third-party security service edge (SSE) solutions.
Securing the modern hybrid workforce is complex, driven by the surge of SaaS and remote work. In fact, many enterprises still employ disparate security stacks for on-premises and remote users. For on-prem deployments — especially for branches and campuses — a common approach is to use a colocation-based architecture, in which regional branches aggregate traffic in a colo using SD-WAN headends or VPN concentrators, and firewalls secure user traffic. However, remote users often connect via SSE, resulting in inconsistent security enforcement policies for remote and on-prem users.
To state the obvious, managing separate solutions for on-prem and remote user access to public and private applications can be challenging for security administrators.
For on-prem users and applications:
There’s no good scalable and cost-efficient way to send aggregated traffic from a colocation facility to SSE, so organizations continue to use firewalls to secure access to public and private applications. This results in complex configurations, lengthy onboarding processes, and costly infrastructure upgrades.
Firewalls in colocation facilities need to be sized for peak capacity and high availability, increasing total cost of ownership (TCO).
Remote users:
Disjointed security approaches across SSE and colocation firewalls creates inconsistencies between remote and on-prem users’ security postures.
Using VPN tunnels or application connectors for remote access to cloud resources introduces considerable overhead. This stems from the performance limitations of these connections and the operational complexity of managing numerous tunnels, resulting in higher latency for remote users.
The cloud-first era demands simpler, cloud-delivered security, without the complexity of traditional on-prem routing. Critically, businesses need a single, cloud-native security approach that delivers consistent controls and policies across every application and user, spanning cloud, on-prem, and SaaS environments. Cloud WAN using NCC Gateway is the first major cloud solution to offer managed integration of security service edge (SSE) for users accessing private and public applications. With integration to SSE solutions like Palo Alto Networks Prisma Access and Broadcom Cloud SWG, NCC Gateway offers enterprises a streamlined approach to securing their distributed workforce and applications using the provider of their choice.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 to try Google Cloud networking’), (‘body’, <wagtail.rich_text.RichText object at 0x3e1b95a84250>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectpath=/products?#networking’), (‘image’, None)])]>
What is NCC Gateway?
For organizations managing complex hybrid and multi-cloud environments, Google Cloud’s Network Connectivity Center has long provided a simplified, unified management experience powered by Google’s global infrastructure. Now, we’re thrilled to announce an evolution that takes security to the next level with NCC Gateway.
Imagine a unified security solution that protects all your users, regardless of their location or how they connect — whether it’s through Cloud Interconnect, SD-WAN, Cloud VPN, or even the public internet. With NCC Gateway’s managed integration of third-party SSE, secure access to your private, public, and Google Cloud APIs is now a reality across your entire distributed infrastructure.
NCC Gateway does all this by eliminating the complexities of traditional IPSec tunnel management and traffic steering, enabling rapid onboarding of branch locations and helping to optimize performance for high-bandwidth applications. This ensures that user traffic is securely routed through the chosen SSE stack, while maintaining privacy and integrity within Google Cloud’s private network to minimize latency and enhance the overall user experience.
Key use cases
Here are three key use cases where NCC Gateway simplifies your network security and boosts performance:
1. Streamlined, high-bandwidth on-ramp for branch users
NCC Gateway provides a high-performance on-ramp for branch users connecting over 10 or 100 Gbps Cloud Interconnect, a substantial improvement over single-gigabit IPsec tunnels. This ensures dedicated, high-throughput connectivity, for optimal application performance. Then, following SSE inspection, traffic is efficiently routed to public applications on the internet, to private applications over the Google backbone, or to applications in other clouds via Cross-Cloud Interconnect.
2. High-performance, private off-ramp to private applications for remote users
NCC Gateway natively integrates third-party SSE stacks within Google Cloud’s private backbone. This removes the need for internet-based encryption, while maintaining privacy and integrity with higher performance. For applications running in other clouds, customers can leverage private connectivity with dedicated bandwidth through Cross-Cloud Interconnect backed with an SLA.
3. Protected application access to the internet
NCC Gateway provides a unified secure internet gateway for users and applications that are on-prem or in other clouds, while offering streamlined multi-gigabit onboarding with minimal configuration, eliminating complex tunnel management, and enabling rapid, secure deployment. For internet-bound SaaS traffic, Google’s Premium Tier network sends data to the best peering location, for optimized and secure access.
Key benefits of Cloud WAN with NCC Gateway
The addition of NCC Gateway to Cloud WAN brings a number of advantages:
Unified security posture: Enhance your security posture by consolidating your security stack and minimizing the attack surface. NCC Gateway enforces consistent ingress and egress security through Cloud WAN, providing a uniform security experience for all users, regardless of location or device, with your preferred SSE provider.
Improved application experience: Deliver a superior user experience with lower latency for both SaaS and private applications, powered by our premium backbone and native encryption. Cloud WAN provides up to 40% improved performance compared to the public internet.1
Lower costs: Achieve significant cost savings by streamlining multi-cloud connectivity and adopting a consumption-based model. Cloud WAN provides up to a 40% savings in total cost of ownership (TCO) over a customer-managed WAN solution.2
What our partners are saying
This is what our SSE partners had to say about NCC Gateway integration.
Palo Alto Networks:
“The integration of Prisma SASE with Cloud WAN unlocks new possibilities for customers, offering a high-bandwidth on-ramp to Prisma Access from large branches and campuses, while providing a high-performance private off-ramp to optimize secure access to private applications in Google Cloud or any other clouds.” – Anupam Upadhyaya, Vice President, Product Management, Palo Alto Networks
Broadcom:
“In an age where enterprises demand cutting-edge security at line speed, we’re proud to partner with Google Cloud to deliver a game-changing solution: Symantec Security Service Edge (SSE) natively integrated into Cloud WAN via Symantec Cloud SWG Express Connect. This gives our joint customers a secure express lane to critical data and seamless access to world-class AI capabilities. It’s another major step in our mission to bring enterprise-grade security to all.” – Jason Rolleston, General Manager, Enterprise Security Group, Broadcom
Learn more
Be among the first to experience the power of NCC Gateway, which will be available in preview in Q2 ‘25. You can learn more about Cloud WAN on the Cross-Cloud Network solution page.
1. During testing, network latency was more than 40% lower when traffic to a target traveled over the Cross-Cloud Network compared to when traffic to the same target traveled across the public internet. 2. Architecture includes SD-WAN and 3rd party firewalls, and compares a customer-managed WAN using multi-site colocation facilities to a WAN managed and hosted by Google Cloud.
Google Cloud Next 25 took place this week and we’re all still buzzing! It was a jam-packed week in Las Vegas complete with interactive experiences, including more than 10 keynotes and spotlights, 700 sessions, and 350+ sponsoring partners joining us for an incredible Expo show. Attendees enjoyed hands-on learning across AI innovation, data cloud, modern infrastructure, security, Google Workspace, and more.
At our opening keynote, we showcased cutting-edge product innovations across our AI-optimized platform and featured hundreds of customers and partners building with Google Cloud as well as five awesome demos. You can catch up on all the highlights in our 10-minute keynote recap.
Our developer keynoteshowed how AI is revolutionizing the developer workflow, and featured seven incredible demos on everything from building with Gemini to creating multi-agent systems.
1_next25 wrap
2_next25 wrap
3_next25 wrap
4_next25 wrap
5_next25 wrap
6_next25 wrap
Last year, we shared how customers were exploring the exciting potential of generative AI to transform the way they work. This year, we showcased how customers are getting real business value from Google AI, celebrating hundreds of customer stories across the event, including the amazing story of how The Sphere is using Google AI to enrich their fully immersive The Wizard of Oz experience.
It was a busy week, so we’ve prepared a summary of all the 228 announcements from Next ‘25 below:
AI and Multi-Agent Systems
Models: Building on Google DeepMind research, we announced the addition of a variety of first-party models, as well as new third-party models to Vertex AI Model Garden.
1. Gemini 2.5 Pro is available in public preview on Vertex AI, AI Studio, and in the Gemini app. Gemini 2.5 Pro is engineered for maximum quality and tackling the most complex tasks demanding deep reasoning and coding expertise. It is ranked #1 on Chatbot Arena.
2. Gemini 2.5 Flash — our low latency and most cost-efficient thinking model — is coming soon to Vertex AI, AI Studio, and in the Gemini app.
3. Imagen 3: Our highest quality text-to-image model now has improved image generation and inpainting capabilities for reconstructing missing or damaged portions of an image.
5. Lyria: The industry’s first enterprise-ready, text-to-music model, transforms simple text prompts into 30-second music clips.
6. Veo 2: Our advanced video generation model has new editing and camera control features to help customers refine and repurpose video content with precision.
9. Vertex AI Dashboards: These help you monitor usage, throughput, latency, and troubleshoot errors, providing you with greater visibility and control.
10. Model Customization and Tuning: You can also manage custom training and tuning with your own data on top of foundational models in a secure manner across all first-party model families including Gemini, Imagen, Veo, embedding, and translation models, as well as open models like Gemma, Llama, and Mistral.
11. Vertex AI Model Optimizer: Automatically generate the highest quality response for each prompt based on your desired balance of quality and cost
12. Live API: Offers streaming audio and video directly into Gemini. Now your agents can process and respond to rich media in real time, opening new possibilities for immersive, multimodal applications.
13. Vertex AI Global Endpoint: Provides capacity-aware routing for our Gemini models across multiple regions, maintaining application responsiveness even during peak traffic or regional service fluctuations.
We also introduced new capabilities to help you build and manage multi-agent systems — regardless of which technology framework or model you’ve chosen.
14. Agent Development Kit (ADK): This open-source framework simplifies the process of building sophisticated multi-agent systems while maintaining precise control over agent behavior.Agent Development Kit supports the Model Context Protocol (MCP)which provides a unified way for AI models to access and interact with various data sources and tools, rather than requiring custom integrations for each.
15. Agent2Agent (A2A) protocol: We’re proud to be the first hyperscaler to create an open Agent2Agent protocol to help enterprises support multi-agent ecosystems, so agents can communicate with each other, regardless of the underlying framework or model. More than 50 partners, including Accenture, Box, Deloitte, Salesforce, SAP, ServiceNow, and TCSare actively contributing to defining this protocol, representing a shared vision of multi-agent systems.
16. Agent Garden: This collection of ready-to-use samples and tools is directly accessible in ADK. Leverage pre-built agent patterns and components to accelerate your development process and learn from working examples.
17. Agent Engine: This fully managed agent runtime in Vertex AI helps you deploy your custom agents to production with built-in testing, release, and reliability at a global, secure scale.
18. Grounding with Google Maps1: For agents that rely on geospatial context, you can now ground your agents with Google Maps, so they can provide responses with geospatial information tied to places in the U.S.
19. Customer Engagement Suite: This latest version includes human-like voices; the ability to understand emotions so agents can adapt better during conversation; streaming video support so AI agents can interpret and respond to what they see in real-time through customer devices; and AI assistance to build agents in a no-code interface.
We announced exciting enhancements to Google Agentspace to help scale the adoption of enterprise search and AI agents across the enterprise. Agentspace puts the latest Google foundation models, Google-quality search, powerful AI agents, and actionable enterprise knowledge in the hands of every employee.
20. Integrated with Chrome Enterprise: Bringing Agentspace directly into Chrome helps employees easily and securely find information, including data and resources, right within their existing workflows.
21. Agent Gallery: This provides employees a single view of available agents across the enterprise, including those from Google, internal teams, and partners — making agents easy to discover and use.
22. Agent Designer: A no-code interface for creating custom agents that automate everyday work tasks or enhance knowledge. Agent Designer helps employees adapt agents to their individual workflows and needs, no matter their technical experience.
23. Idea Generation agent: Helps employees innovate by autonomously developing novel ideas in any domain, then evaluating them to find the best solutions via a competitive system inspired by the scientific method.
24. Deep Research agent: Explores complex topics on the employee’s behalf, synthesizing information across internal and external sources into comprehensive, easy-to-read reports — all with a single prompt.
We brought the best of Google DeepMind and Google Research together with new infrastructure and AI capabilities in Google Cloud, including:
25. AlphaFold 3: Developed by Google DeepMind and Isomorphic Labs, the new AlphaFold 3 High-Throughput Solution, available for non-commercial use and deployable via Google Cloud Cluster Toolkit, enables efficient batch processing of up to tens of thousands of protein sequences while minimizing cost through autoscaling infrastructure.
26. WeatherNext AI models: Google DeepMind and Google Research WeatherNext models enable fast, accurate weather forecasting, and are now available in Vertex AI Model Garden, allowing organizations to customize and deploy them for various research and industry applications.
27. Ironwood: Our 7th generation TPUjoins our AI-optimized hardware portfolio to power thinking, inferential AI models at scale (coming later in 2025). Read more here.
28. Google Distributed Cloud (GDC): We have partnered with NVIDIA to bring Gemini to NVIDIA Blackwell systems, with Dell as a key partner, so Gemini can be used locally in air-gapped and connected environments. Read more here.
29. Pathways on Cloud: Developed by Google DeepMind, Pathways is a distributed runtime that powers all of AI at Google, and is now available for the first time on Google Cloud.
30. vLLM on TPU: We’re bringing vLLM to TPUs to make it easy to run inference on TPUs. Customers who have optimized PyTorch with vLLM can how run inference on TPUs without changing their software stack, and also serve on both TPUs and GPUs if needed.
31. Dynamic Workload Scheduler resource management and job scheduling platform now features support for Trillium, TPU v5e, A4 (NVIDIA B200), and A3 Ultra (NVIDIA H200) VMs in preview via Flex Start mode, with Calendar mode support for TPUs coming later this month.
32. A4 and A4X VMs: We’ve significantly enhanced our GPU portfolio with the availability of A4 and A4X VMs powered by NVIDIA’s B200 and GB200 Blackwell GPUs, respectively, and A4X VMs are now in preview. We were the first cloud provider to offer both of these options.
33. NVIDIA Vera Rubin GPUs: Google Cloud will be among the first to offer NVIDIA’s next-generation Vera Rubin GPUs, which offer up to 15 exaflops of FP4 inference performance per rack.
34. Cluster Director (formerly Hypercompute Cluster) lets you deploy and manage a group of accelerators as a single unit with physically colocated VMs, targeted workload placement, advanced cluster maintenance controls, and topology-aware scheduling. New updates coming later this year include Cluster Director for Slurm, 3600 observability features, and job continuity capabilities. Register to join the preview.
Application Development
Developing on top of Google Cloud, and with Google Cloud tools, gets better every day.
35. The new Application Design Center, now in preview, provides a visual, canvas-style approach to designing and modifying application templates, and lets you configure application templates for deployment, view infrastructure as code in-line, and collaborate with teammates on designs.
36. The new Cloud Hub service, in preview, is the central command center for your entire application landscape, providing insights into deployments, health and troubleshooting, resource optimization, maintenance, quotas and reservations, and support cases. Try Cloud Hub here.
38. Application Monitoring, in public preview, supports automatically tagging telemetry (logs, metrics, and traces) with application context, application-aware alerts, and out-of-the-box application dashboards.
39. Cost Explorer, in private preview, provides visibility into granular application costs and utilization metrics, allowing you to identify efficiency opportunities; sign up here to try it out.
40. Gemini Code Assistagents can help with common developer tasks such as code migration, new feature implementation, code review, test generation, model testing, and documentation, and their progress can be tracked on the new Gemini Code Assist Kanban board.
41. Gemini Code Assist is now available in Android Studio for professional developers who want AI coding assistance with enterprise security and privacy features.
42. Gemini Code Assist tools, now in preview, helps you access information from Google apps and tools from partners including Atlassian, Sentry, Snyk, and more.
43. An App Prototyping agent in preview for Gemini Code Assist within the new Firebase Studio development environment turns your app ideas into fully functional prototypes, including the UI, backend code, and AI flows.
44. Gemini Cloud Assist is integrated with Application Design Center in preview to accelerate application infrastructure design and deployment.
45. Gemini Cloud Assist Investigations leverages data in your cloud environment to accelerate troubleshooting and issue resolution. Register for the private preview here.
46. Gemini Cloud Assist is now integrated across Google Cloud services including Storage Insights, Cloud Observability, Firebase, Database Center, Flow Analyzer, FinOps Hub, as well as security- and compliance-related services.
47. FinOps Hub 2.0 now includes waste insights and cost optimization opportunities from Gemini Cloud Assist.
48. The new Enterprise tier of the Google Developer Program is in limited preview, providing a safe and affordable way to explore Google Cloud and its AI products for a set monthly cost of $75/month per seat. Learn more here.
Compute
Whatever your workload, there’s a Compute Engine virtual machine to help you run it at the price, performance and reliability levels you need.
49. New C4D VMs built on AMD’s 5th Gen EPYC processors and paired with Google Titanium deliver impressive performance gains over prior generations— up to 30% vs C3D on the estimated SPECrate®2017_int_base benchmark. Currently in preview,try out C4D today.
50. C4 VMs built on the 6th generation Intel Granite Rapids CPUs feature the highest frequency of any Compute Engine VM — up to 4.2 GHz.
51. C4 shapes with Titanium Local SSD offer improved performance for I/O-intensive workloads like databases and caching layers, achieving Local SSD latency reductions of up to 35%.
52. C4 bare metal instances provide performance gains of up to 35% for general compute and up to 65% for ML recommendation workloads compared to the prior generation.
53. New, larger C4 VM shapes scale up to 288 vCPU, with 2.2TB of high-performing DDR5 memory and larger cache sizes. Request preview access here.
Compute Engine also features a variety of specialized VM families and unique capabilities:
54. New H4D VMs for demanding HPC workloads are built on the 5th gen AMD EPYC CPUs, and offer the highest whole-node VM performance of more than 12,000 flops, the highest per-core performance, and the best memory bandwidth of more than 950 GB/s of our VM families. Sign up for the H4D preview.
55. M4 VMs are certified for business-critical, in-memory SAP HANA workloads ranging from 744GB to 3TB, and for SAP NetWeaver Application Server, and offer up to 65% better price-performance and 2.25x more SAP Application Performance Standard (SAPS) compared to the previous memory-optimized M3.
56. The Z3 storage-optimized family now features new Titanium SSDs and offers nine new smaller shapes, ranging from 3TB to 18TB per instance. The Z3 family also introducing new storage-optimized bare-metal instance which include up to 72TB of Titanium SSDs and direct access to the physical server CPUs. Now in preview, register your interest here.
57. Nutanix Cloud Clusters (NC2) on Google Cloud let you run, manage, and operate apps, data, and AI across private and public clouds. Sign up for the public preview here.
58. Google Cloud VMware Engine now comes in 18 additional node shapes, bringing the total number of node shapes across VMware Engine v1 and v2 to 26.
59. Within the Titanium family, Titanium ML Adapter securely integrates NVIDIA ConnectX-7 network interface cards (NICs), providing 3.2 Tbps of non-blocking GPU-to-GPU bandwidth.
60. Titanium offload processors now integrate our GPU clusters with the Jupiter data center fabric, for greater cluster scale.
62. MIGs now support committed use discounts (CUDs) and reservation sharing with Vertex AI and Autopilot.
Containers & Kubernetes
The case for running on Google Kubernetes Engine (GKE) keeps on getting stronger, across an ever expanding class of workloads, most recently — AI.
63. GKE Inference Gatewayoffers intelligent scaling and load-balancing capabilities,helping you handle request scheduling and routing with gen AI model-aware scaling and load-balancing techniques.
64. With GKE Inference Quickstart, you can choose an AI model and your desired performance, and GKE configures the right infrastructure, accelerators, and Kubernetes resources to match.
66. Cluster Director for GKE (formerly Hypercompute Cluster) is now generally available, letting you deploy and manage large clusters of accelerated VMs with compute, storage, and networking — all operating as a single unit.
67. We announced performance improvements to GKE Autopilot, including faster pod scheduling, scaling reaction time, and capacity right-sizing.
68. Starting in Q3, Autopilot’s container-optimized compute platform will also be available to standard GKE clusters, without requiring a specific cluster configuration.
Customers
We shared hundreds of new customer stories across every industry and region, highlighting the ways they’re using Google Cloud to drive real impact. Here are some highlights:
69. Agoda, one of the world’s largest digital travel platforms, creates unique visuals and videos of travel destinations with Imagen and Veo on Vertex AI.
70. Bayer built an agent that uses predictive AI and advanced analytics to predict flu trends.
71. Bending Spoonsintegrated Imagen 3 into its Remini app to launch a popular new AI filter, processing an astounding 60 million photos per day.
72. BloombergConnects is using Gemini to explore new ways to help museums and other cultural institutions make their digital content accessible to more visitors.
73. Citi is using Vertex AI to rapidly deploy generative AI-powered productivity tools to more than 150,000 employees.
74. DBS, a leading Asian financial services group, is using Customer Engagement Suite to reduce customer call handling times by 20%.
75. Deutsche Bankbuilt DB Lumina, a new Gemini-powered tool that can synthesize financial data and research, turning, for example, a report that’s hundreds of pages into a one-page brief, delivering it in a matter of seconds to traders and wealth managers.
76. Deutsche Telekom has announced an expanded strategic partnership with Google Cloud, focusing on cloud and AI integration to modernize Deutsche Telekom’s IT, networks, and business applications, including migrating its SAP landscape.
77. Dun & Bradstreet is using Security Command Center to centralize monitoring of AI security threats.
78. Fanatics is partnering with Google Cloud to use AI technology to enhance every aspect of the fan journey. With Vertex AI Search for Commerce, Fanatics has developed an intelligent search ecosystem that understands and anticipates fan preferences, improves quality assurance and delivers intelligent customer service, and more.
79. Freshfieldsis using Gemini for Google Workspace and Google Cloud’s Vertex AI to enhance client services, including powering Freshfields’ Dynamic Due Diligence solution.
80. Globo, Latin America’s largest media company, used Vertex AI Search to create a recommendations experience inside its streaming platform that more than doubled their click-through-play rate on videos.
81. Gordon Food Services is simplifying insight discovery and recommending next steps with Agentspace.
82. The Home Depot built Magic Apron, an agent that offers expert guidance 24/7, providing detailed how-to instructions, product recommendations, and review summaries to make home improvement easier.
83. Honeywell has incorporated Gemini into its product development.
84. KPMG is building Google AI into in its newly formed KPMG Law firm and implementing Agentspace to enhance its own workplace operations.
85. L’Oreal is using Gemini, Imagen and Veo to accelerate creative ideation and production for marketing and product design, significantly speeding up workflows while maintaining ethical standards.
86. Lloyds Banking Group has taken a significant step in its strategic transformation by migrating its major platforms to Google Cloud. The transition is unlocking new opportunities to innovate with AI, enhancing the customer experience.
87. Lowe’sis revolutionizing product discovery with Vertex AI Search to generate dynamic product recommendations and address customers’ complex search queries.
89. Nokia built a coding tool to speed up app development with Gemini, enabling developers to create 5G applications faster.
90. Nuro, an autonomous driving company, uses vector search in AlloyDB to identify challenging scenarios on the road.
91. Mercado Libre deployed Vertex AI Search across 150M items in 3 pilot countries that is helping their 100M customers find the products they love faster, already delivering millions of dollars in incremental revenue.
92. Papa Johns is using AI to transform the ordering and delivery experience for its global customers. With Google Cloud’s AI, data analytics, and machine learning capabilities, Papa Johns can anticipate customer needs and personalize their pizza experience, as well as provide a consistent customer experience both inside the restaurants and online.
93. Redditis using Gemini on Vertex AI to power “Reddit Answers,” Reddit’s AI-powered conversation platform. Additionally, Reddit is using Enterprise Search to improve its homepage experience.
94. Samsung is integrating Gemini on Google Cloud into Ballie, its newest AI home companion robot, enabling more personalized and intelligent interactions for users.
95. Seattle Children’s hospitalis launching Pathway Assistant, a gen AI-powered agent with Gemini that improves clinicians’ access to complex information and the latest evidence-based best practices needed to treat patients.
96. Government of Singapore uses Google Cloud Web Risk to protect their residents online.
97. The Wizard of Oz at The Sphere is an immersive experience that reconceptualizes the 1939 film classic through the magic of AI, bringing it to life on a whole new scale for the colossal 160,000-square-foot domed screen at The Sphere in Las Vegas. It’s a collaboration between Sphere Entertainment, Google DeepMind, Google Cloud, Hollywood production company Magnopus, and five others.
98. Spotify uses BigQuery to harness enormous amounts of data to deliver personalized experiences to over 675 million users worldwide.
99. Intuitis using Google Cloud’s Document AI and Gemini models to simplify tax preparation for millions of TurboTax consumers this tax season, ultimately saving time and reducing errors.
100. United Wholesale Mortgage is using Google Cloud’s gen AI and data analytics to improve the mortgage process for 50,000 mortgage brokers and their clients, focusing on speed, efficiency, and personalized service.
101. Verizon is using Google Cloud’s Customer Engagement Suite to enhance its customer service for more than 115 million connections with AI-powered tools, like the Personal Research Assistant.
102. Vodafoneused Vertex AI along with open-source tools and Google Cloud’s security foundation to establish an AI security governance layer.
103. Wayfairupdates product attributes 5x faster with Vertex AI.
104. WPP built Open as a platform powered by Google models that all of its employees worldwide can use to concept, produce, and measure campaigns.
106. The next-generation of AlloyDB natural language lets you query structured data in AlloyDB securely and accurately, enabling natural language text modality in apps.
108. AlloyDB AI includes three new AI models: one that improves the relevance of vector search results using cross attention reranking; a multimodal embeddings model that supports text, images, and videos, and a new Gemini Embedding text model.
109. The new AlloyDB AI query engine lets developers use natural language expressions and constructs within SQL queries. Sign up for the preview of these AlloyDB features here.
111. Firestore with MongoDB compatibility, in preview, lets developers take advantage of MongoDB’s API portability along with Firestore’s multi-region replication with strong consistency, virtually unlimited scalability, a 99.999% SLA, and single-digit milliseconds read latency. Get started here today.
112. The new Oracle Base Database Service offers a flexible and controllable way to run Oracle Databases in the cloud.
113. Oracle Exadata X11M is now GA, bringing the Oracle Exadata platform to Google Cloud and adding additional enterprise-ready capabilities, including customer managed encryption keys (CMEK).
114. Database Migration Service (DMS) now supportsSQL Server to PostgreSQL migrationsfor Cloud SQL and AlloyDB, allowing you to fully execute on your database modernization strategy.
115. Cloud SQL and AlloyDB are available on C4A instances, our Arm-based Google Axion Processors delivering higher price-performace and throughput. Learn more here.
116. Database Center is now generally available and supports every database in our portfolio, providing a unified, AI-powered fleet management solution.
117. Spanner vector search is now generally available, designed to work with our SQL, Graph, Key-Value, and Full-Text Search modalities.
118. Graph Visualization for Spanner is now generally available, allowing users to visually explore valuable information from graph data.
120. Aiven for AlloyDB Omni, a fully-managed AlloyDB Omni service from our partner Aiven that runs on AWS, Azure, and Google Cloud, is now generally available.
122. New Cassandra-compatible APIs and live-migration tooling for zero-downtime migrations from Cassandra to Bigtable and Spanner.
123. Memorystore for Valkey is now generally available, with support for 7.2 and 8.0 engine versions.
124. Firebase Data Connect is now GA, offering the reliability of Cloud SQL for PostgreSQL with instant GraphQL APIs and type-safe SDKs
Data analytics
We announced several new innovations with our autonomous data to AI platform powered by BigQuery, alongside our unified, trusted, and conversational Looker BI platform:
127. BigQuery anomaly detection, now in preview, maintains data quality and automates metadata generation.
128.Data science agent, now GA, is embedded within Google’s Colab notebook, provides intelligent model selection, enabling scalable training, and faster iteration.
131. BigQuery knowledge engine, in preview, leverages Gemini to analyze schema relationships, table descriptions, and query histories to generate metadata on the fly, model data relationships, and recommend business glossary terms.
132. BigQuery semantic search, is now GA, providing AI-powered data insights and across BigQuery, grounding AI and agents in business context.
133. BigQuery’s contribution analysis feature, now GA, helps you pinpoint the key factors (or combinations of factors) responsible for the most significant changes in a metric.
135. BigQuery pipe syntax is GA, letting you apply operators in any order and as often as you need, and is compatible with most standard SQL operators.
Then, for data science and analyst teams, we added AI-driven data science and workflows as part of BigQuery notebook:
136. New intelligent SQL cells understand your data’s context and provide smart suggestions as you write code, and let you join data sources directly within your notebook.
137. Native exploratory analysis and visualization capabilities in BigQuery make it easy to explore data, as well as add features to enable easier collaboration with colleagues. Data scientists can also schedule analyses to run and refresh insights periodically.
138. The new BigQuery AI query engine lets data scientists process structured and unstructured data together with added real-world context, co-processing traditional SQL alongside Gemini to inject runtime access to real-world knowledge, linguistic understanding, and reasoning abilities.
139. Google Cloud for Apache Kafka, now GA, facilitates real-time data pipelines for event sourcing, model scoring, messaging and real-time analytics.
141. New dataset-level insights in BigQuery data canvas, in preview, surface hidden relationships between tables and generate cross-table queries by integrating query usage analysis and metadata.
142. BigQuery ML includes the new AI.GENERATE_TABLE in preview to capture the output of LLM inference within SQL clauses.
144. BigQuery vector search includes a new index type, now GA, based on Google’s ScaNN model that’s coupled with a CPU-optimized distance computation algorithm for scalable, faster and more cost-efficient processing.
145. The preview of BigQuery ML’s pre-trained TimesFM model developed by Google Research simplifies time-series forecasting.
146. We integrated new Google Maps Platform datasets directly into BigQuery, to make it easier for data analysts and decision makers to access insights.
147. In addition, Earth Engine in BigQuery brings the best of Earth Engine’s geospatial raster data analytics directly into BigQuery. Learn more here.
148. GrowthLoopintroduced its Compound Marketing Engine built on BigQuery with Growth Agents powered by Gemini, so marketing can build personalized audiences and journeys that drive rapidly compounding growth.
149. Informaticaexpanded its services on Google Cloud to enable sophisticated analytical and AI governance use cases.
150. Fivetranintroduced its Managed Data Lake Service for Cloud Storage with native integration with BigQuery metastore and automatic data conversion to open table formats like Apache Iceberg and Delta Lake
151. DBTis now integrated with BigQuery DataFrames and DBT Cloud is now on Google Cloud.
152. Datadogintroduced expanded monitoring capabilities for BigQuery, providing granular visibility into query performance, usage attribution, and data quality metrics.
BigQuery’s autonomous data foundation provides governance, orchestration for diverse data workloads, and a commitment to flexibility via open formats. Announcements in this area include:
153. BigQuery makes unstructured data a first-class citizen with multimodal tables in preview, bringing rich, complex data types alongside structured data for unified storage and querying via the new ObjectRef data type.
154. BigQuery governance in previewprovides a single, unified view for data stewards and professionals to handle discovery, classification, curation, quality, usage, and sharing.
156. BigQuery metastore, now GA, enable engine interoperability across BigQuery, Apache Spark, and Apache Flink engines, with support for the Iceberg Catalog.
157. BigQuery business glossary, now GA, lets you define and administer company terms, identify data stewards for these terms, and attach them to data asset fields.
158. BigQuery continuous queries, now GA,enable instant analysis and actions on streaming data using SQL, regardless of its original format.
159. BigQuery tables for Apache Iceberg in preview, lets you connect your Iceberg data to SQL, Spark, AI and third-party engines.
160. New advanced workload management capabilities, now GA,scale resources, manage workloads, and help ensure their cost-effectiveness.
161. BigQuery spend commit, now GA,simplifies purchasing, unifying spend across BigQuery data processing engines, streaming, governance, and more.
162. BigQuery DataFrames now has AI code assist capabilities in preview, letting you use natural language prompts to generate or suggest code in SQL or Python, or to explain an existing SQL query.
163. SQL translation assistance, now GA, is an AI-based translator that lets you create Gemini-enhanced rules to customize your SQL translations, to accelerate BigQuery migrations.
164. Catalog metadata export, GA, enables bulk extract of catalog entries into Cloud Storage.
165. BigQuery can now perform automatic at-scale cataloging of BigLake and object tables, now GA.
166. BigQuery managed disaster recovery is now GA, featuring automatic failover coordination, continuous near-real-time data replication to a secondary region, and fast, transparent recovery during outages.
167. Newworkload management capabilities in preview include reservation-level fair sharing of slots, predictability in performance of reservations, and enhanced observability through reservation attribution in billing.
Looker, is adding a host of new conversational and visual capabilities, aimed at making BI accessible and useful to all users, accelerated by AI.
168. Gemini in Looker features are now available to all Looker platform users, including Conversational Analytics, Visualization Assistant, Formula Assistant, Automated Slide Generation, and LookML Code Assistant.
169. Code Interpreter for Conversational Analytics is in preview, allowing business users to perform forecasting and anomaly detection using natural language without needing deep Python expertise. Learn more and sign up for it here.
170. New Looker reports feature an intuitive drag-and-drop interface, granular design controls, a rich library of visualizations and templates, and real-time collaboration capabilities, now in the core Looker platform.
171. With Google Cloud’s acquisition of Spectacles.dev, developers can automate testing and validation of SQL and LookML changes using CI/CD practices.
Firebase
172. The new Firebase Studio, available to everyone in preview, is a cloud-based, agentic development environment powered by Gemini that includes everything developers need to create and publish production-quality full-stack AI apps quickly, all in one place. Gemini Code Assist agents are available via private preview.
173. Genkit, an open-souce framework for building AI-powered applications, using your preferred language, now has early support for Python and expanded support for Go. Try this template in Firebase Studio to build with Genkit.
174. Vertex AI in Firebase now includes support for the Live API for Gemini models, enabling more conversational interactions in apps such as allowing customers to ask audio questions and get responses.
175. Firebase Data Connectis now GA,offering the reliability of Cloud SQL for PostgreSQL with instant GraphQL APIs and type-safe SDKs.
176. Firebase App Hosting is also GA, providing an opinionated, git-centric hosting solution for modern, full-stack web apps.
177. A new App Testing agent within Firebase App Distribution, also in preview, prepares mobile apps for production by generating, managing, and executing end-to-end tests.
Google Cloud Consulting
Google Cloud Consulting introduced several new pre-packaged service offerings:
178. Agentspace Accelerator provides a structured approach to connecting and deploying AI-powered search within organizations, so employees can easily gain access to relevant internal information and resources when they need it.
180. Oracle on Google Cloud lets customers combine Oracle databases and applications with Google Cloud’s advanced platform and AI capabilities for enhanced database and network performance.
181. We expanded access to Delivery Navigator, a series ofproven delivery methodologies and best practices to help with migrations and technology implementations to customers as well as partners, in preview.
182. Cloud WAN, a Cross-Cloud Network solution, is a fully managed, reliable, and secure enterprise backbone that makes Google’s global private network available to all Google Cloud customers. Cloud WAN delivers up to 40% improved network performance, while reducing total cost of ownership by up to 40%. Read more here.
183. The new 400G Cloud Interconnect and Cross-Cloud Interconnect, available later this year, offers up to 4x more bandwidth than our 100G Cloud Interconnect and Cross-Cloud Interconnect, providing connectivity from on-premises or other cloud environments to Google Cloud.
184. Build massive AI services with networking support for up to 30,000 GPUs per cluster in a non-blocking configuration, available in preview now.
185. Zero-Trust RDMA security helps you secure your high-performance GPU and TPU traffic with our RDMA firewall, featuring dynamic enforcement policies. Available later this year.
186. Get accelerated GPU-to-GPU communication, with up to 3.2Tbps of non-blocking GPU-to-GPU bandwidth with our high-throughput, low-latency RDMA networking, now generally available.
188. Cloud Load Balancing has optimizations for LLM inference,letting you leverage NVIDIA GPU capacity across multiple cloud providers or on-prem infrastructure.
189. New Service Extensions plugins, powered by WebAssembly (Wasm), let you automate, extend, and customize your applications with plugin examples in Rust, C++, and Go. Support for Cloud Load Balancing is now generally available, and Cloud CDN support will follow later this year.
190. Cloud CDN‘s fast cache invalidation delivers static and dynamic content at global scale with improved performance, now in preview.
191. TLS 1.3 0-RTT in Cloud CDN boosts application performance for resumed connections, now in preview.
192. App Hub provides streamlined service discovery and management by automating service discovery and cataloging.
193. App Hub service health enables resilient global services with network-driven cross-regional failover. Available later this year.
194. Later in 2025, you’ll be able to use Private Service Connect to publish multiple services within a single GKE cluster, making them natively accessible from non-peered GKE clusters, Cloud Run, or Service Mesh.
Then, to help you secure your workloads, we introduced enhancements to protect distributed applications and internet-facing services against network attacks:
195. The new DNS Armor detects DNS-based data exfiltration attacks performed using DNS tunneling, domain generation algorithms (DGA) and other sophisticated techniques. Available in preview later this year.
196. New hierarchical policies for Cloud Armor let you enforce granular protection of your network architecture.
197. There are new network types and firewall tags for Cloud NGFW hierarchical firewall policies, coming this quarter in preview.
198. Cloud NGFW adds new layer 7 domain filtering, allowing firewall administrators to monitor and control outbound web traffic to only allowed destinations. Coming later in 2025.
199. Inline network DLP for Secure Web Proxy and Application Load Balancer provides real-time protection for sensitive data-in-transitvia integration with third-party (Symantec DLP) solutions using Service Extensions. In preview this quarter.
200. Network Security Integration, now generally available, helps you maintain consistent policies across hybrid and multi-cloud environments without changing your routing policies or network architecture.
We’ve always taken an open approach to AI, and the same is true for agentic AI. With updates this week at Next ‘25, we’re now infusing partners at every layer of our agentic AI stack to enable multi-agent ecosystems. Here’s a closer look:
202. Expert AI services: Our ecosystem of services partners — including Accenture, BCG, Capgemini, Cognizant, Deloitte, HCLTech, Infosys, KPMG, McKinsey, PwC, TCS, and Wipro — have actively contributed to the A2A protocol and will support its implementation.
203. AI Agent Marketplace: We launched a new AI Agent Marketplace — a dedicated sectionwithin Google Cloud Marketplace that allows customers to browse, purchase, and manage AI agents built by partners including Accenture, BigCommerce, Deloitte Elastic, UiPath, Typeface, and VMware, with more launching soon.
204. Power agents with all your enterprise data: We are partnering with NetApp, Oracle, SAP, Salesforce, and ServiceNow to allow agents to access data stored in these popular platforms.
205. Better field alignment and co-sell: We introduced new processes to better capture and share partners’ critical contributions with our sales team, including increased visibility into co-selling activities like workshops, assessments, and proofs-of-concept, as well as partner-delivered services.
206. More partner earnings: We are evolving incentives to help partners capitalize on the biggest opportunities, such as a 2x increase in partner funding for AI opportunities over the past year. We also introduced new AI-powered capabilities inEarnings Hub, our destination for tracking incentives and growth.
207. We partnered with Adobe, the leader in creativity, to bring our advanced Imagen 3 and Veo 2 models to applications like Adobe Express.
208. Together with Salesforce’s Agentforce, we’re leading the digital labor revolution, driving massive gains in human augmentation, productivity, efficiency, and customer success.
Security
We offer critical cyber defense capabilities for today’s challenging threat environment, and introduced a number of new innovations:
209. Google Unified Security: This solution brings together our visibility, threat detection, AI powered security operations, continuous virtual red-teaming, the most trusted enterprise browser, and Mandiant expertise — in one converged security solution running on a planet-scale data fabric.
210. Alert triage agent: This agent performs dynamic investigations on behalf of users. It analyzes the context of each alert, gathers relevant information, and renders a verdict on the alert, along with a history of the agent’s evidence and decision making.
211. Malware analysis agent: This agent investigates whether code is safe or harmful. It builds onCode Insight to analyze potentially malicious code, including the ability to create and execute scripts for deobfuscation.
212. In Google Security Operations, new data pipeline management capabilities can help customers better manage scale, reduce costs, and satisfy compliance mandates.
213. We also expanded our Risk Protection Program, which provides discounted cyber-insurance coverage based on cloud security posture, to welcome new program partners Beazley and Chubb, two of the world’s largest cyber-insurers.
214. New employee phishing protections in Chrome Enterprise Premium use Google Safe Browsing data to help protect employees against lookalike sites and portals attempting to capture credentials.
215. TheMandiant Retainer provides on-demand access to Mandiant experts. Customers now can redeem prepaid funds for investigations, education, and intelligence to boost their expertise and resilience.
216. Mandiant Consulting is also partnering with Rubrik and Cohesity to create a solution to minimize downtime and recovery costs after a cyberattack.
Storage
Storage is a critical component for minimizing bottlenecks in both training and inference, and we introduced new innovations to help:
217. We expanded Hyperdisk Storage Pools to store up to 5 PiB of data in a single pool — a 5x increase from before.
218. Hyperdisk Exapools is the biggest and fastest block storage in any public cloud, with exabytes of storage delivering terabytes per second of performance.
219. Hyperdisk ML can now hydrate from Cloud Storage using GKE volume populator.
220. Rapid Storage is a new Cloud Storage zonal bucket with <1ms random read and write latency, and compared to other leading hyperscalers, 20x faster data access, 6 TB/s of throughput, and 5x lower latency for random reads and writes.
221. Anywhere Cacheis a new strongly consistent cache that works seamlessly with existing regional buckets to cache data within a selected zone. Reduces latency up to 70% and 2.5TB/s accelerating AI workloads; maximizing goodput by keeping data close to GPU/TPUs.
222. The new Google Cloud Managed Lustre high-performance, fully managed parallel file system built on DDN EXAScaler. This zonal storage solution provides PB scale <1ms latency, millions of IOPS, and TB/s of throughput for AI workloads.
223. Storage Intelligence, the industry’s first offering enabling customers to generate storage insights specific to their environment by querying object metadata at scale, uses LLMs to provide insights into data estates, as well as take actions on them.
Startups
224. We announced a significant new partnership with the leading venture capital firm Lightspeed, which will make it easier for Lightspeed-backed startups to access technology and resources through the Google for Startups Cloud Program. This includes upwards of $150,000 in cloud credits for Lightspeed’s AI portfolio companies, on top of existing credits available to all qualified startups through the Google for Startups Cloud Program.
225. The new Startup Perks program provides early stage startups with preferred access to solutions from our partners like Datadog, Elastic, ElevenLabs, GitLab, MongoDB, NVIDIA, Weights & Biases, and more.
226. Google for Startups Cloud Program members will receive an additional $10,000 in credits to use exclusively on Partner Models through Vertex AI Model Garden, so they can quickly start using both Gemini models and models from partners like Anthropic and Meta.
Google Workspace: AI-powered productivity
Gemini not only powers best-in-class AI capabilities as a model, but through its own products, like Google Workspace, which includes popular apps like Gmail, Docs, Drive and Meet. We announced a number of new Workspace innovations to further empower users with AI, including:
227. Help me Analyze: This powerful feature transforms Google Sheets into your personal business analyst, intelligently identifying insights from your data without the need for explicit prompting, empowering you to make data-driven decisions with ease.
228. Docs Audio Overview: With audio overviews in Docs, you can create high-quality, human-like audio read-outs or podcast-style summaries of your documents.
229. Google Workspace Flows: Workspace Flows helps you automate daily work and repetitive tasks like managing approvals, researching customers, organizing your email, summarizing your daily agenda, and much more.
There’s no place like home
And with that, we’ve come to the end of Next 25. We hope you’ve enjoyed your time in Las Vegas, and wish you safe travels.
See you in Vegas next year for Google Cloud Next: April 22 – 24, 2026.
aside_block
<ListValue: [StructValue([(‘title’, ‘Turn your new insights from Google Cloud Next into action’), (‘body’, <wagtail.rich_text.RichText object at 0x3e62dd34bfa0>), (‘btn_text’, ”), (‘href’, ”), (‘image’, <GAEImage: next 25>)])]>
1. Grounding with Google Maps is currently available as an experimental release in the United States, providing access to only places data in the United States.
We are excited to announce that AWS Mainframe Modernization service is now available with greater control of managed runtime environments that run modernized mainframe applications.
For both refactored and replatformed applications, you can now export data sets to an Amazon S3 bucket. Optionally, you can choose to encrypt the exported data set. This export feature makes it easier to move data set across environments, or to archive data sets.
For applications refactored with AWS Blu Age, you can now restart a batch job at a specific step. This enables advanced batch operational and recovery procedures.
For applications replatformed with Rocket Software, you can now configure your managed runtime application using a base configuration compatible with Rocket Enterprise Server deployed on non-managed environments. This base configuration provides flexibility by allowing numerous advanced configuration parameters supported by Rocket Enterprise Server, such as CICS or IMS granular parameters. It also allows transferring exported configuration parameters from a Rocket Enterprise Server deployed on Amazon EC2 to an AWS Mainframe Modernization managed runtime application.
These new features are available in any AWS Region where AWS Mainframe Modernization managed runtime is already deployed. To learn more, please visit AWS Mainframe Modernization product and documentation pages.
Starting today, Amazon Elastic Compute Cloud (EC2) I4g storage-optimized instances powered by AWS Graviton2 processors and 2nd generation AWS Nitro SSDs are now available in the South America (Sao Paulo) region.
I4g instances are optimized for workloads performing a high mix of random read/write operations and requiring very low I/O latency and high compute performance, such as transactional databases (MySQL, and PostgreSQL), real-time databases including in-memory databases, NoSQL databases, time-series databases (Clickhouse, Apache Druid, MongoDB) and real-time analytics such as Apache Spark.
Amazon Redshift Concurrency Scaling is now available in Israel (Tel Aviv) and Canada West (Calgary) regions.
Amazon Redshift Concurrency Scaling elastically scales query processing power to provide consistently fast performance for hundreds of concurrent queries. Concurrency Scaling resources are added to your Redshift cluster transparently in seconds, as concurrency increases, to process queries without wait time. Amazon Redshift customers with an active Redshift cluster earn up to one hour of free Concurrency Scaling credits, which is sufficient for the concurrency needs of most customers. Concurrency scaling allows you to specify usage control providing customers with predictability in their month-to-month cost, even during periods of fluctuating analytical demand.
To enable Concurrency Scaling, set the Concurrency Scaling Mode to Auto in your Amazon Web Services Management Console. You can allocate Concurrency Scaling usage to specific user groups and workloads, control the number of Concurrency Scaling clusters that can be used, and monitor Cloudwatch performance and usage metrics.
To learn more about concurrency scaling including regional-availability, see our documentation and pricing page.
Starting today, we are making it easier for customers to understand their inter-availability zone (AZ) VPC Peering usage within the same AWS Region by introducing a new usage type in their bill. These changes won’t affect customers’ charges and will help them easily understand their VPC Peering costs, enabling them to choose the right architecture based on cost, performance, and ease of management.
VPC Peering is an Amazon VPC feature that allows customers to establish networking connection between two VPCs, helping them route traffic between two VPCs using private IPv4 or IPv6 addresses. Previously, VPC Peering usage was reported under the intra-regional Data Transfer usage, making it difficult for customers to understand their VPC Peering usage and charges. With this launch, customers can now view their VPC Peering usage using the new usage type “Region_Name-VpcPeering-In/Out-Bytes” in Cost Explorer or Cost and Usage Report. Customers do not need to make any changes to their existing VPC Peering connections to benefit from this change, as these changes will be automatically applied.
There are no changes to the pricing for data transferred over VPC Peering connections. These changes will apply to all AWS commercial and the AWS Gov Cloud (US) Regions.
Attending a tech conference like Google Cloud Next can feel like drinking from a firehose — all the news, all the sessions, and breakouts, all the learning and networking… But after a busy couple of days, watching the developer keynote makes it seem like there’s a method to the madness. A coherent picture starts to emerge from all the things that you’ve seen, pointing the way to all the cool things you can do when you get back to your desk.
This year, the developer keynote was hosted by the inimitable duo of Richard Seroter, Google Cloud Chief Evangelist, and Stephanie Wong, Head of Developer Skills and Community, plus a whole host of experts from around Google Cloud product, engineering, and developer advocacy teams. The keynote itself was organized around a noble, relatable goal: Use AI to help remodel AI Developer Experience Engineer Paige Bailey’s 1970s era kitchen. But how?
It all starts with a prompt
The generative AI experience starts by prompting a model with data and your intent. Paige was joined on stage by Logan Kilpatrick, Senior Product Manager at Google DeepMind. There, Logan and Paige prompted AI Studio to analyze Paige’s kitchen, supplying it with text descriptions, floor plans, and images. In return, it suggested cabinets, a cohesive design, color palette, and materials, relying on Gemini’s native image generation capabilities to bring its ideas to life. Then, to answer important questions on cost, especially for Paige’s area, they used Grounding with Google Search to pull in real-world material costs, local building codes and regulations, and other relevant information.
As Logan said, “From understanding videos, to native image generation, to grounding real information with Google Search – these are things that can only be built with Gemini.”
Gemini 2.5 Flash — our workhorse model optimized specifically for low latency and cost efficiency — is coming soon to Vertex AI, AI Studio, and the Gemini app.
From prompt to agent
We all know that a prompt is the heart of a generative AI query. “But what the heck is an agent?” asked Richard. “That’s the million-dollar question.”
“An agent is a service that talks to an AI model to perform a goal-based operation using the tools and context it has,” Stephanie explained. And how do you go from prompt to agent? One way is to use Vertex AI, our comprehensive platform for building and managing AI applications and agents, and Agent Development Kit (ADK), an open-source framework for designing agents.ADK makes it easier than ever to get started with agents powered by Gemini models and Google AI tools.
Dr. Fran Hinkelman, Developer Relations Engineering Manager at Google Cloud, took the stage to show off ADK. An agent needs three things, Fran explained: 1) instructions to define your agent’s goal, 2) tools to enable them to perform, and 3) a model to handle the LLM’s tasks.
Fran wrote the agent code using Python, and in a matter of minutes, deployed it, and got a professionally laid out PDF that outlined everything a builder might need to get started on a kitchen remodel. “What a massive time-saver,” Fran said.
New things that make this possible:
Agent Development Kit (ADK)is our new open-source framework that simplifies the process of building agents and sophisticated multi-agent systems while maintaining precise control over agent behavior. With ADK, you can build an AI agent in under 100 lines of intuitive code.
ADK support for Model Context Protocol (MCP), which creates a standardized structure and format for all the information an LLM needs to process a data request.
From one agent to many
It’s one thing to build an agent. It’s another to orchestrate a collection of agents — exactly the kind of thing you need for a complex process like remodeling a kitchen. To show you how, Dr. Abirami Sukumaran, Staff Developer Advocate at Google Cloud, used ADK to create a multi-agent ecosystem with three types of agents: 1) a construction proposal agent 2) a permits and compliance agent 3) an agent for ordering and delivering materials.
And when the multi-agent system was ready, she deployed it directly from ADK to Vertex AI Agent Engine, a fully managed agent runtime that supports many agent frameworks including ADK.
It gets better: After deploying her agent, Abirami tested it out in Google Agentspace, a hub for sharing your own agents and those from third-parties.
There was a problem, though. Midway through, the agent system appeared to fail. Abirami sprung into action, launching Gemini Cloud Assist Investigations, which used Logs Explorer to return relevant observations and hypotheses about the source of the problem. It even supplied a recommended code fix for the agents. Abirami examined the code, accepted it, redeployed her agents, and saved the day.
This is really key. “It’s hard enough to build systems that orchestrate complex agents and services,” Abirami said. “Developers shouldn’t have to sit around debugging multiple dependencies — getting to the logs, going through the code, all of this can take a lot of time and resources that devs typically don’t have.”
New things that make this possible:
Vertex AI Agent Engine is a fully managed runtime in Vertex AI that helps you deploy your custom agents to production with built-in testing, release, and reliability at a global, secure scale.
Cloud Assist Investigations helps diagnose problems with infrastructure and even issues in the code.
Agent2Agent (A2A) protocol: We’re proud to be the first hyperscaler to create an open protocol to help enterprises support multi-agent ecosystems, so agents can communicate with each other, regardless of the underlying technology.
Choose your own IDE and models
“Have you heard of vibe coding?” i.e., agentic coding, asked our next presenter, Debi Cabrera, Senior Developer Advocate at Google Cloud. Essentially, people can prompt an agent with ideas as well as code to get to an effective programming output. People are doing it more and more using Windsurf, a popular new Integrated Development Environment (IDE), and she’s a fan.
Debi also showed using Gemini in Cursor and IntelliJ with Copilot, but you could also use Visual Studio Code, Tabnine, Cognition, or Aider. (She even wrote her prompts in Spanish, which Gemini handled sin problema). At the end of the day, “we’re enabling devs to use Gemini wherever it suits you best,” Debi said.
Conversely, if you don’t want to use Gemini as your model, you can also use one of the more than 200 models in Vertex AI Model Garden, including Llama, Gemma 3, Anthropic, and Mistral, or open source models from Hugging Face.
“No matter what you use, we’re excited to see what you come up with!”
Android Studiosupport for Gemini Code Assist is now available in preview.
Gemini in Firebase provides complete AI assistance in the new Firebase Studio.
In a field of dreams
Next up, presenters took a break from Paige’s kitchen remodel to tackle another high-value problem: how to throw a pitch.
With all the data that Major League Baseball processes with Google Cloud — 25 million data points per game — pitching technique is a problem that’s ripe for AI.
Jake DiBattista, winner of the recent Google Cloud x MLB Hackathon, started by analyzing a video of a great left-handed pitcher, Clayton Kershaw. He pre-processed the video using a computer vision library, and stored it in Google Cloud, using selections such as pitch type and game state to pull MLB data. Finally, after sending all this information to the Gemini API, he got his answer: Kershaw threw his signature curveball with nearly no deviation from his ideal.
Impressive, but how well does it work for those of us who aren’t pros? Jake created an “amateur mode” for less experienced players, and used a video of our host, Richard, throwing a pitch! After some prompt engineering to adapt from the professional model for Kershaw to an amateur model for Richard, the results were a little more prescriptive: He has potential, he just needs to tighten up his arm a little, and use more leg drive to maximize his power.
Jake shared the inspiration for his project: As a shot putter in college, he wanted to measure the accuracy of his throwing technique. How can you improve if you don’t know what you’re doing wrong – or right? Back then, having this kind of data would have been incredibly valuable for his development.
But what’s truly amazing is that Jake built this fully customizable prompt generator for analyzing pitches in just one week. “This essentially worked out of the box,” Jake said. “I didn’t need to implement a custom model or build overly complex datasets.”
Get back to work
Meanwhile, back at his day job, our next presenter Jeff Nelson, Developer Advocate at Google Cloud, took the stage with a clear goal: to turn raw data into a data application for use by sales managers. He started in BigQuery Notebook to build a forecast and wrote some SQL code. BigQuery loaded the results into a Python DataFrame, because Python makes it easy to use libraries to execute code over tables of any size.
But how can you actually use this agent to forecast sales? Jeff selected the Gemini Data Science Agent built into the Notebook, hit “Ask Agent,” and inputted a prompt that asked for a sales forecast from his table. The best part – from that point onward, all code was generated and executed by the Gemini Data Science Agent.
Plus, he pointed out that the agent used Spark for feature engineering, which is only possible because of our new Serverless Spark engine in BigQuery. Switching between SQL, Spark, and Python is easy, so you can use the right tool for the job.
To build the forecast itself, Jeff used a new Google foundation model, TimesFM, that’s accessible directly from BigQuery.Unlike traditional models, this one’s pre-trained and on massive times-series datasets, so you get forecasts by simply inputting data. “The forecast becomes a data app accessible to everyone,” Jeff said.
As a developer, how would you like it if you could hand off boring things like creating technical design or product requirement docs? Scott Densmore, Senior Director of Engineering, closed out the demos to show us an incredible way to cut through tedious work: Gemini Code Assist and its new Kanban board.
Code Assist can help you orchestrate agents in all aspects of the software development lifecycle, including with what Scott calls a “backpack” that holds all your engineering context. Using a technical design doc for a Java migration as an example, Scott created a comment and assigned it to Code Assist right from the Google doc. Instantly, the new task shows up on the Kanban board, ready to be tracked. Nor is this capability limited to Google Docs — you can also assign tasks directly from your chatrooms and bug trackers, or have Code Assist proactively find them for you.
Then, he took a tougher example: he asked Code Assist to create a prototype for a product requirement doc. He told Code Assist the changes he wanted, and hit repeat until he was happy with what he saw. Easy.
“Gemini Code Assist provides an extra pair of coding hands to help you create applications and remove repetitive and mundane tasks — so you can focus on the fun stuff.”
New things that make this possible:
Gemini Code Assist Kanban boardlets you interact with our agents, review the workplan that Gemini creates to complete the tasks, and track the progress of the various jobs/requests.
Pretty amazing, right? But don’t just take our word for it, for a true sense of all the magic that we demonstrated here, go ahead and rewatch the full developer keynote. We promise that it will be an hour well spent.
Today, AWS announces that AWS Control Tower supports an additional 223 managed Config rules in Control Catalog for various use cases such as security, cost, durability, and operations. With this launch, you can now search, discover, enable and manage these additional rules directly from AWS Control Tower and govern more use cases for your multi-account environment.
To get started, in AWS Control Tower go to the Control Catalog and search for controls with the implementation filter AWS Config, you will then see all the AWS Config rules present in the Catalog. If you find rules that are relevant for you, you can then directly enable them from the AWS Control Tower console. You can also use ListControls, GetControl and EnableControl APIs. With this launch we’ve updated ListControls and GetControl APIs to support three new fields: Create Time, Severity and Implementation, that you can use when searching for a control in Control Catalog. For example, you can now programmatically find high severity Config rules which were created after your previous evaluation.
You can search the new AWS Config rules in all AWS Regions where AWS Control Tower is available, including AWS GovCloud (US). When you want to deploy a rule, reference the list of supported regions for that rule to see where it can be enabled. To learn more, visit the AWS Control Tower User Guide.
AWS announces the end of sale for AWS Elemental Link HD devices effective April 15, 2024. AWS Elemental Link UHD devices will continue to be available for purchase. To support HD content contribution workflows, Link UHD has now added HD ingest pricing, providing a seamless path for new deployments. Existing Link HD devices will continue to be supported, with Link UHD now serving as the recommended solution for both HD and UHD contribution workflows.
To enable HD pricing on Link UHD devices, you can configure the device’s input resolution on the Link device configuration page when the device is not actively streaming. The configuration option provides the flexibility to optimize costs when contributing HD content through Link UHD devices.
This feature is available immediately in all AWS Regions where Link UHD is supported. The input resolution configuration option is accessible through the AWS Management Console for all Link UHD devices.
Gateway Load Balancer (GWLB) now supports Load Balancer Capacity Unit (LCU) Reservation that allows you to proactively set a minimum bandwidth capacity for your load balancer, complementing its existing ability to auto-scale based on your traffic pattern.
Gateway Load Balancer helps you deploy, scale, and manage third-party virtual appliances. With this feature, you can reserve a guaranteed capacity for anticipated traffic surge. The LCU reservation is ideal for scenarios such as onboarding and migrating new workload to your GWLB gated services without the need to wait for organic scaling, or maintaining a minimum bandwidth capacity for your firewall applications to meet specific SLA or compliance requirements. When using this feature, you pay only for the reserved LCUs and any additional usage above the reservation. You can easily configure this feature through the ELB console or API.
The feature is available for GWLB in US East (N. Virginia), US East (Ohio), US West (Oregon), Asia Pacific (Hong Kong), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm) AWS Regions. This feature is not supported on Gateway Load Balancer Endpoint (GWLBe). To learn more, please refer to the GWLB documentation.
Starting today, Amazon S3 Express One Zone has reduced pricing for storage by 31%, PUT requests by 55%, and GET requests by 85%. In addition, S3 Express One Zone has reduced its per-gigabyte data upload and retrieval charges by 60% and now applies these charges to all bytes rather than just portions of requests exceeding 512 kilobytes.
Amazon S3 Express One Zone is a high-performance, single-Availability Zone storage class purpose-built to deliver consistent single-digit millisecond data access for your most frequently accessed data and latency-sensitive applications, such as machine learning training, analytics for live streaming events, and market analysis for financial services.
Amazon Bedrock Knowledge Bases now extends support for hybrid search to knowledge bases created using Amazon Aurora PostgreSQL and MongoDB Atlas vector stores. This capability, which can improve relevance of the results, previously only worked with Opensearch Serverless and Opensearch Managed Clusters in Bedrock Knowledge Bases.
Retrieval augmented generation (RAG) applications use semantic search, based on vectors, to search unstructured text. These vectors are created using foundation models to capture contextual and linguistic meaning within data to answer human-like questions. Hybrid search merges semantic and full-text search methods, executing dual queries and combining results. This approach improves results relevance by retrieving documents that match conceptually from semantic search or that contain specific keywords found in full-text search. The wider search scope enhances result quality, particularly for keyword-based queries.
You can enable hybrid search through the Knowledge Base APIs or through the Bedrock console. In the console, you can select hybrid search as your preferred search option within Knowledge Bases, or choose the default search option to use semantic search only. Hybrid search with Aurora PostgreSQL is available in all AWS Regions where Bedrock Knowledge Bases is available, excluding Europe (Zurich) and GovCloud (US) Regions. Hybrid search with Mongo DB Atlas is available in the US West (Oregon) and US East (N. Virginia) AWS Regions. To learn more, refer to Bedrock Knowledge Bases documentation. To get started, visit the Amazon Bedrock console.