Customers can now create Amazon FSx for Lustre file systems in the AWS Asia Pacific (Thailand) Region, providing fully managed shared storage with the scalability and performance of the popular Lustre file system.
Amazon FSx makes it easier and more cost effective to launch, run, and scale feature-rich, high-performance file systems in the cloud. It supports a wide range of workloads with its reliability, security, scalability, and broad set of capabilities. Amazon FSx for Lustre provides fully managed shared storage built on the world’s most popular high-performance file system, designed for fast processing of workloads such as machine learning, high performance computing (HPC), video processing, financial modeling, and electronic design automation (EDA).
To learn more about Amazon FSx for Lustre, visit our product page, and see the AWS Region Table for complete regional availability information.
From retail to gaming, from code generation to customer care, an increasing number of organizations are running LLM-based applications, with 78% of organizations in development or production today. As the number of generative AI applications and volume of users scale, the need for performant, scalable, and easy to use inference technologies is critical. At Google Cloud, we’re paving the way for this next phase of AI’s rapid evolution with our AI Hypercomputer.
At Google Cloud Next 25, we shared many updates to AI Hypercomputer’s inference capabilities, unveiling Ironwood, our newest Tensor Processing Unit (TPU) designed specifically for inference, coupled with software enhancements such as simple and performant inference using vLLM on TPU and the latest GKE inference capabilities — GKE Inference Gateway and GKE Inference Quickstart.
With AI Hypercomputer, we also continue to push the envelope for performance with optimized software, backed by strong benchmarks:
Google’s JetStream inference engine incorporates new performance optimizations, integrating Pathways for ultra-low latency multi-host, disaggregated serving.
MaxDiffusion, our reference implementation of latent diffusion models, delivers standout performance on TPUs for compute-heavy image generation workloads, and now supports Flux, one of the largest text-to-image generation models to date.
The latest performance results from MLPerf™ Inference v5.0 demonstrate the power and versatility of Google Cloud’s A3 Ultra (NVIDIA H200) and A4 (NVIDIA HGX B200) VMs for inference.
Optimizing performance for JetStream: Google’s JAX inference engine
To maximize performance and reduce inference costs, we are excited to offer more choice when serving LLMs on TPU, further enhancing JetStream and bringing vLLM support for TPU, a widely-adopted fast and efficient library for serving LLMs. With both vLLM on TPU and JetStream, we deliver standout price-performance with low-latency, high-throughput inference and community support through open-source contributions and from Google AI experts.
JetStream is Google’s open-source, throughput- and memory-optimized inference engine, purpose-built for TPUs and based on the same inference stack used to serve Gemini models. Since we announced JetStream last April, we have invested significantly in further improving its performance across a wide range of open models. When using JetStream, our sixth-generation Trillium TPU now exceeds throughput performance by 2.9x for Llama2-70B and 2.8x for Mixtral 8x7B compared to TPU v5e (using our reference implementation MaxText).
Figure 1: JetStream throughput (output tokens / second). Google internal data. Measured using Llama2-70B (MaxText) on Cloud TPU v5e-8 and Trillium 8-chips and Mixtral 8x7B (MaxText) on Cloud TPU v5e-4 and Trillium 4-chips. Maximum input length: 1024, maximum output length: 1024. As of April 2025.
Available for the first time for Google Cloud customers, Google’s Pathways runtime is now integrated into JetStream, enabling multi-host inference and disaggregated serving — two important features as model sizes grow exponentially and generative AI demands evolve.
Multi-host inference using Pathways distributes the model across multiple accelerators hosts when serving. This enables the inference of large models that don’t fit on a single host. With multi-host inference, JetStream achieves 1703 token/s on Llama3.1-405B on Trillium. This translates to three times more inference per dollar compared to TPU v5e.
In addition, with Pathways, disaggregated serving capabilities allow workloads to dynamically scale LLM inference’s decode and prefill stages independently. This allows for better utilization of resources and can lead to improvements in performance and efficiency, especially for large models. For Llama2-70B, using multiple hosts with disaggregated serving performs seven times better for prefill (time-to-first-token, TTFT) operations, and nearly three times better for token generation (time-per-output-token, TPOT) compared with interleaving the prefill and decode stages of LLM request processing on the same server on Trillium.
Figure 2: Measured using Llama2-70B (MaxText) on Cloud TPU Trillium 16-chips (8 chips allocated for prefill server, 8 chips allocated for decode server). Measured using the OpenOrca dataset. Maximum input length: 1024, maximum output length: 1024. As of April 2025.
Customers like Osmos are using TPUs to maximize cost-efficiency for inference at scale:
“Osmos is building the world’s first AI Data Engineer. This requires us to deploy AI technologies at the cutting edge of what is possible today. We are excited to continue our journey building on Google TPUs as our AI infrastructure for training and inference. We have vLLM and JetStream in scaled production deployment on Trillium and are able to achieve industry leading performance at over 3500 tokens/sec per v6e node for long sequence inference for 70B class models. This gives us industry leading tokens/sec/$, comparable to not just other hardware infrastructure, but also fully managed inference services. The availability of TPUs and the ease of deployment on AI Hypercomputer lets us build out an Enterprise software offering with confidence.” – Kirat Pandya, CEO, Osmos
MaxDiffusion: High-performance diffusion model inference
Beyond LLMs, Trillium demonstrates standout performance on compute-heavy workloads like image generation. MaxDiffusion delivers a collection of reference implementations of various latent diffusion models. In addition to Stable Diffusion inference, we have expanded MaxDiffusion to now support Flux; with 12 billion parameters, Flux is one of the largest open source text-to-image models to date.
As demonstrated on MLPerf 5.0, Trillium now delivers 3.5x throughput improvement for queries/second on Stable Diffusion XL (SDXL) compared to last performance round for its predecessor, TPU v5e. This further improves throughput by 12% since the MLPerf 4.1 submission.
Figure 3: MaxDiffusion throughput (images per second). Google internal data. Measured using the SDXL model on Cloud TPU v5e-4 and Trillium 4-chip. Resolution: 1024×1024, batch size per device: 16, decode steps: 20. As of April 2025.
With this throughput, MaxDiffusion delivers a cost-efficient solution. The cost to generate 1000 images is as low as 22 cents on Trillium, 35% less compared to TPU v5e.
Figure 4: Diffusion cost to generate 1000 images. Google internal data. Measured using the SDXL model on Cloud TPU v5e-4 and Cloud TPU Trillium 4-chip. Resolution: 1024×1024, batch size per device: 2, decode steps: 4. Cost is based on the 3Y CUD prices for Cloud TPU v5e-4 and Cloud TPU Trillium 4-chip in the US. As of April 2025.
A3 Ultra and A4 VMs MLPerf 5.0 Inference results
For MLPerf™ Inference v5.0, we submitted 15 results, including our first submission with A3 Ultra (NVIDIA H200) and A4 (NVIDIA HGX B200) VMs. The A3 Ultra VM is powered by eight NVIDIA H200 Tensor Core GPUs and offers 3.2 Tbps of GPU-to-GPU non-blocking network bandwidth and twice the high bandwidth memory (HBM) compared to A3 Mega with NVIDIA H100 GPUs. Google Cloud’s A3 Ultra demonstrated highly competitive performance, achieving results comparable to NVIDIA’s peak GPU submissions across LLMs, MoE, image, and recommendation models.
Google Cloud was the only cloud provider to submit results on NVIDIA HGX B200 GPUs, demonstrating excellent performance of A4 VM for serving LLMs including Llama 3.1 405B (a new benchmark introduced in MLPerf 5.0). A3 Ultra and A4 VMs both deliver powerful inference performance, a testament to our deep partnership with NVIDIA to provide infrastructure for the most demanding AI workloads.
Customers like JetBrains are using Google Cloud GPU instances to accelerate their inference workloads:
“We’ve been using A3 Mega VMs with NVIDIA H100 Tensor Core GPUs on Google Cloud to run LLM inference across multiple regions. Now, we’re excited to start using A4 VMs powered by NVIDIA HGX B200 GPUs, which we expect will further reduce latency and enhance the responsiveness of AI in JetBrains IDEs.” – Vladislav Tankov, Director of AI, JetBrains
AI Hypercomputer is powering the age of AI inference
Google’s innovations in AI inference, including hardware advancements in Google Cloud TPUs and NVIDIA GPUs, plus software innovations such as JetStream, MaxText, and MaxDiffusion, are enabling AI breakthroughs with integrated software frameworks and hardware accelerators. Learn more about using AI Hypercomputer for inference. Then, check out these JetStream and MaxDiffusion recipes to get started today.
Amazon DocumentDB (with MongoDB compatibility), a fully managed, native JSON database that makes it simple and cost-effective to operate critical document workloads at virtually any scale without managing infrastructure, is now available in the AWS Europe (Stockholm) Region. Amazon DocumentDB provides scalability and durability for mission-critical MongoDB workloads, supporting millions of requests per second and can be scaled to 15 low latency read replicas in minutes without application downtime. Storage scales automatically up to 128 TiB without any impact to your application. In addition, Amazon DocumentDB natively integrates with AWS Database Migration Service (DMS), Amazon CloudWatch, AWS CloudTrail, AWS Lambda, AWS Backup and more.
To learn more about Amazon DocumentDB, please visit the Amazon DocumentDB product page, and see the AWS Region Table for complete regional availability.
Amazon Connect now has new pricing models for external voice transfer and Contact Lens with external voice systems. The new pricing models have independent pricing for external voice connectors and external voice minutes and are effective for all customers from May 1, 2025.
External voice transfer directly transfers voice calls and metadata from Amazon Connect to another voice system, so you can use Amazon Connect telephony and Interactive Voice Response (IVR) to help improve customer experience. Each external transfer connector is now $3,100 per month and each external voice transfer is $0.005 per minute.
Contact Lens with external voice enables Connect Contact Lens contact records, call recording, real-time and post-call analytics, and agent evaluations with your existing voice system to help improve customer experience and agent performance. Each external voice connector is now $3,100 per month and each external voice minute is $0.012 per minute. Standard Contact Lens conversational analytics and performance evaluation charges apply when used with external voice.
Amazon Connect external voice transfer and Contact Lens with external voice are available in the US East (N. Virginia) and US West (Oregon) AWS Regions. To learn more about Amazon Connect and other voice systems, review the following resources:
Today, AWS announced the opening of a new AWS Direct Connect location within the NEXTDC B2 data center near Brisbane, Australia. By connecting your network to AWS at the new location, you gain private, direct access to all public AWS Regions (except those in China), AWS GovCloud Regions, and AWS Local Zones. This site is the first AWS Direct Connect location in Brisbane and the eight AWS Direct connect location within Australia. This Direct Connect location offers dedicated 10 Gbps and 100 Gbps connections with MACsec encryption available.
The Direct Connect service enables you to establish a private, physical network connection between AWS and your data center, office, or colocation environment. These private connections can provide a more consistent network experience than those made over the public internet.
For more information on the over 148 Direct Connect locations worldwide, visit the locations section of the Direct Connect product detail pages. Or, visit our getting started page to learn more about how to purchase and deploy Direct Connect.
Amazon Aurora PostgreSQL Limitless Database is now available with PostgreSQL version 16.8 compatibility, bringing significant improvements and new features. This release contains product improvements and bug fixes made by the PostgreSQL community, along with Aurora Limitless-specific additions such as support for the ltree extension, the btree_gist extension, and improved query performance.
Aurora PostgreSQL Limitless Database makes it easy for you to scale your relational database workloads by providing a serverless endpoint that automatically distributes data and queries across multiple Amazon Aurora Serverless instances while maintaining the transactional consistency of a single database. Aurora PostgreSQL Limitless Database offers capabilities such as distributed query planning and transaction management, removing the need for you to create custom solutions or manage multiple databases to scale. As your workloads increase, Aurora PostgreSQL Limitless Database adds additional compute resources while staying within your specified budget, so there is no need to provision for peak, and compute automatically scales down when demand is low.
Aurora PostgreSQL Limitless Database is available in the following AWS Regions: US East (N. Virginia), US East (Ohio), US West (Oregon), Asia Pacific (Hong Kong), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Europe (Frankfurt), Europe (Ireland), and Europe (Stockholm).
Amazon Relational Database Service (RDS) for PostgreSQL now supports the latest minor versions 17.5, 16.9, 15.13, 14.18, and 13.21. We recommend that you upgrade to the latest minor versions to fix known security vulnerabilities in prior versions of PostgreSQL, and to benefit from the bug fixes added by the PostgreSQL community. This release also includes updates for PostgreSQL extensions such as pg_repack 1.5.1, pg_logical 2.4.5 and others.
You can use automatic minor version upgrades to automatically upgrade your databases to more recent minor versions during scheduled maintenance windows. You can also use Amazon RDS Blue/Green deployments for RDS for PostgreSQL using physical replication for your minor version upgrades. Learn more about upgrading your database instances, including automatic minor version upgrades and Blue/Green Deployments in the Amazon RDS User Guide.
Amazon RDS for PostgreSQL makes it simple to set up, operate, and scale PostgreSQL deployments in the cloud. See Amazon RDS for PostgreSQL Pricing for pricing details and regional availability. Create or update a fully managed Amazon RDS database in the Amazon RDS Management Console.
Amazon SQS now supports VPCE endpoints that have been validated under the Federal Information Processing Standard (FIPS) 140-3 program. You can now easily use AWS PrivateLink with Amazon SQS for regulated workloads that require a secure connection using a FIPS 140-3 validated cryptographic module.
FIPS compliant endpoints help companies contracting with the US federal government meet the FIPS security requirement to encrypt sensitive data in supported regions. To create an interface VPC endpoint that connects to an Amazon SQS FIPS endpoint, see Internetwork traffic privacy in Amazon SQS.
The new capability is available in all AWS Commercial Regions in the United States and Canada. To learn more about FIPS 140-3 at AWS, visit FIPS 140-3 Compliance.
Starting today, customers can use AWS Control Tower in the AWS Asia Pacific (Thailand) and AWS Mexico (Central) Regions. With this launch, AWS Control Tower is available in 32 AWS Regions and the AWS GovCloud (US) Regions. AWS Control Tower offers the easiest way to set up and govern a secure, multi-account AWS environment. It simplifies AWS experiences by orchestrating multiple AWS services on your behalf while maintaining the security and compliance needs of your organization. You can set up a multi-account AWS environment within 30 minutes or less, govern new or existing account configurations, gain visibility into compliance status, and enforce controls at scale.
If you are new to AWS Control Tower, you can launch it today in any of the supported regions, and you can use AWS Control Tower to build and govern your multi-account environment in all supported Regions. If you are already using AWS Control Tower and you want to extend its governance features to the newly supported regions in your accounts, you can go to the settings page in your AWS Control Tower dashboard, select your regions, and update your landing zone. Once you update all accounts that are governed by AWS Control Tower, your landing zone, managed accounts, and registered OUs will be under governance in the new regions.
For a full list of Regions where AWS Control Tower is available, see the AWS Region Table. To learn more, visit the AWS Control Tower homepage or see the AWS Control Tower User Guide.
AWS Security Incident Response is now available to customers in three additional AWS Regions: Asia Pacific (Mumbai), Europe (Paris), and South America (São Paulo). You can now use these additional regions to prepare for, respond to, and recover from security events.
With AWS Security Incident Response, you can enhance your organization’s overall security posture and incident response readiness. AWS Security Incident Response offers three core features: monitoring and triaging of security findings from Amazon GuardDuty and third-party tools through AWS Security Hub; integrated communication and collaboration tools to streamline security escalation and response; and access to self-managed security investigation tools and 24/7 support from the AWS Customer Incident Response Team (CIRT), who can assist you in investigating, containing, eradicating, and recovering from security events. The AWS CIRT has years of experience helping customers recover from security events, building up deep institutional knowledge based on real-world scenarios.
Starting today, you can use AWS Shield Advanced in the AWS Asia Pacific (Thailand) and AWS Mexico (Central) regions. AWS Shield Advanced is a managed application security service that safeguards applications running on AWS from distributed denial of service (DDoS) attacks. Shield Advanced provides always-on detection and automatic inline mitigations that minimize application downtime and latency from DDoS attacks. Also, it provides protections against more sophisticated and larger attacks for your applications running on Amazon Elastic Compute Cloud (EC2), Amazon Elastic Load Balancing (ELB), Amazon CloudFront, AWS Global Accelerator, and Amazon Route 53. To learn more visit, the AWS Shield Advanced product page.
For a full list of AWS regions where AWS Shield Advanced is available, visit the AWS Regional Services page. AWS Shield Advanced pricing may vary between regions. For more information about pricing, visit the AWS Shield Pricing page.
AWS CodePipeline now enables you to use AWS Secrets Manager credentials in your Commands actions by specifying the secrets as environment variables in the action declaration. Additionally, Commands actions now support Windows commands and larger instance types, allowing you to run more complex workloads and accelerate execution times. To learn more about these new capabilities, visit our documentation.
For more information about AWS CodePipeline, visit our product page. This feature is available in all regions where AWS CodePipeline is supported.
Today, AWS announces significant enhancements to Amazon Q Developer in Amazon SageMaker AI Jupyter Lab, introducing customization of code suggestions based on private code repositories and the ability to include entire workspace context for improved code assistance. These new features empower organizations to leverage their proprietary code and improve the relevance of code suggestions, ultimately enhancing developer productivity and code quality within Jupyter Lab environments.
With the new customization feature, Amazon Q Developer can now assist with software development in ways that conform to your team’s internal libraries, proprietary algorithmic techniques, and enterprise code style. An Amazon Q Developer customization is a set of elements that enables Amazon Q to provide suggestions based on your company’s code-base. This ensures that code suggestions, both inline and chat based, align perfectly with your organization’s specific coding practices and standards.
Additionally, the workspace context enables Amazon Q Developer to locate files, understand how code is used across multiple files, and generate code that leverages multiple files, including those that aren’t currently opened. This contextual awareness results in more accurate and relevant code assistance, helping developers better understand their entire project structure before they start coding. Users can access the workspace features through the chat interface, ensuring a seamless development experience that takes into account the full scope of their project.
These enhancements to Amazon Q Developer in Amazon SageMaker AI Jupyter Lab are now available in All Regions where Amazon SageMaker AI is offered.
To learn more about these new features see documentation.
BigQuery delivers optimized search/lookup query performance by efficiently pruning irrelevant files. However, in some cases, additional column information is required for search indexes to further optimize query performance. To help, we recently announced indexing with column granularity, which lets BigQuery pinpoint relevant data within columns, for faster search queries and lower costs.
BigQuery arranges table data into one or more physical files, each holding N rows. This data is stored in a columnar format, meaning each column has its own dedicated file block. You can learn more about this in the BigQuery Storage Internals blog. The default search index is at the file level, which means it maintains mappings from a data token to all the files containing it. Thus, at query time, the search index helps reduce the search space by only scanning those relevant files. This file-level indexing approach excels when search tokens are selective, appearing in only a few files. However, scenarios arise where search tokens are selective within specific columns but common across others, causing these tokens to appear in most files, and thus diminishing the effectiveness of file-level indexes.
For example, imagine a scenario where we have a collection of technical articles stored in a simplified table named TechArticles with two columns — Title and Content. And let’s assume that the data is distributed across four files, as shown below.
Our goal is to search for articles specifically related to Google Cloud Logging. Note that:
The tokens “google”, “cloud”, and “logging” appear in every file.
Those three tokens also appear in the “Title” column, but only in the first file.
Therefore, the combination of the three tokens is common overall, but highly selective in the “Title” column.
Now, let’s say, we create a search index on both columns of the table with the following DDL statement:
CREATE SEARCH INDEX myIndex ON myDataset.TechArticles(Title, Content);
The search index stores the mapping of data tokens to the data files containing the tokens, without any column information; the index looks like the following (showing the three tokens of interest: “google”, “cloud”, and “logging”):
With the usual query SELECT * FROM TechArticles WHERE SEARCH(Title, "Google Cloud Logging"), using the index without column information, BigQuery ends up scanning all four files, adding unnecessary processing and latency to your query.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud data analytics’), (‘body’, <wagtail.rich_text.RichText object at 0x3e4f83a65a00>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/bigquery/’), (‘image’, None)])]>
Indexing with column granularity
Indexing with column granularity, a new public preview feature in BigQuery, addresses this challenge by adding column information in the indexes. This lets BigQuery leverage the indexes to pinpoint relevant data within columns, even when the search tokens are prevalent across the table’s files.
Let’s go back to the above example. Now we can create the index with COLUMN granularity as follows:
CREATE SEARCH INDEX myIndex ON myDataset.TechArticles(Title, Content)
OPTIONS (default_index_column_granularity = 'COLUMN');
The index now stores the column information associated with each data token. The index is as follows:
Using the same query SELECT * FROM TechArticles WHERE SEARCH(Title, "Google Cloud Logging") as above but using the index with column information, BigQuery now only needs to scan file1 since the index lookup is the intersection of the following:
Files where Token=’google’ AND Column=’Title’ (file1)
Files where Token=’cloud’ AND Column=’Title’ (file1, file2, file3, and file4)
Files where Token’=’logging’ AND Column=’Title’ (file1).
Performance improvement benchmark results
We benchmarked query performance on a 1TB table containing Google Cloud Logging data of an internal Google test project with the following query:
SELECT COUNT(*)
FROM `dataset.log_1T`
WHERE SEARCH((logName, trace, labels, metadata), 'appengine');
In this benchmark query, the token ‘appengine’ appears infrequently in the columns used for query filtering, but is more common in other columns. The default search index already helped reduce a large portion of the search space, resulting in half the execution time, reducing processed bytes and slot usage. By employing column granularity indexing, the improvements are even more significant.
In short, column-granularity indexing in BigQuery offers the following benefits:
Enhanced query performance: By precisely identifying relevant data within columns, column-granularity indexing significantly accelerates query execution, especially for queries with selective search tokens within specific columns.
Improved cost efficiency: Index pruning results in reduced bytes processed and/or slot time, translating to improved cost efficiency.
This is particularly valuable in scenarios where search tokens are selective within specific columns but common across others, or where queries frequently filter or aggregate data based on specific columns.
Best practices and getting started
Indexing with column granularity represents a significant advancement in BigQuery’s indexing capabilities, letting you achieve greater query performance and cost efficiency.
For best results, consider the following best practices:
Identify high-impact columns: Analyze your query patterns to identify columns that are frequently used in filters or aggregations and would benefit from column-granularity indexing.
Monitor performance: Continuously monitor query performance and adjust your indexing strategy as needed.
Consider indexing and storage costs: While column-granularity indexing can optimize query performance, be mindful of potential increases in indexing and storage costs.
At Google Cloud Next 25, we announced a major step forward in geospatial analytics: Earth Engine in BigQuery. This new capability unlocks Earth Engine raster analytics directly in BigQuery, making advanced analysis of geospatial datasets derived from satellite imagery accessible to the SQL community.
Before we get into the details of this new capability and how it can power your use cases, it’s helpful to distinguish between two types of geospatial data and where Earth Engine and BigQuery have historically excelled:
Raster data: This type of data represents geographic information as a grid of cells, or pixels, where each pixel stores a value that represents a specific attribute such as elevation, temperature, or land cover. Satellite imagery is a prime example of raster data. Earth Engine excels at storing and processing raster data, enabling complex image analysis and manipulation.
Vector data: This type of data represents geographic features such as points, lines, or polygons. Vector data is ideal for representing discrete objects like buildings, roads, or administrative boundaries. BigQuery is highly efficient at storing and querying vector data, making it well-suited for large-scale geographic analysis.
Earth Engine and BigQuery are both powerful platforms in their own right. By combining their geospatial capabilities, we are bringing the best of both raster and vector analytics to one place. That’s why we created Earth Engine in BigQuery, an extension to BigQuery’s current geospatial capabilities that will broaden access to raster analytics and make it easier than ever before to answer a wide range of real-world enterprise problems.
Earth Engine in BigQuery: Key features
You can use the two key features of Earth Engine in BigQuery to perform raster analytics in BigQuery:
A new function in BigQuery: Run ST_RegionStats(), a new BigQuery geography function that lets you efficiently extract statistics from raster data within specified geographic boundaries.
New Earth Engine datasets in BigQuery Sharing (formerly Analytics Hub): Access a growing collection of Earth Engine datasets in BigQuery Sharing (formerly Analytics Hub), simplifying data discovery and access. Many of these datasets are analysis-ready, immediately usable for deriving statistics for an area of interest, and providing valuable information such as elevation, emissions, or risk prediction.
Five easy steps to raster analytics
The new ST_RegionStats() function is similar to Earth Engine’s reduceRegion function, which allows you to compute statistics for one or more regions of an image. The ST_RegionStats() function is a new addition to BigQuery’s set of geography functions invoked as part of any BigQuery SQL expression. It takes an area of interest (e.g., a county, parcel of land, or zip code) indicated by a geography and an Earth Engine-accessible raster image and computes a set of aggregate values for the pixels that intersect with the specified geography. Examples of aggregate statistics for an area of interest would be maximum flood depth or average methane emissions for a certain county.
These are the five steps to developing meaningful insights for an area of interest:
Identify a BigQuery table with vector data: This could be data representing administrative boundaries (e.g., counties, states), customer locations, or any other geographic areas of interest. You can pull a dataset from BigQuery public datasets or use your own based on your needs.
Identify a raster dataset: You can discover Earth Engine raster datasets in BigQuery Sharing, or you can use raster data stored as a Cloud GeoTiff or Earth Engine image asset. This can be any raster dataset that contains the information you want to analyze within the vector boundaries.
Use ST_RegionStats() to bring raster data into BigQuery: The ST_RegionStats() geography function takes the raster data (raster_id), vector geometries (geography), and optional band (band_name) as inputs and calculates aggregate values (e.g., mean, min, max, sum, count) on the intersecting raster data and vector feature.
Analyze the results: You can use the output of running ST_RegionStats() to analyze the relationship between the raster data and the vector features, generating valuable insights about an area of interest.
Visualize the results: Geospatial analysis is usually most impactful when visualized on a map. Tools like BigQuery Geo Viz allow you to easily create interactive maps that display your analysis results, making it easier to understand spatial patterns and communicate findings.
Toward data-driven decision making
The availability of Earth Engine in BigQuery opens up new possibilities for scaled data-driven decision-making across various geospatial and sustainability use cases, by enabling raster analytics on datasets that were previously unavailable in BigQuery. These datasets can be used with the new ST_RegionStats() geography function for a variety of use cases, such as calculating different land cover types within specific administrative boundaries or analyzing the average elevation suitability within proposed development areas. You can also find sample queries for these datasets in BigQuery Sharing’s individual dataset pages. For example, if you navigate to the GRIDMET CONUS Drought Indices dataset page, you can find a sample query for calculating mean Palmer Drought Severity Index (PDSI) for each county in California, used to monitor drought conditions across the United States.
Let’s take a deeper look at some of the use cases that this new capability unlocks:
1. Climate, physical risk, and disaster response
Raster data can provide critical insights on weather patterns and natural disaster monitoring. Many of the raster datasets available in BigQuery Sharing provide derived data on flood mapping, wildfire risk assessment, drought conditions, and more. These insights can be used for disaster risk and response, urban planning, infrastructure development, transportation management, and more. For example, you could use the Wildfire Risk to Communities dataset for predictive analytics, allowing you to assess wildfire hazard risk, exposure of communities, and vulnerability factors, so you can develop effective resilience strategies. For flood mapping, you could use the Global River Flood Hazard dataset to understand regions in the US that have the highest predicted inundation depth, or water height above ground surface.
2. Sustainable sourcing and agriculture
Raster data also provides insights on land cover and land use over time. Several of the new Earth Engine datasets in BigQuery include derived data on terrain, elevation, and land-cover classification, which are critical inputs for supply chain management and assessing agriculture and food security. For businesses that operate in global markets, sustainable sourcing requires bringing transparency and visibility to supply chains, particularly as regulatory requirements are shifting commitments to deforestation-free commodity production from being voluntary to mandatory. With the new Forest Data Partnership maps for cocoa, palm and rubber, you can analyze where commodities are grown over time, and add in the Forest Persistence or the JRC Global Forest Cover datasets to understand if those commodities are being grown in areas that had not been deforested or degraded before 2020. With a simple SQL query, you could, for instance, determine the estimated fraction of Indonesia’s land area that had undisturbed forest in 2020.
3. Methane emissions monitoring
Reducing methane emissions from the oil and gas industry is crucial to slow the rate of climate change. The MethaneSAT L4 Area Sources dataset, which can be used as an Earth Engine Image asset with the ST_RegionStats() function, provides insights into small, dispersed area emissions of methane from various sources. This type of diffuse but widespread emissions can make up the majority of methane emissions in an oil and gas basin. You can analyze the location, magnitude, and trends of these emissions to identify hotspots, inform mitigation efforts, and understand how emissions are characterized across large areas, such as basins.
4. Custom use cases
In addition to these datasets, you can bring your own raster datasets via Cloud Storage GeoTiffs or Earth Engine image assets, to support other use cases, while still benefiting from BigQuery’s scalability and analytical tools.
aside_block
<ListValue: [StructValue([(‘title’, ‘$300 in free credit to try Google Cloud data analytics’), (‘body’, <wagtail.rich_text.RichText object at 0x3e4f833ed370>), (‘btn_text’, ‘Start building for free’), (‘href’, ‘http://console.cloud.google.com/freetrial?redirectPath=/bigquery/’), (‘image’, None)])]>
Bringing it all together with an example
Let’s take a look at a more advanced example based on modeled wildfire risk and AI-driven weather forecasting technology. The SQL below uses the Wildfire Risk to Communities dataset listed in BigQuery Sharing, which is designed to help communities understand and mitigate their exposure to wildfire. The data contains bands that index the likelihood and consequence of wildfire across the landscape. Using geometries from a public dataset of census-designated places, you can compute values from this dataset using ST_RegionStats() to compare communities’ relative risk exposures. You can also combine weather data from WeatherNext Graph forecasts to see how imminent fire weather is predicted to affect those communities.
To start, head to the BigQuery Sharing console, click “Search listings”, filter to “Climate and environment,” select the “Wildfire Risk to Community” dataset (or search for the dataset in the search bar), and click “Subscribe” to add the Wildfire Risk dataset to your BigQuery project. Then search for “WeatherNext Graph” and subscribe to the WeatherNext Graph dataset.
1
2
3
With these subscriptions in place, run a query to combine these datasets across many communities with a single query. You can break this task into subqueries using the SQL WITH statement for clarity:
First, select the input tables that you subscribed to in the previous step.
Second, compute the weather forecast using WeatherNext Graph forecast data for a specific date and for the places of interest. The result is the average and maximum wind speeds within each community.
Third, use the ST_RegionStats() function to sample the Wildfire Risk to Community raster data for each community. Since we are only concerned with computing mean values within regions, you can set the scale to 1 kilometer in the function options in order to use lower-resolution overviews and thus reduce compute time. To compute at the full resolution of the raster (in this case, 30 meters), you can leave this option out.
The result is a table containing the mean values of wildfire risk for both bands within each community and wind speeds projected over the course of a day. In addition, you can combine the computed values for wildfire risk, wildfire consequence, and maximum wind speed into a single composite index to show relative wildfire exposure for a selected day in Colorado.
WITH
-- Step 1: Select inputs from datasets that we've subscribed to
wildfire_raster AS (
SELECT
id
FROM
`wildfire_risk_to_community_v0_mosaic.fire`
),
places AS (
SELECT
place_id,
place_name,
place_geom AS geo,
FROM
`bigquery-public-data.geo_us_census_places.places_colorado`
),
-- Step 2: Compute the weather forecast using WeatherNext Graph forecast data
weather_forecast AS (
SELECT
ANY_VALUE(place_name) AS place_name,
ANY_VALUE(geo) AS geo,
AVG(SQRT(POW(t2.`10m_u_component_of_wind`, 2)
+ POW(t2.`10m_v_component_of_wind`, 2))) AS average_wind_speed,
MAX(SQRT(POW(t2.`10m_u_component_of_wind`, 2)
+ POW(t2.`10m_v_component_of_wind`, 2))) AS maximum_wind_speed
FROM
`weathernext_graph_forecasts.59572747_4_0` AS t1,
t1.forecast AS t2
JOIN
places
ON
ST_INTERSECTS(t1.geography_polygon, geo)
WHERE
t1.init_time = TIMESTAMP('2025-04-28 00:00:00 UTC')
AND t2.hours < 24
GROUP BY
place_id
),
-- Step 3: Combine with wildfire risk for each community
wildfire_risk AS (
SELECT
geo,
place_name,
ST_REGIONSTATS( -- Wildfire likelihood
geo, -- Place geometry
(SELECT id FROM wildfire_raster), -- Raster ID
'RPS', -- Band name (Risk to Potential Structures)
OPTIONS => JSON '{"scale": 1000}' -- Computation resolution in meters
).mean AS wildfire_likelihood,
ST_REGIONSTATS( -- Wildfire consequence
geo, -- Place geometry
(SELECT id FROM wildfire_raster), -- Raster ID
'CRPS', -- Band name (Conditional Risk to Potential Structures)
OPTIONS => JSON '{"scale": 1000}' -- Computation resolution in meters
).mean AS wildfire_consequence,
weather_forecast.* EXCEPT (geo, place_name)
FROM
weather_forecast
)
-- Step 4: Compute a composite index of relative wildfire risk.
SELECT
*,
PERCENT_RANK() OVER (ORDER BY wildfire_likelihood)
* PERCENT_RANK() OVER (ORDER BY wildfire_consequence)
* PERCENT_RANK() OVER (ORDER BY average_wind_speed)
AS relative_risk
FROM
wildfire_risk
Mean values of wildfire risk and wind speeds for each community
You can save this output in Google Sheets to visualize how wildfire risk and consequences are related among communities statewide.
Google sheet visualizing relationship between wildfire risk (x-axis) and wildfire consequence (y-axis) colored by wind speed
Alternatively, you can visualize relative wildfire risk exposure in BigQuery GeoViz with the single composite index to show relative wildfire exposure for a selected day in Colorado.
GeoViz map showing a composite index combining values for wildfire risk, wildfire consequence, and maximum wind speed for each community
What’s next for Earth Engine in BigQuery?
Earth Engine in BigQuery marks a significant advancement in geospatial analytics, and we’re excited to further expand raster analytics in BigQuery, making sustainability decision-making easier than ever before. Learn more about this new capability in the BigQuery documentation for working with raster data, and stay tuned for new Earth Engine capabilities in BigQuery in the near future!
Today, we are announcing the availability of Route 53 Resolver Query Logging in the Asia Pacific (Thailand) and Mexico (Central) Regions, enabling you to log DNS queries that originate in your Amazon Virtual Private Clouds (Amazon VPCs). With query logging enabled, you can see which domain names have been queried, the AWS resources from which the queries originated – including source IP and instance ID – and the responses that were received.
Route 53 Resolver is the Amazon DNS server that is available by default in all Amazon VPCs. Route 53 Resolver responds to DNS queries from AWS resources within a VPC for public DNS records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones. With Route 53 Resolver Query Logging, customers can log DNS queries and responses for queries originating from within their VPCs, whether those queries are answered locally by Route 53 Resolver, or are resolved over the public internet, or are forwarded to on-premises DNS servers via Resolver Endpoints. You can share your query logging configurations across multiple accounts using AWS Resource Access Manager (RAM). You can also choose to send your query logs to Amazon S3, Amazon CloudWatch Logs, or Amazon Kinesis Data Firehose.
There is no additional charge to use Route 53 Resolver Query Logging, although you may incur usage charges from Amazon S3, Amazon CloudWatch, or Amazon Kinesis Data Firehose. To learn more about Route 53 Resolver Query Logging or to get started, visit the Route 53 product page or the Route 53 documentation.
Starting today, the Amazon Elastic Compute Cloud (Amazon EC2) Inf2 instances, optimized for generative AI, are generally available in the AWS Asia Pacific (Seoul) Region. Amazon EC2 Inf2 instances deliver up to 40% lower inference costs over comparable Amazon EC2 instances.
You can use Inf2 instances to run popular applications such as text summarization, code generation, video and image generation, speech recognition, personalization, and more. Inf2 instances are the first inference-optimized instances in Amazon EC2 to introduce scale-out distributed inference supported by NeuronLink, a high-speed, nonblocking interconnect. Inf2 instances offer up to 2.3 petaflops and up to 384 GB of total accelerator memory with 9.8 TB/s bandwidth.
The AWS Neuron SDK integrates natively with popular machine learning frameworks, so you can continue using your existing frameworks to deploy on Inf2. Developers can get started with Inf2 instances using AWS Deep Learning AMIs, AWS Deep Learning Containers, or managed services such as Amazon Elastic Container Service (Amazon ECS), Amazon Elastic Kubernetes Service (Amazon EKS), and Amazon SageMaker.
Inf2 instances are now available in four sizes: inf2.xlarge, inf2.8xlarge, inf2.24xlarge, inf2.48xlarge in 14 AWS Regions as On-Demand Instances, Reserved Instances, and Spot Instances, or as part of a Savings Plan.
Amazon S3 Tables are now available in eleven additional AWS Regions: Africa (Cape Town), Asia Pacific (Hong Kong), Asia Pacific (Hyderabad), Asia Pacific (Jakarta), Asia Pacific (Malaysia), Asia Pacific (Melbourne), Canada West (Calgary), Europe (Milan), Europe (Zurich), Israel (Tel Aviv), and Middle East (Bahrain). S3 Tables deliver the first cloud object store with built-in Apache Iceberg support, and the easiest way to store tabular data at scale.
Amazon Web Services (AWS) has expanded the regional availability for Amazon AppStream 2.0. Starting today, AWS customers can deploy their applications and desktops in the AWS Europe (Paris) Region and stream them using AppStream 2.0.
Deploying your applications on AppStream 2.0 in a region closer to your end users helps provide a more responsive experience. Additionally, European Union customers now have another AWS region option to deploy their workloads on AppStream 2.0.
Amazon AppStream 2.0 is a fully managed, secure application streaming service that provides users with instant access to their desktop applications from anywhere. It allows users to stream applications and desktops from AWS to their devices, without requiring them to download, install, or manage any software locally. AppStream 2.0 manages the AWS resources required to host and run your applications, scales automatically, and provides access to your users on demand.
To get started with Amazon AppStream 2.0, sign into the AppStream 2.0 management console and select Europe (Paris) Region. For the full list of Regions where AppStream 2.0 is available, see the AWS Region Table. AppStream 2.0 offers pay-as-you-go pricing. For more information, see Amazon AppStream 2.0 Pricing.
Amazon Elastic Compute Cloud (Amazon EC2) C8gd instances, Amazon EC2 M8gd instances, and Amazon EC2 R8gd instances with up to 11.4 TB of local NVMe-based SSD block-level storage are now available in AWS Region Europe (Frankfurt). These instances are powered by AWS Graviton4 processors, delivering up to 30% better performance over Graviton3-based instances. They have up to 40% higher performance for I/O intensive database workloads, and up to 20% faster query results for I/O intensive real-time data analytics than comparable AWS Graviton3-based instances. These instances are built on the AWS Nitro System and are a great fit for applications that need access to high-speed, low latency local storage.
Each instance is available in 12 different sizes. They provide up to 50 Gbps of network bandwidth and up to 40 Gbps of bandwidth to the Amazon Elastic Block Store (Amazon EBS). Additionally, customers can now adjust the network and Amazon EBS bandwidth on these instances by 25% using EC2 instance bandwidth weighting configuration, providing greater flexibility with the allocation of bandwidth resources to better optimize workloads. These instances offer Elastic Fabric Adapter (EFA) networking on 24xlarge, 48xlarge, metal-24xl, and metal-48xl sizes.
These instances are now available in AWS Regions US East (Ohio, N. Virginia), US West (Oregon), and Europe (Frankfurt).