Back to Blog
awscloudserverlessnews

AWS Lambda & S3 Express One Zone: A 2025 Deep Dive into re:Invent 2023's Impact

Discover how AWS re:Invent 2023's Lambda scaling and S3 Express One Zone have reshaped high-performance serverless & storage in 2025. Uncover benchmarks, trade-offs, and real-world impact.

DataFormatHub Team
December 21, 2025
Share:
AWS Lambda & S3 Express One Zone: A 2025 Deep Dive into re:Invent 2023's Impact

It’s late 2025, and another AWS re:Invent has just concluded, but as an engineer who's been knee-deep in the trenches, I find myself reflecting less on the fresh-off-the-press announcements and more on the ones that have truly matured over the past year. Specifically, the game-changing developments unveiled at re:Invent 2023 concerning AWS Lambda's scaling capabilities and the introduction of Amazon S3 Express One Zone have fundamentally reshaped how we approach high-performance serverless and storage architectures. These aren't just new features; they're hardened tools we've been putting through their paces, and the numbers tell an interesting story.

Let's cut through the marketing fluff and dive into the practical realities of what these updates mean for your production workloads, complete with benchmarks and the inevitable trade-offs.

AWS Lambda's New Scaling Paradigm: Breaking the Bottlenecks

For years, the Achilles' heel for many highly burstable serverless applications on AWS Lambda wasn't the execution speed of individual functions, but rather the rate at which Lambda could provision new execution environments. Prior to re:Invent 2023, Lambda scaled synchronously invoked functions by creating 500 to 3,000 new execution environments in the first minute (depending on the region), followed by an additional 500 environments every minute thereafter. Crucially, these scaling quotas were shared across all Lambda functions within an account in a given region. This meant that a sudden surge in traffic to one hot function could starve another critical function of the necessary scaling capacity, leading to throttles and increased latency.

The Numbers Tell an Interesting Story: 12x Faster, Independent Scaling

The re:Invent 2023 announcement fundamentally altered this dynamic. AWS Lambda now scales each synchronously invoked function up to 12 times faster, allowing it to provision 1,000 concurrent executions every 10 seconds. Even more impactful, each function now scales independently until the account's concurrency limit is reached. This is a significant shift.

Let's look at the raw comparison:

  • Pre-re:Invent 2023: Initial burst of 500-3000 concurrent executions in the first minute, then +500/minute, shared across the account.
  • Post-re:Invent 2023: Initial burst of 1,000 concurrent executions every 10 seconds (i.e., 6,000 per minute), per function, independently.

This translates into a vastly more responsive system for event-driven architectures. Consider a scenario with an API endpoint backed by Lambda and an SQS queue processing asynchronous tasks, both experiencing simultaneous spikes. In the old model, the API's scaling could be hampered by the queue processor's demand, or vice-versa. Now, both can aggressively scale out to meet demand, each consuming its share of the account's total concurrency without directly impeding the other's ramp-up speed.

For instance, processing messages from SQS and Kafka event sources also benefits, allowing for quicker message processing and reduced queue backlogs during peak times. We've observed this dramatically reduce the need for aggressive pre-warming strategies or over-provisioning for many of our bursty workloads.

Reality Check: Still an Account-Level Game

While the per-function independent scaling is a robust improvement, it's vital to remember that the account-level concurrency limit still exists. If your aggregate demand across all functions exceeds this limit, you'll still hit throttles. The default quota is typically 1,000 concurrent executions, though it can be increased significantly upon request. The implication here is that while individual functions are more agile, overall capacity planning and monitoring of account-level concurrency remain critical.

Furthermore, this rapid scaling can expose bottlenecks in downstream services. An API Gateway, with its default limit of 10,000 requests per second, could become the new throttle point if your Lambda functions are now scaling much faster than your API Gateway can handle requests. Architectural review of your entire request path, not just Lambda, is more important than ever.

A Quick Code Example (Conceptual):

Imagine a Python Lambda function triggered by an HTTP API Gateway request:

# app.py
import json
import os
import time

# Simulate some initial setup/dependency loading (pre-handler initialization)
# This code runs once per execution environment (cold start)
GLOBAL_RESOURCE = os.getenv('GLOBAL_RESOURCE', 'initialized')
print(f"[{os.getpid()}] Global resource: {GLOBAL_RESOURCE} (initialized at {time.time()})")

def lambda_handler(event, context):
    start_time = time.monotonic()
    
    # Simulate work that scales with invocation
    payload = json.loads(event.get('body', '{}'))
    task_duration = int(payload.get('duration_ms', 50)) / 1000.0
    time.sleep(task_duration) # Simulate I/O or CPU work
    
    end_time = time.monotonic()
    response_time_ms = (end_time - start_time) * 1000
    
    print(f"[{os.getpid()}] Invocation completed in {response_time_ms:.2f}ms")
    
    return {
        'statusCode': 200,
        'body': json.dumps({
            'message': f'Processed in {response_time_ms:.2f}ms',
            'pid': os.getpid(),
            'global_resource': GLOBAL_RESOURCE
        })
    }

With the improved scaling, deploying this function and hitting it with a sudden surge of requests would see new pids (new execution environments) spin up far more rapidly and consistently than before, allowing the system to absorb the load much more effectively, provided the account concurrency and downstream services can keep pace.

Underlying Platform Stability: Amazon Linux 2023

A quieter but foundational improvement from re:Invent 2023 was the introduction of Amazon Linux 2023 (AL2023) as a managed runtime and container base image for Lambda. AL2023 provides an OS-only environment with a smaller deployment footprint, updated libraries (like glibc), and a new package manager compared to its predecessor, AL2. This isn't a direct performance booster in the same vein as scaling, but it's a sturdy platform improvement that contributes to more efficient custom runtimes and will serve as the base for future Lambda managed runtimes (e.g., Node.js 20, Python 3.12, Java 21). Smaller base images mean potentially faster download times during cold starts and a more modern, secure environment.

S3 Express One Zone: A New Tier for the Performance-Critical

For nearly two decades, Amazon S3 has been the workhorse of cloud storage, renowned for its scalability, durability, and versatility. However, for extremely low-latency, high-QPS (Queries Per Second) workloads like machine learning training, interactive analytics, or high-performance computing, S3 Standard's multi-AZ architecture, while providing incredible durability and availability, introduced inherent network latency that could become a bottleneck. Custom caching layers were often employed, adding complexity and operational overhead.

The "Why" and "What": Single-Digit Milliseconds, Millions of Requests

Enter Amazon S3 Express One Zone, announced at re:Invent 2023. This new storage class is purpose-built to deliver the fastest cloud object storage, promising consistent single-digit millisecond latency and the ability to scale to hundreds of thousands of requests per second, even millions of requests per minute, for frequently accessed data. The key architectural differentiator is its single Availability Zone (AZ) deployment.

Architectural Nuances for Performance:

  1. Directory Buckets: To achieve its high TPS goals, S3 Express One Zone introduces a new bucket type: Directory Buckets. Unlike general-purpose S3 buckets that scale incrementally, directory buckets are designed for instantaneous scaling to hundreds of thousands of requests per second. This is a crucial distinction when optimizing for extreme throughput.
  2. Co-location with Compute: By storing data within a single AZ, you can co-locate your compute resources (EC2, ECS, EKS) in the same AZ, dramatically reducing network latency between compute and storage. This is where a significant portion of the performance gain comes from, minimizing inter-AZ hops.
  3. Session-based Authentication: A new CreateSession API is introduced, optimized for faster authentication and authorization of requests, further shaving off precious milliseconds from the request path.

Benchmarks & Comparisons: The Raw Performance

AWS claims S3 Express One Zone is up to 10x faster than S3 Standard. For small objects, where time to first byte is a dominant factor, the benefit is particularly pronounced. In internal tests at re:Invent 2023, downloading 100,000 objects showed S3 Express achieving about 9 GB/s throughput compared to S3 Standard's 1 GB/s, with average latencies of 80ms for S3 Standard dropping to single-digit milliseconds for S3 Express.

Beyond raw speed, S3 Express One Zone also boasts 50% lower request costs compared to S3 Standard. This, combined with more efficient compute utilization (less idle time waiting for storage), can lead to overall cost reductions, with some customers seeing up to a 60% reduction in total cost of ownership for specific applications.

This storage class is a practical choice for:

  • AI/ML Training and Inference: Where models frequently access vast, often small, datasets.
  • Interactive Analytics: Accelerating query times for services like Athena or EMR.
  • Media Processing: Especially for workflows requiring rapid access to many small media assets.
  • High-Performance Computing: Any workload that is extremely I/O bound.

Reality Check: Durability and Management Trade-offs

The primary trade-off with S3 Express One Zone is its single-AZ durability model. While it offers 11 nines of durability within that single AZ (achieved through end-to-end integrity checks, redundant storage across multiple devices, and continuous monitoring), it is not resilient to the loss or damage of an entire Availability Zone. This means that in the event of a catastrophic AZ failure (e.g., fire, water damage), data stored only in S3 Express One Zone in that AZ could be lost.

For mission-critical data, customers must explicitly build cross-AZ redundancy or backup solutions (e.g., replicating to S3 Standard in another AZ). This adds a layer of architectural responsibility that S3 Standard's regional durability abstracts away.

Another point of consideration is the introduction of a new bucket type ("Directory"). While functionally powerful, it adds a slight complexity to S3 bucket management, requiring developers to choose between general-purpose and directory buckets based on their access patterns and performance needs. The storage cost per GB is also higher than S3 Standard, though, as noted, this is often offset by reduced request costs and increased compute efficiency.

Practical Implications and the Road Ahead

A year after their announcement, both Lambda's enhanced scaling and S3 Express One Zone have proven to be sturdy, efficient additions to the AWS toolkit. We've seen them enable more responsive applications, simplify certain architectural patterns (like removing custom caching layers for high-performance S3 access), and provide tangible cost savings through optimized compute usage.

The independent scaling of Lambda functions has significantly improved our ability to handle unpredictable traffic spikes without complex pre-warming or fear of resource contention between services. For S3, the Express One Zone class has opened doors for workloads previously constrained by object storage latency, especially in the burgeoning AI/ML space. The explicit trade-off in durability for extreme performance is a clear design choice that developers must actively consider, not an oversight.

These developments from re:Invent 2023 underscore AWS's continued focus on performance, efficiency, and giving developers finer-grained control over their infrastructure, even within serverless and managed services. As we continue to push the boundaries of cloud-native applications, these foundational improvements provide a solid, pragmatic bedrock for innovation.


Sources


πŸ› οΈ Related Tools

Explore these DataFormatHub tools related to this topic:


πŸ“š You Might Also Like