Catalog
concept#Architecture#Cloud#DevOps#Observability#Platform

Serverless Architecture

An architectural paradigm where applications run on managed cloud services and event-driven functions without managing server infrastructure.

Serverless architecture is a cloud-native paradigm where applications run on fully managed services and event-driven functions without managing server infrastructure.
Established
Medium

Classification

  • Medium
  • Technical
  • Architectural
  • Intermediate

Technical context

API Gateway / load balancerManaged databases (e.g. DynamoDB, Cloud SQL)Observability tools (tracing, logs, metrics)

Principles & goals

Design for loose coupling and short execution times.Treat functions as immutable, isolated deployables.Instrument observability and cost metrics from day one.
Build
Enterprise, Domain, Team

Use cases & scenarios

Compromises

  • Unexpected costs from faulty scaling or hot loops.
  • Difficulties in distributed tracing and debugging.
  • Security risks from misconfigured managed services.
  • Keep functions small and single-responsibility.
  • Implement idempotency and clear retry strategies.
  • Instrument cost metrics and set budget limits.

I/O & resources

  • Defined event sources and triggers
  • Minimal function code with clear interface contracts
  • Managed services for persistence, auth and messaging
  • Scalable functions as deployable artifacts
  • Metrics for latency, cost and errors
  • Integrated observability and alerting pipelines

Description

Serverless architecture is a cloud-native paradigm where applications run on fully managed services and event-driven functions without managing server infrastructure. It reduces operational overhead and enables elastic scaling, but introduces design trade-offs around cold starts, dependency management, cost model and potential vendor lock-in. Compared to traditional architectures, billing and observability approaches change and new operational patterns are required.

  • Reduced operational overhead via managed services.
  • Automatic, fine-grained scaling on demand.
  • Pay-per-use cost model for irregular workloads.

  • Cold-start latencies for infrequently used functions.
  • Limited execution time or resources per invocation.
  • Vendor lock-in due to proprietary services and APIs.

  • Cold-start latency

    Time between function invocation and fully operational execution.

  • Cost per request

    Average billing cost per function invocation or API call.

  • Error rate / retries

    Share of failed executions and number of automatic retries.

Web API with AWS Lambda

API endpoints served by Lambda functions; API Gateway handles authentication and rate limiting.

Image processing pipeline in functions

Upload triggers a function that transforms images and stores them in object storage; processing scales with input rate.

Scheduled report generation

Scheduled Lambda functions generate reports and deliver results to storage or notification systems.

1

Conduct architecture workshop and select use cases.

2

Implement a proof-of-concept for a critical path.

3

Integrate observability, cost alerts and security policies.

⚠️ Technical debt & bottlenecks

  • Hidden dependencies on provider-specific services.
  • Insufficient test coverage for asynchronous processes.
  • Monoliths that become hard to split into functions later.
Cold startsVendor lock-inObservability
  • Migration without redesign: lift-and-shift of monolithic services into individual functions.
  • High-frequency low-latency applications that suffer from cold starts.
  • Uncontrolled cost-intensive events resulting from misconfiguration.
  • Underestimating total operating costs despite pay-per-use.
  • Missing end-to-end traces across multiple managed services.
  • Untested provider failure scenarios.
Cloud architecture and service selectionAsynchronous design and event modelingObservability, monitoring and cost analysis
Scalability for variable loadsMinimization of operational effortFast time-to-market
  • Maximum execution time per function (provider limit)
  • Limited native resources (CPU/memory)
  • Network latency to managed services