Integrating the AWS Lambda Telemetry API with Prometheus and OpenSearch

January 24, 2024 By Mark Otto 0

The AWS Lambda service automatically captures telemetry, such as metrics, logs, and traces, and sends it to Amazon CloudWatch and AWS X-Ray. There’s nothing you need to do for it to happen, this is available out-of-the-box. However, if your enterprise has adopted an open source observability solution like Prometheus or OpenSearch, you can use the AWS Lambda Telemetry API to stream telemetry to the open source tool of your choice directly from the AWS Lambda service.

This blog covers the details of how the AWS Lambda Telemetry API works and how to integrate it with open source observability and telemetry solutions such as Prometheus and OpenSearch.

Overview

Observability is one of the critical traits of any modern application. Observability helps you to gain insights necessary to understand what’s happening in the application environment and how to improve its performance. The AWS Cloud Adoption Framework defines monitoring, tracing, and logging as three pillars of observability.

AWS provides out-of-the-box solutions for addressing observability requirements when building serverless applications using the AWS Lambda service.

  • AWS Lambda service sends function logs to Amazon CloudWatch Logs. Those logs include both system logs generated by the AWS Lambda service (also known as platform logs), and application logs produced by your functions (also known as function logs). You can use additional Amazon CloudWatch capabilities such as Logs Insights to query the logs or Live Tail to debug your functions.
  • AWS Lambda service emits operational and performance metrics to Amazon CloudWatch Metrics, which you can use to build operational and business dashboards. You can build widgets visualizing these metrics and set alarms to respond to changes in utilization or error rates.
  • When you enable Active Tracing for your Lambda functions, or when a Lambda function is invoked by a service or application that also has AWS X-Ray tracing enabled, the AWS Lambda service will start collecting initialization and invocation traces and sending them to the AWS X-Ray service to visualize, identify bottlenecks, and debug suspicious behaviors.

AWS Lambda service provides these capabilities out-of-the-box through native integrations with Amazon CloudWatch and AWS X-Ray, making it easy for application builders to meet their observability needs without any extra instrumentation or code changes. However, some customers choose to implement their own centralized enterprise observability platform using open source alternatives such as Prometheus or OpenSearch. The AWS Lambda Telemetry API enables you to use these open source tools to receive fine-grained, per-invoke telemetry data directly from the AWS Lambda service.

Understanding the AWS Lambda Telemetry API

When using AWS Lambda Telemetry API, your extensions receive telemetry data directly from the AWS Lambda service. During function initialization and invocation, AWS Lambda automatically captures telemetry, such as logs, platform metrics, and platform traces. Extensions subscribed to the Telemetry API get this telemetry data directly from the AWS Lambda service in near real time.

The following diagram illustrates the core components of the Extensions API and the Telemetry API. You use Extensions APIs to register your extension with the AWS Lambda service and subscribe to telemetry streams. You also use the Telemetry API to receive the telemetry data in your extension. You can read more about building extensions that use the Telemetry API in the announcement blog, and AWS Lambda service documentation.

Lambda service diagram

There are three types of telemetry streams your extensions can subscribe to

  • Platform telemetry – these are logs, traces, and metrics produced by the AWS Lambda service. This type of telemetry helps you to get insights about execution environment lifecycle, extension lifecycle, and function invocations
  • Function logs – these are custom logs produced by the function code, such as console.log() output
  • Extension logs – similar to function logs, but produced by extensions

Once your extension subscribes to the Telemetry API, as described in the documentation, it starts receiving the telemetry events stream from the AWS Lambda service.

Telemetry API event schema

In order to build the integration with Prometheus and OpenSearch, you need to understand the event schema emitted by the Telemetry API. Here is a sample stream of events that denotes a single Lambda function invocation:

  • The telemetry stream starts with an event of type platform.start. This event type is emitted upon function invocation start. It contains timestamps, request ID, function version, and the tracing segment.
  • The next event in the sequence is platform.runtimeDone. This event type is emitted when the runtime has completed function invocation. This event will have additional information about invocation status, such as success or failure, and metrics, such as invocation duration and produced bytes.
  • The last event has a platform.report type. This event is emitted once invocation has completed, and it provides invocation metrics such as invocation duration, initialization duration, memory usage.
{ "time": "2023-06-02T12:01:23.521Z", "type": "platform.start", "record": { "requestId": "e6b761a9-c52d-415d-b040-7ba94b9452f3", "version": "$LATEST", "tracing": { "spanId": "54565fb41ac79632", "type": "X-Amzn-Trace-Id", "value": "Root=1-62e900b2-710d76f009d6e7785905449a;Parent=0efbd19962d95b05;Sampled=1" } } } { "time": "2023-06-02T12:01:23.521Z", "type": "platform.runtimeDone", "record": { "requestId": "e6b761a9-c52d-415d-b040-7ba94b9452f3", "status": "success", "tracing": { "spanId": "54565fb41ac79632", "type": "X-Amzn-Trace-Id", "value": "Root=1-62e900b2-710d76f009d6e7785905449a;Parent=0efbd19962d95b05;Sampled=1" }, "spans": [ { "name": "responseLatency", "start": "2022-08-02T12:01:23.521Z", "durationMs": 23.02 }, { "name": "responseDuration", "start": "2022-08-02T12:01:23.521Z", "durationMs": 20 } ], "metrics": { "durationMs": 200.0, "producedBytes": 15 } } } { "time": "2023-06-02T12:01:23.521Z", "type": "platform.report", "record": { "requestId": "e6b761a9-c52d-415d-b040-7ba94b9452f3", "metrics": { "durationMs": 220.0, "billedDurationMs": 300, "memorySizeMB": 128, "maxMemoryUsedMB": 90, "initDurationMs": 200.0 }, "tracing": { "spanId": "54565fb41ac79632", "type": "X-Amzn-Trace-Id", "value": "Root=1-62e900b2-710d76f009d6e7785905449a;Parent=0efbd19962d95b05;Sampled=1" } } }

You can see the full list of events and event schemas in the Telemetry API docs.

Dispatching telemetry to Prometheus

Prometheus uses a polling model for retrieving metrics, where it pulls metrics from a service at a regular interval. This works well for long running processes, but it’s not a good fit for the ephemeral nature of Lambda functions. To address this, Prometheus provides a PushGateway component for jobs that are not semantically related to a specific machine or job instance. The Lambda function extension can push metrics through the PushGateway to be picked up by Prometheus. This is illustrated in this diagram:

Prometheus PushGateway diagram

Prometheus supports multiple types of metrics, such as Counters, Gauges, Summaries, and Histograms. When the extension receives a telemetry event, your extension code can parse the event and push the corresponding metric to the PushGateway. At this stage, the extension also adds dimensions to the metrics, and applies labels, such as function name.

Your extension code can collect the results of an individual function invocation from the platform.report event and send it to the PushGateway using the Prometheus Client Library. This sample code records each metric as observations of a Summary metric. Each observation has a set of labels that are used to indicate where the metric was observed, such as the request ID, the function name, and an UUID that was generated when the Lambda execution environment was initialized.

# Define a set of standard labels for each of the Summary metrics
metric_labels = ["request_id", "instance", "function"] # Create Summary metrics for each of the metrics in the platform report object
report_metrics = { "durationMs": Summary('durationMs', 'Duration in ms', registry=registry, labelnames=metric_labels), "billedDurationMs": Summary('billedDurationMs', 'Billed Duration', registry=registry, labelnames=metric_labels), "memorySizeMB": Summary('memorySizeMB', 'Memory Size MB', registry=registry, labelnames=metric_labels), "maxMemoryUsedMB": Summary('maxMemoryUsedMB', 'Max Memory Used MB', registry=registry, labelnames=metric_labels), "initDurationMs": Summary('initDurationMs', 'Init Duration MS', registry=registry, labelnames=metric_labels),
} # Define a Counter to count the total number of events logged
counter = Counter('events', 'Number of events', registry=registry) # Dispatch a batch of metrics to prometheus
def dispatch_to_prometheus(batch): for metric in batch: # Increment the count of all events counter.inc() # Get the event time and increment the associated Counter for that event type metric_type = metric['type'] metric_name = metric_type.replace(".","_") if not metric_type in event_types: # If this is the first time we've seen this event type, create a new Counter event_types[metric_type] = Counter(metric_name, f'Number of {metric_type}', registry=registry, labelnames=["instance", "function"]).labels(instance=UUID, function=FUNCTION_NAME) event_types[metric_type].inc() # For the platform.report type, record the Summary data for each metric if metric_type == "platform.report": record = metric['record'] for summary_name, summary_metric in report_metrics.items(): if summary_name in record["metrics"]: sum_obj = summary_metric.labels(request_id=record['requestId'], instance=UUID, function=FUNCTION_NAME) sum_obj.observe(record["metrics"][summary_name]) # Push all metrics to the PushGateway push_to_gateway(GATEWAY, job=FUNCTION_NAME, registry=registry, handler=gateway_auth_handler, grouping_key={"id":UUID})

Once Prometheus ingests the metrics from PushGateway, you can visualize them using Grafana:

Grafana display

Dispatching telemetry to OpenSearch

OpenSearch can receive events from a Lambda extension and store them in one or more indices. Each index has a mapping which defines the format of the events in that index, including the JSON structure of the document that makes up the event.

Your Lambda extension code can push telemetry to OpenSearch by either directly using the OpenSearch API, or using an OpenSearch client library. By using the bulk operations supported by the OpenSearch API, the Lambda extension can batch telemetry events together and send them to OpenSearch in a single request.

OpenSearch batch request diagram

Since the data contained in the telemetry events differs based on the event type, you can use dynamic mapping for the “record” object. For example, the following operation creates an OpenSearch index with “record” as a dynamic object:

PUT /telemetry { "settings": { "index": { "number_of_shards": 3 } }, "mappings": { "properties": { "time": {"type": "date"}, "type": {"type": "keyword"}, "record": { "dynamic": true, "type": "object" } } }, "aliases": { "sample-alias1": {} }
}

Now you can send a batch of telemetry events to OpenSearch, as demonstrated in this sample code snippet:

# The indices for the various types of events, defined in environment variables
indices = { "platform": os.getenv("PLATFORM_INDEX"), "function": os.getenv("FUNCTION_INDEX"), "extension": os.getenv("EXTENSION_INDEX")
} def dispatch_to_opensearch(batch): # Each action specifies the destination index in the _index field, # and the event is in the doc field actions = [ {"_index": indices[doc['type'].split('.')[0]], "doc": doc} for doc in batch if doc['type'].split('.')[0] in indices.keys()] try: # Use the bulk helper to send a batch of events resp = helpers.bulk(client, actions) except helpers.errors.BulkIndexError as bulk_err: print(f'[telementry_dispatcher] Unable to submit telemetry data batch: {bulk_err}')

After sending the telemetry events to OpenSearch, you can use the OpenSearch Dashboard to query and visualize results.

OpenSearch dashboard screenshot

Sample implementations

Use the following sample extension implementations to get started with forwarding AWS Lambda telemetry data to your open source solutions, such as Prometheus and OpenSearch. Follow the steps in README for setup instructions.

Extensions using the Telemetry API, like other extensions, share the same billing model as Lambda functions. When using Lambda functions with extensions, you pay for requests served, and the combined compute time used to run your code and all extensions, in 1-ms increments. When sending telemetry from your extensions, you might be subject to data transfer costs. To learn more about the billing for extensions, visit the AWS Lambda pricing page.

Security best practices

Extensions run within the same execution environment as the function, so they have the same level of access to resources such as file system, networking, and environment variables. AWS Identity and Access Management (IAM) permissions assigned to the function are shared with extensions. AWS guidance is to assign the least required privileges to your functions.

Always install extensions from a trusted source only. Use Infrastructure as Code (IaC) tools, such as AWS CloudFormation, to simplify the task of attaching the same extension configuration, including IAM permissions, to multiple functions. Additionally, IaC tools allow you to have an audit record of extensions and versions you’ve used previously.

When building extensions, do not log sensitive data. Sanitize payloads and metadata before logging or persisting them for audit purposes.

Conclusion

The AWS Lambda Telemetry API enables Lambda extensions to receive logs, platform traces, and invocation-level metrics directly from the AWS Lambda service. Using this capability developers and operators can collect telemetry from the AWS Lambda service and dispatch it to open source monitoring and observability tools they are already familiar with, such as Prometheus and OpenSearch.

Useful links