title: Monitor Azure Functions description: Learn how to use Azure Application Insights with Azure Functions for monitoring function execution. services: functions author: tdykstra manager: cfowler editor: '' tags: '' keywords: azure functions, functions, event processing, webhooks, dynamic compute, serverless architecture ms.assetid: 501722c3-f2f7-4224-a220-6d59da08a320 ms.service: functions ms.devlang: multiple ms.topic: article ms.tgt_pltfrm: multiple ms.workload: na ms.date: 09/15/2017 ms.author: tdykstra
Azure Functions offers built-in integration with Azure Application Insights for monitoring functions. This article shows how to configure Functions to send telemetry data to Application Insights.
Functions also has built-in monitoring that doesn't use Application Insights. We recommend Application Insights because it offers more data and better ways to analyze the data. For information about the built-in monitoring, see the last section of this article.
For a function app to send data to Application Insights, it needs to know the instrumentation key of an Application Insights instance. There are two ways to make that connection in the Azure portal:
- Create a connected Application Insights instance when you create the function app.
- Connect an Application Insights instance to an existing function app.
Enable Application Insights on the Function App Create page:
Get an instrumentation key and save it in a function app:
-
Create the Application Insights instance. Set application type to General.
-
Copy the instrumentation key from the Essentials page of the Application Insights instance. Hover over the end of the displayed key value to get a Click to copy button.
-
In the function app's Application settings page, add an app setting named APPINSIGHTS_INSTRUMENTATIONKEY and paste the instrumentation key.
-
Click Save.
To navigate to Application Insights from a function app in the portal, select the Application Insights link on the function app's Overview page.
For information about how to use Application Insights, see the Application Insights documentation. This section shows some examples of how to view data in Application Insights. If you are already familiar with Application Insights, you can go directly to the sections about configuring and customizing the telemetry data.
In Metrics Explorer, you can create charts and alerts based on metrics such as number of function invocations, execution time, and success rate.
On the Failures tab, you can create charts and alerts based on function failures and server exceptions. The Operation Name is the function name. Failures in dependencies are not shown unless you implement custom telemetry for dependencies.
On the Performance tab, you can analyze performance issues.
The Servers tab shows resource utilization and throughput per server. This data can be useful for debugging scenarios where functions are bogging down your underlying resources. Servers are referred to as Cloud role instances.
The Live Metrics Stream tab shows metrics data as it is created in real time.
Application Insights Analytics gives you access to all of the telemetry data in the form of tables in a database. Analytics provides a query language for extracting and manipulating the data.
Here's a query example. This one shows the distribution of requests per worker over the last 30 minutes.
requests
| where timestamp > ago(30m)
| summarize count() by cloud_RoleInstance, bin(timestamp, 1m)
| render timechart
The tables that are available are shown in the Schema tab of the left pane. You can find data generated by function invocations in the following tables:
- traces - Logs created by the runtime and by function code.
- requests - One for each function invocation.
- exceptions - Any exceptions thrown by the runtime.
- customMetrics - Count of successful and failing invocations, success rate, duration.
- customEvents - Events tracked by the runtime, for example: HTTP requests that trigger a function.
- performanceCounters - Info about the performance of the servers that the functions are running on.
The other tables are for availability tests and client/browser telemetry. You can implement custom telemetry to add data to them.
Within each table, some of the Functions-specific data is in a customDimensions
field. For example, the following query retrieves all traces that have log level Error
.
traces
| where customDimensions.LogLevel == "Error"
The runtime provides customDimensions.LogLevel
and customDimensions.Category
. You can provide additional fields in logs you write in your function code. See Structured logging later in this article.
You can use Application Insights without any custom configuration, but the default configuration can result in high volumes of data. If you're using a Visual Studio Azure subscription, you might hit your data cap for App Insights. The remainder of this article shows how to configure and customize the data that your functions send to Application Insights.
The Azure Functions logger includes a category for every log. The category indicates which part of the runtime code or your function code wrote the log.
The Functions runtime creates logs that have a category beginning with "Host". For example, the "function started," "function executed," and "function completed" logs have category "Host.Executor".
If you write logs in your function code, their category is "Function".
The Azure functions logger also includes a log level with every log. LogLevel is an enumeration, and the integer code indicates relative importance:
LogLevel | Code |
---|---|
Trace | 0 |
Debug | 1 |
Information | 2 |
Warning | 3 |
Error | 4 |
Critical | 5 |
None | 6 |
Log level None
is explained in the next section.
The host.json file configures how much logging a function app sends to Application Insights. For each category, you indicate the minimum log level to send. Here's an example:
{
"logger": {
"categoryFilter": {
"defaultLevel": "Information",
"categoryLevels": {
"Host.Results": "Error",
"Function": "Error",
"Host.Aggregator": "Information"
}
}
}
}
This example sets up the following rules:
- For logs with category "Host.Results" or "Function", send only
Error
level and above to Application Insights. Logs forInformation
level and below are ignored. - For logs with category Host. Aggregator, send only
Information
level and above to Application Insights. Logs forDebug
level and below are ignored. - For all other logs, send only
Information
level and above to Application Insights.
The category value in host.json controls logging for all categories that begin with the same value. For example, "Host" in host.json controls logging for "Host.General", "Host.Executor", "Host.Results", and so forth.
If host.json includes multiple categories that start with the same string, the longer ones are matched first. For example, suppose you want everything from the runtime except "Host.Aggregator" to log at Error
level, while "Host.Aggregator" logs at Information
level:
{
"logger": {
"categoryFilter": {
"defaultLevel": "Information",
"categoryLevels": {
"Host": "Error",
"Function": "Error",
"Host.Aggregator": "Information"
}
}
}
}
To suppress all logs for a category, you can use log level None
. No logs are written with that category and there is no log level above it.
The following sections describe the main categories of logs that the runtime creates.
These logs show as "requests" in Application Insights. They indicate success or failure of a function.
All of these logs are written at Information
level, so if you filter at Warning
or above, you won't see any of this data.
These logs provide counts and averages of function invocations over a configurable period of time. The default period is 30 seconds or 1,000 results, whichever comes first.
The logs show as "customMetrics" in Application Insights. Examples are number of runs, success rate, and duration.
All of these logs are written at Information
level, so if you filter at Warning
or above, you won't see any of this data.
All logs for categories other than the ones already listed show as "traces" in Application Insights.
All logs with categories that begin with "Host" are written by the Functions runtime. The "Function started" and "Function completed" logs have category "Host.Executor". For successful runs, these logs are Information
level; exceptions are logged at Error
level. The runtime also creates Warning
level logs, for example: queue messages sent to the poison queue.
Logs written by your function code have category "Function" and may be any log level.
As noted in the previous section, the runtime aggregates data about function executions over a period of time. The default period is 30 seconds or 1,000 runs, whichever comes first. You can configure this setting in the host.json file. Here's an example:
{
"aggregator": {
"batchSize": 1000,
"flushTimeout": "00:00:30"
}
}
Application Insights has a sampling feature that can protect you from producing too much telemetry data at times of peak load. When the number of telemetry items exceeds a specified rate, Application Insights starts to randomly ignore some of the incoming items. You can configure sampling in host.json. Here's an example:
{
"applicationInsights": {
"sampling": {
"isEnabled": true,
"maxTelemetryItemsPerSecond" : 5
}
}
}
You can write logs in your function code that appear as traces in Application Insights.
Use an ILogger parameter in your functions instead of a TraceWriter
parameter. Logs created by using TraceWriter
do go to Application Insights, but ILogger
lets you do structured logging.
With an ILogger
object you call Log<level>
extension methods on ILogger to create logs. For example, the following code writes Information
logs with category "Function".
public static async Task<HttpResponseMessage> Run(HttpRequestMessage req, ILogger logger)
{
logger.LogInformation("Request for item with key={itemKey}.", id);
The order of placeholders, not their names, determines which parameters are used in the log message. For example, suppose you have the following code:
string partitionKey = "partitionKey";
string rowKey = "rowKey";
logger.LogInformation("partitionKey={partitionKey}, rowKey={rowKey}", partitionKey, rowKey);
If you keep the same message string and reverse the order of the parameters, the resulting message text would have the values in the wrong places.
Placeholders are handled this way so that you can do structured logging. Application Insights stores the parameter name-value pairs in addition to the message string. The result is that the message arguments become fields that you can query on.
For example, if your logger method call looks like the previous example, you could query the field customDimensions.prop__rowKey
. The prefix is added to ensure that there are no collisions between fields the runtime adds and fields your function code adds.
You can also query on the original message string by referencing the field customDimensions.prop__{OriginalFormat}
.
Here's a sample JSON representation of customDimensions
data:
{
customDimensions: {
"prop__{OriginalFormat}":"C# Queue trigger function processed: {message}",
"Category":"Function",
"LogLevel":"Information",
"prop__message":"c9519cbf-b1e6-4b9b-bf24-cb7d10b1bb89"
}
}
In C# script functions, you can use the LogMetric
extension method on ILogger
to create custom metrics in Application Insights. Here's a sample method call:
logger.LogMetric("TestMetric", 1234);
This code is an alternative to calling TrackMetric
using the Application Insights API for .NET.
In Node.js functions, use context.log
to write logs. Structured logging is not enabled.
context.log('JavaScript HTTP trigger function processed a request.' + context.invocationId);
In Node.js functions, you can use the context.log.metric
method to create custom metrics in Application Insights. Here's a sample method call:
context.log.metric("TestMetric", 1234);
This code is an alternative to calling trackMetric
using the Node.js SDK for Application Insights.
You can use the Microsoft.ApplicationInsights NuGet package to send custom telemetry data to Application Insights.
Here's an example of C# code that uses the custom telemetry API. The example is for a .NET class library, but the Application Insights code is the same for C# script.
using System;
using System.Net;
using Microsoft.ApplicationInsights;
using Microsoft.ApplicationInsights.Extensibility;
using Microsoft.Azure.WebJobs;
using System.Net.Http;
using System.Threading.Tasks;
using Microsoft.Azure.WebJobs.Extensions.Http;
using Microsoft.Extensions.Logging;
using System.Linq;
namespace functionapp0915
{
public static class HttpTrigger2
{
private static string key = TelemetryConfiguration.Active.InstrumentationKey =
System.Environment.GetEnvironmentVariable(
"APPINSIGHTS_INSTRUMENTATIONKEY", EnvironmentVariableTarget.Process);
private static TelemetryClient telemetry =
new TelemetryClient() { InstrumentationKey = key };
[FunctionName("HttpTrigger2")]
public static async Task<HttpResponseMessage> Run(
[HttpTrigger(AuthorizationLevel.Anonymous, "get", "post", Route = null)]
HttpRequestMessage req, ExecutionContext context, ILogger log)
{
log.LogInformation("C# HTTP trigger function processed a request.");
DateTime start = DateTime.UtcNow;
// parse query parameter
string name = req.GetQueryNameValuePairs()
.FirstOrDefault(q => string.Compare(q.Key, "name", true) == 0)
.Value;
// Get request body
dynamic data = await req.Content.ReadAsAsync<object>();
// Set name to query string or body data
name = name ?? data?.name;
telemetry.Context.Operation.Id = context.InvocationId.ToString();
telemetry.Context.Operation.Name = "cs-http";
if (!String.IsNullOrEmpty(name))
{
telemetry.Context.User.Id = name;
}
telemetry.TrackEvent("Function called");
telemetry.TrackMetric("Test Metric", DateTime.Now.Millisecond);
telemetry.TrackDependency("Test Dependency",
"swapi.co/api/planets/1/",
start, DateTime.UtcNow - start, true);
return name == null
? req.CreateResponse(HttpStatusCode.BadRequest,
"Please pass a name on the query string or in the request body")
: req.CreateResponse(HttpStatusCode.OK, "Hello " + name);
}
}
}
Don't call TrackRequest
or StartOperation<RequestTelemetry>
, because you'll see duplicate requests for a function invocation. The Functions runtime automatically tracks requests.
Set telemetry.Context.Operation.Id
to the invocation ID each time your function is started. This makes it possible to correlate all telemetry items for a given function invocation.
telemetry.Context.Operation.Id = context.InvocationId.ToString();
The Application Insights Node.js SDK is currently in beta. Here's some sample code that sends custom telemetry to Application Insights:
const appInsights = require("applicationinsights");
appInsights.setup();
const client = appInsights.defaultClient;
module.exports = function (context, req) {
context.log('JavaScript HTTP trigger function processed a request.');
client.trackEvent({name: "my custom event", tagOverrides:{"ai.operation.id": context.invocationId}, properties: {customProperty2: "custom property value"}});
client.trackException({exception: new Error("handled exceptions can be logged with this method"), tagOverrides:{"ai.operation.id": context.invocationId}});
client.trackMetric({name: "custom metric", value: 3, tagOverrides:{"ai.operation.id": context.invocationId}});
client.trackTrace({message: "trace message", tagOverrides:{"ai.operation.id": context.invocationId}});
client.trackDependency({target:"http://dbname", name:"select customers proc", data:"SELECT * FROM Customers", duration:231, resultCode:0, success: true, dependencyTypeName: "ZSQL", tagOverrides:{"ai.operation.id": context.invocationId}});
client.trackRequest({name:"GET /customers", url:"http://myserver/customers", duration:309, resultCode:200, success:true, tagOverrides:{"ai.operation.id": context.invocationId}});
if (req.query.name || (req.body && req.body.name)) {
context.res = {
// status: 200, /* Defaults to 200 */
body: "Hello " + (req.query.name || req.body.name)
};
}
else {
context.res = {
status: 400,
body: "Please pass a name on the query string or in the request body"
};
}
context.done();
};
The tagOverrides
parameter sets operation_Id
to the function's invocation ID. This setting enables you to correlate all of the automatically-generated and custom telemetry for a given function invocation.
Dependencies don't show up automatically, but you can write custom code to show dependencies. The sample code in the C# custom telemetry section shows how. The sample code results in an application map in Application Insights that looks like this:
To report an issue with Application Insights integration in Functions, or to make a suggestion or request, create an issue in GitHub.
We recommend Application Insights for monitoring functions because it offers more data and better ways to analyze the data. But you can also find telemetry and logging data in the Azure portal pages for a Function app.
Select the Monitor tab for a function and you get a list of function executions. Select a function execution to review the duration, input data, errors, and associated log files.
Important
When using the Consumption hosting plan for Azure Functions, the Monitoring tile in the Function App does not show any data. This is because the platform dynamically scales and manages compute instances for you. These metrics are not meaningful on a Consumption plan.
Real-time monitoring is available by clicking live event stream on the function Monitor tab. The live event stream is displayed in a graph in a new tab in the browser
Note
There is a known issue that may cause your data to fail to be populated. You may need to close the browser tab containing the live event stream and then click live event stream again to allow it to properly populate your event stream data.
These statistics are real-time but the actual graphing of the execution data may have around 10 seconds of latency.
You can stream log files to a command-line session on a local workstation using the Azure Command Line Interface (CLI) 1.0 or PowerShell.
To get started, install the Azure CLI 1.0 and sign in to Azure.
Use the following commands to enable classic Service Management mode, choose your subscription, and stream log files:
azure config mode asm
azure account list
azure account set <subscriptionNameOrId>
azure site log tail -v <function app name>
To get started, install and configure Azure PowerShell.
Use the following commands to add your Azure account, choose your subscription, and stream log files:
PS C:\> Add-AzureAccount
PS C:\> Get-AzureSubscription
PS C:\> Get-AzureSubscription -SubscriptionName "MyFunctionAppSubscription" | Select-AzureSubscription
PS C:\> Get-AzureWebSiteLog -Name MyFunctionApp -Tail
For more information, see How to: Stream logs for web apps.
[!div class="nextstepaction"] Learn more about Application Insights
[!div class="nextstepaction"] Learn more about the logging framework that Functions uses