AWS Kinesis Firehose for Logs Source
An AWS Kinesis Firehose for Logs Source allows you to ingest CloudWatch logs or any other logs streamed and delivered via Amazon Kinesis Data Firehose.
Amazon Kinesis Data Firehose is an AWS service that can reliably load streaming data into any analytics platform, such as Sumo Logic. It is a fully managed service that automatically scales to match the throughput of data and requires no ongoing administration. With Kinesis Data Firehose, you do not need to write applications or manage resources. You configure your AWS service logs like VPC flow logs to send logs to AWS CloudWatch that can then stream logs to Kinesis Data Firehose which automatically delivers the logs to your Sumo Logic account. This eliminates the need for creating separate log processors or forwarders such as AWS Lambda functions, that are limited by time out, concurrency, and memory limits.
The following diagram shows the flow of data with an AWS Kinesis Firehose for Logs Source:
For failed logs messages, AWS will send them into the backup S3 bucket. Sumo Logic will ingest those failed logs through S3, and not the firehose.
Create an AWS Kinesis Firehose for Logs Source
When you create an AWS Kinesis Firehose for Logs Source, you add it to a Hosted Collector. Before creating the Source, identify the Hosted Collector you want to use or create a new Hosted Collector. For instructions, see Configure a Hosted Collector.
To create an AWS Kinesis Firehose for Logs Source:
-
Classic UI. In the main Sumo Logic menu, select Manage Data > Collection > Collection.
New UI. In the Sumo Logic top menu select Configuration, and then under Data Collection select Collection. You can also click the Go To... menu at the top of the screen and select Collection. -
On the Collectors page, click Add Source next to a Hosted Collector.
-
Select AWS Kinesis Firehose for Logs Source.
-
Enter a Name for the Source. A description is optional.
-
(Optional) The Enable S3 Replay option allows you to collect any logs that were deemed undelivered by Kinesis.
Kinesis puts undelivered logs into a backup directory within your S3 bucket with the path
http-endpoint-failed/yyyy/MM/dd/00/
(00
indicates UTC time zone). This is useful when you need to meet compliance requirements.With this option enabled, you must specify a Path Expression that points to this backup directory. Start with the path prefix to your Kinesis bucket and append
http-endpoint-failed/
to it, followed by a wildcard *. For example,prefix-http-endpoint-failed/*
.You can narrow the path by dates if needed. See Custom Prefixes for Amazon S3 Objects for details on the format of the backup path. The built-in metadata field
_sourceName
is given the object path name from the backup directory.infoDo not use the same bucket and prefix for different Kinesis Sources. We cannot differentiate logs from multiple Kinesis Data Firehoses.
-
(Optional) For Source Host and Source Category, enter any string to tag the output collected from the Source. (Category metadata is stored in a searchable field called _sourceCategory.)
-
SIEM Processing. Check the checkbox to forward your data to Cloud SIEM.
-
Fields. Click the +Add Field