Fluentd
1.0
1.0
  • Introduction
  • Overview
    • Life of a Fluentd event
    • Support
    • FAQ
    • Logo
    • fluent-package v5 vs td-agent v4
  • Installation
    • Before Installation
    • Install fluent-package
      • RPM Package (Red Hat Linux)
      • DEB Package (Debian/Ubuntu)
      • .dmg Package (macOS)
      • .msi Installer (Windows)
    • Install calyptia-fluentd
      • RPM Package (Red Hat Linux)
      • DEB Package (Debian/Ubuntu)
      • .dmg Package (macOS)
      • .msi Installer (Windows)
    • Install by Ruby Gem
    • Install from Source
    • Post Installation Guide
    • Obsolete Installation
      • Treasure Agent v4 (EOL) Installation
        • Install by RPM Package v4 (Red Hat Linux)
        • Install by DEB Package v4 (Debian/Ubuntu)
        • Install by .dmg Package v4 (macOS)
        • Install by .msi Installer v4 (Windows)
      • Treasure Agent v3 (EOL) Installation
        • Install by RPM Package v3 (Red Hat Linux)
        • Install by DEB Package v3 (Debian/Ubuntu)
        • Install by .dmg Package v3 (macOS)
        • Install by .msi Installer v3 (Windows)
  • Configuration
    • Config File Syntax
    • Config File Syntax (YAML)
    • Routing Examples
    • Config: Common Parameters
    • Config: Parse Section
    • Config: Buffer Section
    • Config: Format Section
    • Config: Extract Section
    • Config: Inject Section
    • Config: Transport Section
    • Config: Storage Section
    • Config: Service Discovery Section
  • Deployment
    • System Configuration
    • Logging
    • Signals
    • RPC
    • High Availability Config
    • Performance Tuning
    • Multi Process Workers
    • Failure Scenarios
    • Plugin Management
    • Trouble Shooting
    • Fluentd UI
    • Linux Capability
    • Command Line Option
    • Source Only Mode
    • Zero-downtime restart
  • Container Deployment
    • Docker Image
    • Docker Logging Driver
    • Docker Compose
    • Kubernetes
  • Monitoring Fluentd
    • Overview
    • Monitoring by Prometheus
    • Monitoring by REST API
  • Input Plugins
    • tail
    • forward
    • udp
    • tcp
    • unix
    • http
    • syslog
    • exec
    • sample
    • monitor_agent
    • windows_eventlog
  • Output Plugins
    • file
    • forward
    • http
    • exec
    • exec_filter
    • secondary_file
    • copy
    • relabel
    • roundrobin
    • stdout
    • null
    • s3
    • kafka
    • elasticsearch
    • opensearch
    • mongo
    • mongo_replset
    • rewrite_tag_filter
    • webhdfs
    • buffer
  • Filter Plugins
    • record_transformer
    • grep
    • parser
    • geoip
    • stdout
  • Parser Plugins
    • regexp
    • apache2
    • apache_error
    • nginx
    • syslog
    • ltsv
    • csv
    • tsv
    • json
    • msgpack
    • multiline
    • none
  • Formatter Plugins
    • out_file
    • json
    • ltsv
    • csv
    • msgpack
    • hash
    • single_value
    • stdout
    • tsv
  • Buffer Plugins
    • memory
    • file
    • file_single
  • Storage Plugins
    • local
  • Service Discovery Plugins
    • static
    • file
    • srv
  • Metrics Plugins
    • local
  • How-to Guides
    • Stream Analytics with Materialize
    • Send Apache Logs to S3
    • Send Apache Logs to Minio
    • Send Apache Logs to Mongodb
    • Send Syslog Data to Graylog
    • Send Syslog Data to InfluxDB
    • Send Syslog Data to Sematext
    • Data Analytics with Treasure Data
    • Data Collection with Hadoop (HDFS)
    • Simple Stream Processing with Fluentd
    • Stream Processing with Norikra
    • Stream Processing with Kinesis
    • Free Alternative To Splunk
    • Email Alerting like Splunk
    • How to Parse Syslog Messages
    • Cloud Data Logging with Raspberry Pi
  • Language Bindings
    • Java
    • Ruby
    • Python
    • Perl
    • PHP
    • Nodejs
    • Scala
  • Plugin Development
    • How to Write Input Plugin
    • How to Write Base Plugin
    • How to Write Buffer Plugin
    • How to Write Filter Plugin
    • How to Write Formatter Plugin
    • How to Write Output Plugin
    • How to Write Parser Plugin
    • How to Write Storage Plugin
    • How to Write Service Discovery Plugin
    • How to Write Tests for Plugin
    • Configuration Parameter Types
    • Upgrade Plugin from v0.12
  • Plugin Helper API
    • Plugin Helper: Child Process
    • Plugin Helper: Compat Parameters
    • Plugin Helper: Event Emitter
    • Plugin Helper: Event Loop
    • Plugin Helper: Extract
    • Plugin Helper: Formatter
    • Plugin Helper: Inject
    • Plugin Helper: Parser
    • Plugin Helper: Record Accessor
    • Plugin Helper: Server
    • Plugin Helper: Socket
    • Plugin Helper: Storage
    • Plugin Helper: Thread
    • Plugin Helper: Timer
    • Plugin Helper: Http Server
    • Plugin Helper: Service Discovery
  • Troubleshooting Guide
  • Appendix
    • Update from v0.12 to v1
    • td-agent v2 vs v3 vs v4
Powered by GitBook
On this page
  • Background
  • Mechanism
  • Prerequisites
  • Install Kinesis Plugin
  • Configuration
  • Tail Input
  • Amazon Kinesis Output
  • Test
  • FAQs
  • Conclusion
  • Learn More

Was this helpful?

  1. How-to Guides

Stream Processing with Kinesis

PreviousStream Processing with NorikraNextFree Alternative To Splunk

Last updated 4 months ago

Was this helpful?

This article explains how to use 's Output plugin () to aggregate semi-structured logs in real-time. Kinesis plugin is developed and published by Amazon Web Services officially.

Background

Amazon Kinesis is a platform for streaming data on AWS, offering powerful services to make it easy to load and analyze streaming data, and also providing the ability for you to build custom streaming data applications for specialized needs.

Mechanism

In this example, Fluentd does three (3) things:

  1. It continuously "tails" the access log file.

  2. It parses the incoming log entries into meaningful fields (such as

    ip, path, etc.) and buffers them.

  3. It writes the buffered data to Amazon Kinesis periodically.

Prerequisites

The following software/services are required to be set up correctly:

You can install Fluentd via major packaging systems.

Install Kinesis Plugin

If out_kinesis_streams (fluent-plugin-kinesis) is not installed yet, please install it manually.

Configuration

Let's start configuring Fluentd. If you used the deb/rpm package, Fluentd's config file is located at /etc/fluent/fluentd.conf.

Tail Input

For the input source, we will set up Fluentd to track the recent Apache logs (typically found at /var/log/apache2/access_log). The Fluentd configuration file should look like this:

<source>
  @type tail
  path /var/log/apache2/access_log
  pos_file /var/log/fluent/apache2.access_log.pos
  <parse>
    @type apache2
  </parse>
  tag kinesis.apache.access
</source>

Let's go through the configuration line by line:

  1. @type tail: The tail Input plugin continuously tracks the log

    file. This handy plugin is included in Fluentd's core.

  2. @type apache2 in <parse>: Uses Fluentd's built-in Apache log parser.

  3. path /var/log/apache2/access_log: The location of the Apache log.

    This may be different for your particular system.

  4. tag kinesis.apache.access: kinesis.apache.access is used as the

    tag to route the messages within Fluentd.

That's it! You should now be able to output a JSON-formatted data stream for Fluentd to process.

Amazon Kinesis Output

The output destination will be Amazon Kinesis. The output configuration should look like this:

<match **>
  # plugin type
  @type kinesis_streams

  # your kinesis stream name
  stream_name <KINESIS_STREAM_NAME>

  # AWS credentials
  aws_key_id <AWS_KEY_ID>
  aws_sec_key <AWS_SECRET_KEY>

  # AWS region
  region us-east-1

  # Use random value for the partition key
  random_partition_key true

  <buffer>
    # Frequency of ingestion
    flush_interval 5s
    # Parallelism of ingestion
    flush_thread_count
  </buffer>
</match>

The match section specifies the regexp used to look for matching tags. If a matching tag is found in a log, then the config inside <match>...</match> is used (i.e. the log is routed according to the config inside). In this example, the kinesis.apache.access tag (generated by tail) is always used.

The ** in match.** matches zero or more period-delimited tag parts (e.g. match/match.a/match.a.b).

The flush_interval parameter specifies how often the data is written to Kinesis.

Test

Restart fluentd to make sure that the configuration change is available:

# systemd
$ sudo systemctl restart fluentd

To test the configuration, just have a couple of accesses to your Apache server. This example uses the ab (Apache Bench) program:

$ ab -n 100 -c 10 http://localhost/

FAQs

Why we need Fluentd, while Kinesis also offers client libraries?

Conclusion

Fluentd with Amazon Kinesis makes the realtime log collection simple, easy, and robust.

Learn More

is an advanced open-source log collector originally developed at . Because Fluentd can collect logs from various sources, is one of the popular destinations for the output.

This article will show you how to use to import Apache logs into Amazon Kinesis.

(with the Combined Log Format)

See section how to install fluent-plugin-kinesis on your environment.

Please make sure that your Apache outputs are in the default combined format. format apache2 cannot parse custom log formats. Please see the article for more information.

The random_partition_key true option will generate the partition key via UUID v3 (). Kinesis Stream consists of shards, and the processing power of each shard is limited. This partition key will be used by Kinesis to determine which shard has been designated for a specific record.

For additional configurations, see .

For those who are interested in security, all communication between Fluentd and Amazon Kinesis are done via HTTPS. If you do not want to have AES keys in the configuration file, is available too for EC2 nodes.

A lot of people use Fluentd + Kinesis, simply because they want to have more choices for and . For inputs, Fluentd has a lot more community-contributed plugins and libraries. For outputs, you can send not only Kinesis, but multiple destinations like Amazon S3, local file storage, etc.

(Made by Amazon Web Services)

If this article is incorrect or outdated, or omits critical information, please . is an open-source project under . All components are available under the Apache 2 License.

Fluentd
Treasure Data, Inc
Amazon Kinesis
Fluentd
Fluentd
Apache
Kinesis Output plugin
Installation
Plugin Management
in_tail
source
Kinesis Output plugin
IAM Role-based authentication
inputs
outputs
Fluentd Architecture
Fluentd Get Started
Amazon Kinesis
Amazon Kinesis Output Plugin
let us know
Fluentd
Cloud Native Computing Foundation (CNCF)
Fluentd
Amazon Kinesis
out_kinesis
Fluentd + Kinesis