out_webhdfs
TimeSliced Output plugin writes records into HDFS (Hadoop Distributed File System). By default, it creates files on an hourly basis. This means that when you first import records using the plugin, no file is created immediately. The file will be created when the time_slice_format
condition has been met. To change the output frequency, please modify the time_slice_format
value. This document doesn't describe all parameters. If you want to know full features, check the Further Reading section.out_webhdfs
is included in td-agent by default (v1.1.10 or later). Fluentd gem users will have to install the fluent-plugin-webhdfs gem using the following command.webhfds
."#{Socket.gethostname}"
in your path to avoid writing into the same HDFS file from multiple Fluentd instances. This conflict could result in data loss.time_slice_format
section). If path contains time placeholders, webhdfs output configures time_slice_format
automatically with these placeholders.%Y%m%d%H
, which creates one file per hour. This parameter may be overwritten by path
configuration.memory
by default (buf_memory). The file
(buf_file) buffer type can be chosen as well. If you use file
buffer type, buffer_path
parameter is required.memory
buffer type.max_retry_wait
is reached.retry_limit
parameter for details), the sleep interval can be up to approximately 131072 seconds (roughly 36 hours) in the default configurations.retry_limit
is ignored and there is no limit). The default values are 17 and false (not disabled). If the limit is reached, buffered data is discarded and the retry interval is reset to its initial value (retry_wait
).log_level
option allows the user to set different levels of logging for each plugin. The supported log levels are: fatal
, error
, warn
, info
, debug
, and trace
.