APT and RPM repositories at repositories.timber.io will be decommissioned on February 28th Migration instructions

AWS S3

Store observability events in the AWS S3 object storage system

status: stable delivery: at-least-once acknowledgements: yes egress: batch state: stateless

Configuration

Example configurations

{
  "sinks": {
    "my_sink_id": {
      "type": "aws_s3",
      "inputs": [
        "my-source-or-transform-id"
      ],
      "bucket": "my-bucket"
    }
  }
}
[sinks.my_sink_id]
type = "aws_s3"
inputs = [ "my-source-or-transform-id" ]
bucket = "my-bucket"
sinks:
  my_sink_id:
    type: aws_s3
    inputs:
      - my-source-or-transform-id
    bucket: my-bucket
{
  "sinks": {
    "my_sink_id": {
      "type": "aws_s3",
      "inputs": [
        "my-source-or-transform-id"
      ],
      "acl": "authenticated-read",
      "bucket": "my-bucket",
      "compression": "gzip",
      "content_encoding": "gzip",
      "content_type": "application/gzip",
      "endpoint": "http://127.0.0.0:5000/path/to/service",
      "filename_append_uuid": true,
      "filename_extension": "json",
      "filename_time_format": "%s",
      "grant_full_control": "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be",
      "grant_read": "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be",
      "grant_read_acp": "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be",
      "grant_write_acp": "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be",
      "key_prefix": "date=%F",
      "region": "us-east-1",
      "server_side_encryption": "AES256",
      "ssekms_key_id": "abcd1234",
      "storage_class": "STANDARD",
      "tags": {
        "Classification": "confidential",
        "PHI": "True",
        "Project": "Blue"
      },
      "timezone": "local"
    }
  }
}
[sinks.my_sink_id]
type = "aws_s3"
inputs = [ "my-source-or-transform-id" ]
acl = "authenticated-read"
bucket = "my-bucket"
compression = "gzip"
content_encoding = "gzip"
content_type = "application/gzip"
endpoint = "http://127.0.0.0:5000/path/to/service"
filename_append_uuid = true
filename_extension = "json"
filename_time_format = "%s"
grant_full_control = "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
grant_read = "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
grant_read_acp = "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
grant_write_acp = "79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
key_prefix = "date=%F"
region = "us-east-1"
server_side_encryption = "AES256"
ssekms_key_id = "abcd1234"
storage_class = "STANDARD"
timezone = "local"

  [sinks.my_sink_id.tags]
  Classification = "confidential"
  PHI = "True"
  Project = "Blue"
sinks:
  my_sink_id:
    type: aws_s3
    inputs:
      - my-source-or-transform-id
    acl: authenticated-read
    bucket: my-bucket
    compression: gzip
    content_encoding: gzip
    content_type: application/gzip
    endpoint: http://127.0.0.0:5000/path/to/service
    filename_append_uuid: true
    filename_extension: json
    filename_time_format: "%s"
    grant_full_control: 79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be
    grant_read: 79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be
    grant_read_acp: 79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be
    grant_write_acp: 79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be
    key_prefix: date=%F
    region: us-east-1
    server_side_encryption: AES256
    ssekms_key_id: abcd1234
    storage_class: STANDARD
    tags:
      Classification: confidential
      PHI: "True"
      Project: Blue
    timezone: local

acknowledgements

optional object

Controls how acknowledgements are handled for this sink.

See End-to-end Acknowledgements for more information on how event acknowledgement is handled.

Whether or not end-to-end acknowledgements are enabled.

When enabled for a sink, any source connected to that sink, where the source supports end-to-end acknowledgements as well, waits for events to be acknowledged by the sink before acknowledging them at the source.

Enabling or disabling acknowledgements at the sink level takes precedence over any global acknowledgements configuration.

acl

optional string literal enum

Canned ACL to apply to the created objects.

For more information, see Canned ACL.

Enum options string literal
OptionDescription
authenticated-read

Bucket/object can be read by authenticated users.

The bucket/object owner is granted the FULL_CONTROL permission, and anyone in the AuthenticatedUsers grantee group is granted the READ permission.

aws-exec-read

Bucket/object are private, and readable by EC2.

The bucket/object owner is granted the FULL_CONTROL permission, and the AWS EC2 service is granted the READ permission for the purpose of reading Amazon Machine Image (AMI) bundles from the given bucket.

bucket-owner-full-control

Object is semi-private.

Both the object owner and bucket owner are granted the FULL_CONTROL permission.

Only relevant when specified for an object: this canned ACL is otherwise ignored when specified for a bucket.

bucket-owner-read

Object is private, except to the bucket owner.

The object owner is granted the FULL_CONTROL permission, and the bucket owner is granted the READ permission.

Only relevant when specified for an object: this canned ACL is otherwise ignored when specified for a bucket.

log-delivery-write

Bucket can have logs written.

The LogDelivery grantee group is granted WRITE and READ_ACP permissions.

Only relevant when specified for a bucket: this canned ACL is otherwise ignored when specified for an object.

For more information about logs, see Amazon S3 Server Access Logging.

private

Bucket/object are private.

The bucket/object owner is granted the FULL_CONTROL permission, and no one else has access.

This is the default.

public-read

Bucket/object can be read publicly.

The bucket/object owner is granted the FULL_CONTROL permission, and anyone in the AllUsers grantee group is granted the READ permission.

public-read-write

Bucket/object can be read and written publicly.

The bucket/object owner is granted the FULL_CONTROL permission, and anyone in the AllUsers grantee group is granted the READ and WRITE permissions.

This is generally not recommended.

auth

optional object
Configuration of the authentication strategy for interacting with AWS services.

auth.access_key_id

required string literal
The AWS access key ID.
Examples
"AKIAIOSFODNN7EXAMPLE"

auth.assume_role

required string literal
The ARN of an IAM role to assume.
Examples
"arn:aws:iam::123456789098:role/my_role"

auth.credentials_file

required string literal
Path to the credentials file.
Examples
"/my/aws/credentials"

auth.external_id

optional string literal
The optional unique external ID in conjunction with role to assume.
Examples
"randomEXAMPLEidString"

auth.imds

optional object
Configuration for authenticating with AWS through IMDS.
Connect timeout for IMDS.
default: 1 (seconds)
Number of IMDS retries for fetching tokens and metadata.
default: 4
Read timeout for IMDS.
default: 1 (seconds)

Timeout for successfully loading any credentials, in seconds.

Relevant when the default credentials chain or assume_role is used.

Examples
30

auth.profile

optional string literal

The credentials profile to use.

Used to select AWS credentials from a provided credentials file.

Examples
"develop"
default: default

auth.region

optional string literal

The AWS region to send STS requests to.

If not set, this defaults to the configured region for the service itself.

Examples
"us-west-2"

auth.secret_access_key

required string literal
The AWS secret access key.
Examples
"wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"

batch

optional object
Event batching behavior.

batch.max_bytes

optional uint

The maximum size of a batch that is processed by a sink.

This is based on the uncompressed size of the batched events, before they are serialized/compressed.

default: 1e+07 (bytes)

batch.max_events

optional uint
The maximum size of a batch before it is flushed.

batch.timeout_secs

optional float
The maximum age of a batch before it is flushed.
default: 300 (seconds)

bucket

required string literal

The S3 bucket name.

This must not include a leading s3:// or a trailing /.

Examples
"my-bucket"

buffer

optional object

Configures the buffering behavior for this sink.

More information about the individual buffer types, and buffer behavior, can be found in the Buffering Model section.

buffer.max_events

optional uint
The maximum number of events allowed in the buffer.
Relevant when: type = "memory"
default: 500

buffer.max_size

required uint

The maximum size of the buffer on disk.

Must be at least ~256 megabytes (268435488 bytes).

Relevant when: type = "disk"

buffer.type

optional string literal enum
The type of buffer to use.
Enum options
OptionDescription
disk

Events are buffered on disk.

This is less performant, but more durable. Data that has been synchronized to disk will not be lost if Vector is restarted forcefully or crashes.

Data is synchronized to disk every 500ms.

memory

Events are buffered in memory.

This is more performant, but less durable. Data will be lost if Vector is restarted forcefully or crashes.

default: memory

buffer.when_full

optional string literal enum
Event handling behavior when a buffer is full.
Enum options
OptionDescription
block

Wait for free space in the buffer.

This applies backpressure up the topology, signalling that sources should slow down the acceptance/consumption of events. This means that while no data is lost, data will pile up at the edge.

drop_newest

Drops the event instead of waiting for free space in buffer.

The event will be intentionally dropped. This mode is typically used when performance is the highest priority, and it is preferable to temporarily lose events rather than cause a slowdown in the acceptance/consumption of events.

default: block

compression

optional string literal enum

Compression configuration.

All compression algorithms use the default compression level unless otherwise specified.

Some cloud storage API clients and browsers handle decompression transparently, so depending on how they are accessed, files may not always appear to be compressed.

Enum options string literal
OptionDescription
gzipGzip compression.
noneNo compression.
snappySnappy compression.
zlibZlib compression.
zstdZstandard compression.
default: gzip

content_encoding

optional string literal

Overrides what content encoding has been applied to the object.

Directly comparable to the Content-Encoding HTTP header.

If not specified, the compression scheme used dictates this value.

Examples
"gzip"

content_type

optional string literal

Overrides the MIME type of the object.

Directly comparable to the Content-Type HTTP header.

If not specified, the compression scheme used dictates this value. When compression is set to none, the value text/x-log is used.

Examples
"application/gzip"

encoding

required object
Configures how events are encoded into raw bytes.

encoding.avro

required object
Apache Avro-specific encoder options.
Relevant when: codec = "avro"
encoding.avro.schema
required string literal
The Avro schema.
Examples
"{ \"type\": \"record\", \"name\": \"log\", \"fields\": [{ \"name\": \"message\", \"type\": \"string\" }] }"

encoding.codec

required string literal enum
The codec to use for encoding events.
Enum options
OptionDescription
avroEncodes an event as an Apache Avro message.
csv

Encodes an event as a CSV message.

This codec must be configured with fields to encode.

gelf

Encodes an event as a GELF message.

This codec is experimental for the following reason:

The GELF specification is more strict than the actual Graylog receiver. Vector’s encoder currently adheres more strictly to the GELF spec, with the exception that some characters such as @ are allowed in field names.

Other GELF codecs such as Loki’s, use a Go SDK that is maintained by Graylog, and is much more relaxed than the GELF spec.

Going forward, Vector will use that Go SDK as the reference implementation, which means the codec may continue to relax the enforcement of specification.

jsonEncodes an event as JSON.
logfmtEncodes an event as a logfmt message.
native

Encodes an event in the native Protocol Buffers format.

This codec is experimental.

native_json

Encodes an event in the native JSON format.

This codec is experimental.

protobufEncodes an event as a Protobuf message.
raw_message

No encoding.

This encoding uses the message field of a log event.

Be careful if you are modifying your log events (for example, by using a remap transform) and removing the message field while doing additional parsing on it, as this could lead to the encoding emitting empty strings for the given event.

text

Plain text encoding.

This encoding uses the message field of a log event. For metrics, it uses an encoding that resembles the Prometheus export format.

Be careful if you are modifying your log events (for example, by using a remap transform) and removing the message field while doing additional parsing on it, as this could lead to the encoding emitting empty strings for the given event.

Examples
"avro"
"csv"
"gelf"
"json"
"logfmt"
"native"
"native_json"
"protobuf"
"raw_message"
"text"

encoding.csv

required object
The CSV Serializer Options.
Relevant when: codec = "csv"
Set the capacity (in bytes) of the internal buffer used in the CSV writer. This defaults to a reasonable setting.
default: 8192
The field delimiter to use when writing CSV.
default: 44

Enable double quote escapes.

This is enabled by default, but it may be disabled. When disabled, quotes in field data are escaped instead of doubled.

default: true
encoding.csv.escape
optional uint

The escape character to use when writing CSV.

In some variants of CSV, quotes are escaped using a special escape character like \ (instead of escaping quotes by doubling them).

To use this, double_quotes needs to be disabled as well otherwise it is ignored.

default: 34
encoding.csv.fields
required [string]

Configures the fields that will be encoded, as well as the order in which they appear in the output.

If a field is not present in the event, the output will be an empty string.

Values of type Array, Object, and Regex are not supported and the output will be an empty string.

encoding.csv.quote
optional uint
The quote character to use when writing CSV.
default: 34
encoding.csv.quote_style
optional string literal enum
The quoting style to use when writing CSV data.
Enum options
OptionDescription
alwaysAlways puts quotes around every field.
necessaryPuts quotes around fields only when necessary. They are necessary when fields contain a quote, delimiter, or record terminator. Quotes are also necessary when writing an empty record (which is indistinguishable from a record with one empty field).
neverNever writes quotes, even if it produces invalid CSV data.
non_numericPuts quotes around all fields that are non-numeric. Namely, when writing a field that does not parse as a valid float or integer, then quotes are used even if they aren’t strictly necessary.
default: necessary

encoding.except_fields

optional [string]
List of fields that are excluded from the encoded event.

encoding.metric_tag_values

optional string literal enum

Controls how metric tag values are encoded.

When set to single, only the last non-bare value of tags are displayed with the metric. When set to full, all metric tags are exposed as separate assignments.

Relevant when: codec = "json" or codec = "text"
Enum options
OptionDescription
fullAll tags are exposed as arrays of either string or null values.
singleTag values are exposed as single strings, the same as they were before this config option. Tags with multiple values show the last assigned value, and null values are ignored.
default: single

encoding.only_fields

optional [string]
List of fields that are included in the encoded event.

encoding.protobuf

required object
Options for the Protobuf serializer.
Relevant when: codec = "protobuf"
encoding.protobuf.desc_file
required string literal

The path to the protobuf descriptor set file.

This file is the output of protoc -o <path> ...

Examples
"/etc/vector/protobuf_descriptor_set.desc"
encoding.protobuf.message_type
required string literal
The name of the message type to use for serializing.
Examples
"package.Message"

encoding.timestamp_format

optional string literal enum
Format used for timestamp fields.
Enum options
OptionDescription
rfc3339Represent the timestamp as a RFC 3339 timestamp.
unixRepresent the timestamp as a Unix timestamp.
unix_floatRepresent the timestamp as a Unix timestamp in floating point.
unix_msRepresent the timestamp as a Unix timestamp in milliseconds.
unix_nsRepresent the timestamp as a Unix timestamp in nanoseconds.
unix_usRepresent the timestamp as a Unix timestamp in microseconds

endpoint

optional string literal
Custom endpoint for use with AWS-compatible services.
Examples
"http://127.0.0.0:5000/path/to/service"

filename_append_uuid

optional bool

Whether or not to append a UUID v4 token to the end of the object key.

The UUID is appended to the timestamp portion of the object key, such that if the object key generated is date=2022-07-18/1658176486, setting this field to true results in an object key that looks like date=2022-07-18/1658176486-30f6652c-71da-4f9f-800d-a1189c47c547.

This ensures there are no name collisions, and can be useful in high-volume workloads where object keys must be unique.

default: true

filename_extension

optional string literal

The filename extension to use in the object key.

This overrides setting the extension based on the configured compression.

Examples
"json"

filename_time_format

optional string literal

The timestamp format for the time component of the object key.

By default, object keys are appended with a timestamp that reflects when the objects are sent to S3, such that the resulting object key is functionally equivalent to joining the key prefix with the formatted timestamp, such as date=2022-07-18/1658176486.

This would represent a key_prefix set to date=%F/ and the timestamp of Mon Jul 18 2022 20:34:44 GMT+0000, with the filename_time_format being set to %s, which renders timestamps in seconds since the Unix epoch.

Supports the common strftime specifiers found in most languages.

When set to an empty string, no timestamp is appended to the key prefix.

default: %s

framing

optional object
Framing configuration.
Options for the character delimited encoder.
Relevant when: method = "character_delimited"
The ASCII (7-bit) character that delimits byte sequences.

framing.method

required string literal enum
The framing method.
Enum options
OptionDescription
bytesEvent data is not delimited at all.
character_delimitedEvent data is delimited by a single ASCII (7-bit) character.
length_delimited

Event data is prefixed with its length in bytes.

The prefix is a 32-bit unsigned integer, little endian.

newline_delimitedEvent data is delimited by a newline (LF) character.
Examples
"bytes"
"character_delimited"
"length_delimited"
"newline_delimited"

grant_full_control

optional string literal

Grants READ, READ_ACP, and WRITE_ACP permissions on the created objects to the named grantee.

This allows the grantee to read the created objects and their metadata, as well as read and modify the ACL on the created objects.

Examples
"79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
"person@email.com"
"http://acs.amazonaws.com/groups/global/AllUsers"

grant_read

optional string literal

Grants READ permissions on the created objects to the named grantee.

This allows the grantee to read the created objects and their metadata.

Examples
"79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
"person@email.com"
"http://acs.amazonaws.com/groups/global/AllUsers"

grant_read_acp

optional string literal

Grants READ_ACP permissions on the created objects to the named grantee.

This allows the grantee to read the ACL on the created objects.

Examples
"79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
"person@email.com"
"http://acs.amazonaws.com/groups/global/AllUsers"

grant_write_acp

optional string literal

Grants WRITE_ACP permissions on the created objects to the named grantee.

This allows the grantee to modify the ACL on the created objects.

Examples
"79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be"
"person@email.com"
"http://acs.amazonaws.com/groups/global/AllUsers"

healthcheck

optional object
Healthcheck configuration.

healthcheck.enabled

optional bool
Whether or not to check the health of the sink when Vector starts up.
default: true

inputs

required [string]

A list of upstream source or transform IDs.

Wildcards (*) are supported.

See configuration for more info.

Array string literal
Examples
[
  "my-source-or-transform-id",
  "prefix-*"
]

key_prefix

optional string template

A prefix to apply to all object keys.

Prefixes are useful for partitioning objects, such as by creating an object key that stores objects under a particular directory. If using a prefix for this purpose, it must end in / to act as a directory path. A trailing / is not automatically added.

Note: This parameter supports Vector's template syntax, which enables you to use dynamic per-event values.
Examples
"date=%F/hour=%H"
"year=%Y/month=%m/day=%d"
"application_id={{ application_id }}/date=%F"
default: date=%F

proxy

optional object

Proxy configuration.

Configure to proxy traffic through an HTTP(S) proxy when making external requests.

Similar to common proxy configuration convention, you can set different proxies to use based on the type of traffic being proxied, as well as set specific hosts that should not be proxied.

proxy.enabled

optional bool
Enables proxying support.
default: true

proxy.http

optional string literal

Proxy endpoint to use when proxying HTTP traffic.

Must be a valid URI string.

Examples
"http://foo.bar:3128"

proxy.https

optional string literal

Proxy endpoint to use when proxying HTTPS traffic.

Must be a valid URI string.

Examples
"http://foo.bar:3128"

proxy.no_proxy

optional [string]

A list of hosts to avoid proxying.

Multiple patterns are allowed:

PatternExample match
Domain namesexample.com matches requests to example.com
Wildcard domains.example.com matches requests to example.com and its subdomains
IP addresses127.0.0.1 matches requests to 127.0.0.1
CIDR blocks192.168.0.0/16 matches requests to any IP addresses in this range
Splat* matches all hosts

region

optional string literal
The AWS region of the target service.
Examples
"us-east-1"

request

optional object

Middleware settings for outbound requests.

Various settings can be configured, such as concurrency and rate limits, timeouts, retry behavior, etc.

Note that the retry backoff policy follows the Fibonacci sequence.

Configuration of adaptive concurrency parameters.

These parameters typically do not require changes from the default, and incorrect values can lead to meta-stable or unstable performance and sink behavior. Proceed with caution.

The fraction of the current value to set the new concurrency limit when decreasing the limit.

Valid values are greater than 0 and less than 1. Smaller values cause the algorithm to scale back rapidly when latency increases.

Note that the new limit is rounded down after applying this ratio.

default: 0.9

The weighting of new measurements compared to older measurements.

Valid values are greater than 0 and less than 1.

ARC uses an exponentially weighted moving average (EWMA) of past RTT measurements as a reference to compare with the current RTT. Smaller values cause this reference to adjust more slowly, which may be useful if a service has unusually high response variability.

default: 0.4

The initial concurrency limit to use. If not specified, the initial limit will be 1 (no concurrency).

It is recommended to set this value to your service’s average limit if you’re seeing that it takes a long time to ramp up adaptive concurrency after a restart. You can find this value by looking at the adaptive_concurrency_limit metric.

default: 1

The maximum concurrency limit.

The adaptive request concurrency limit will not go above this bound. This is put in place as a safeguard.

default: 200

Scale of RTT deviations which are not considered anomalous.

Valid values are greater than or equal to 0, and we expect reasonable values to range from 1.0 to 3.0.

When calculating the past RTT average, we also compute a secondary “deviation” value that indicates how variable those values are. We use that deviation when comparing the past RTT average to the current measurements, so we can ignore increases in RTT that are within an expected range. This factor is used to scale up the deviation to an appropriate range. Larger values cause the algorithm to ignore larger increases in the RTT.

default: 2.5

request.concurrency

optional string literal enum uint

Configuration for outbound request concurrency.

This can be set either to one of the below enum values or to a positive integer, which denotes a fixed concurrency limit.

Enum options
OptionDescription
adaptiveConcurrency will be managed by Vector’s Adaptive Request Concurrency feature.
none

A fixed concurrency of 1.

Only one request can be outstanding at any given time.

default: adaptive
The time window used for the rate_limit_num option.
default: 1 (seconds)
The maximum number of requests allowed within the rate_limit_duration_secs time window.
default: 9.223372036854776e+18 (requests)
The maximum number of retries to make for failed requests.
default: 9.223372036854776e+18 (retries)

The amount of time to wait before attempting the first retry for a failed request.

After the first retry has failed, the fibonacci sequence is used to select future backoffs.

default: 1 (seconds)

request.retry_jitter_mode

optional string literal enum
The jitter mode to use for retry backoff behavior.
Enum options
OptionDescription
Full

Full jitter.

The random delay is anywhere from 0 up to the maximum current delay calculated by the backoff strategy.

Incorporating full jitter into your backoff strategy can greatly reduce the likelihood of creating accidental denial of service (DoS) conditions against your own systems when many clients are recovering from a failure state.

NoneNo jitter.
default: Full
The maximum amount of time to wait between retries.
default: 30 (seconds)

The time a request can take before being aborted.

Datadog highly recommends that you do not lower this value below the service’s internal timeout, as this could create orphaned requests, pile on retries, and result in duplicate data downstream.

default: 60 (seconds)

server_side_encryption

optional string literal enum

AWS S3 Server-Side Encryption algorithms.

The Server-side Encryption algorithm used when storing these objects.

Enum options string literal
OptionDescription
AES256

Each object is encrypted with AES-256 using a unique key.

This corresponds to the SSE-S3 option.

aws:kms

Each object is encrypted with AES-256 using keys managed by AWS KMS.

Depending on whether or not a KMS key ID is specified, this corresponds either to the SSE-KMS option (keys generated/managed by KMS) or the SSE-C option (keys generated by the customer, managed by KMS).

ssekms_key_id

optional string template

Specifies the ID of the AWS Key Management Service (AWS KMS) symmetrical customer managed customer master key (CMK) that is used for the created objects.

Only applies when server_side_encryption is configured to use KMS.

If not specified, Amazon S3 uses the AWS managed CMK in AWS to protect the data.

Note: This parameter supports Vector's template syntax, which enables you to use dynamic per-event values.
Examples
"abcd1234"

storage_class

optional string literal enum

The storage class for the created objects.

See the S3 Storage Classes for more details.

Enum options string literal
OptionDescription
DEEP_ARCHIVEGlacier Deep Archive.
GLACIERGlacier Flexible Retrieval.
INTELLIGENT_TIERINGIntelligent Tiering.
ONEZONE_IAInfrequently Accessed (single Availability zone).
REDUCED_REDUNDANCYReduced Redundancy.
STANDARDStandard Redundancy.
STANDARD_IAInfrequently Accessed.
default: STANDARD

tags

optional object
The tag-set for the object.

tags.*

required string literal
A single tag.

timezone

optional string literal

Timezone to use for any date specifiers in template strings.

This can refer to any valid timezone as defined in the TZ database, or “local” which refers to the system local timezone. It will default to the globally configured timezone.

Examples
"local"
"America/New_York"
"EST5EDT"

tls

optional object
TLS configuration.

tls.alpn_protocols

optional [string]

Sets the list of supported ALPN protocols.

Declare the supported ALPN protocols, which are used during negotiation with peer. They are prioritized in the order that they are defined.

tls.ca_file

optional string literal

Absolute path to an additional CA certificate file.

The certificate must be in the DER or PEM (X.509) format. Additionally, the certificate can be provided as an inline string in PEM format.

Examples
"/path/to/certificate_authority.crt"

tls.crt_file

optional string literal

Absolute path to a certificate file used to identify this server.

The certificate must be in DER, PEM (X.509), or PKCS#12 format. Additionally, the certificate can be provided as an inline string in PEM format.

If this is set, and is not a PKCS#12 archive, key_file must also be set.

Examples
"/path/to/host_certificate.crt"

tls.key_file

optional string literal

Absolute path to a private key file used to identify this server.

The key must be in DER or PEM (PKCS#8) format. Additionally, the key can be provided as an inline string in PEM format.

Examples
"/path/to/host_certificate.key"

tls.key_pass

optional string literal

Passphrase used to unlock the encrypted key file.

This has no effect unless key_file is set.

Examples
"${KEY_PASS_ENV_VAR}"
"PassWord1"

Enables certificate verification.

If enabled, certificates must not be expired and must be issued by a trusted issuer. This verification operates in a hierarchical manner, checking that the leaf certificate (the certificate presented by the client/server) is not only valid, but that the issuer of that certificate is also valid, and so on until the verification process reaches a root certificate.

Relevant for both incoming and outgoing connections.

Do NOT set this to false unless you understand the risks of not verifying the validity of certificates.

tls.verify_hostname

optional bool

Enables hostname verification.

If enabled, the hostname used to connect to the remote host must be present in the TLS certificate presented by the remote host, either as the Common Name or as an entry in the Subject Alternative Name extension.

Only relevant for outgoing connections.

Do NOT set this to false unless you understand the risks of not verifying the remote hostname.

Environment variables

AWS_ACCESS_KEY_ID

common optional string literal
The AWS access key id. Used for AWS authentication when communicating with AWS services.
Examples
AKIAIOSFODNN7EXAMPLE

AWS_CONFIG_FILE

common optional string literal
Specifies the location of the file that the AWS CLI uses to store configuration profiles.
Default: ~/.aws/config

AWS_DEFAULT_REGION

common optional string literal
The default AWS region.
Examples
/path/to/credentials.json

AWS_PROFILE

common optional string literal
Specifies the name of the CLI profile with the credentials and options to use. This can be the name of a profile stored in a credentials or config file.
Default: default
Examples
my-custom-profile

AWS_ROLE_SESSION_NAME

common optional string literal
Specifies a name to associate with the role session. This value appears in CloudTrail logs for commands performed by the user of this profile.
Examples
vector-session

AWS_SECRET_ACCESS_KEY

common optional string literal
The AWS secret access key. Used for AWS authentication when communicating with AWS services.
Examples
wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY

AWS_SESSION_TOKEN

common optional string literal
The AWS session token. Used for AWS authentication when communicating with AWS services.
Examples
AQoEXAMPLEH4aoAH0gNCAPy...truncated...zrkuWJOgQs8IZZaIv2BXIa2R4Olgk

AWS_SHARED_CREDENTIALS_FILE

common optional string literal
Specifies the location of the file that the AWS CLI uses to store access keys.
Default: ~/.aws/credentials

Telemetry

Metrics

link

buffer_byte_size

gauge
The number of bytes current in the buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_discarded_events_total

counter
The number of events dropped by this non-blocking buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_events

gauge
The number of events currently in the buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_received_event_bytes_total

counter
The number of bytes received by this buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_received_events_total

counter
The number of events received by this buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_sent_event_bytes_total

counter
The number of bytes sent by this buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

buffer_sent_events_total

counter
The number of events sent by this buffer.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

component_discarded_events_total

counter
The number of events dropped by this component.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
intentional
True if the events were discarded intentionally, like a filter transform, or false if due to an error.
pid optional
The process ID of the Vector instance.

component_errors_total

counter
The total number of errors encountered by this component.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
error_type
The type of the error
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.
stage
The stage within the component at which the error occurred.

component_received_event_bytes_total

counter
The number of event bytes accepted by this component either from tagged origins like file and uri, or cumulatively from other origins.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
container_name optional
The name of the container from which the data originated.
file optional
The file from which the data originated.
host optional
The hostname of the system Vector is running on.
mode optional
The connection mode used by the component.
peer_addr optional
The IP from which the data originated.
peer_path optional
The pathname from which the data originated.
pid optional
The process ID of the Vector instance.
pod_name optional
The name of the pod from which the data originated.
uri optional
The sanitized URI from which the data originated.

component_received_events_count

histogram

A histogram of the number of events passed in each internal batch in Vector’s internal topology.

Note that this is separate than sink-level batching. It is mostly useful for low level debugging performance issues in Vector due to small internal batches.

component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
container_name optional
The name of the container from which the data originated.
file optional
The file from which the data originated.
host optional
The hostname of the system Vector is running on.
mode optional
The connection mode used by the component.
peer_addr optional
The IP from which the data originated.
peer_path optional
The pathname from which the data originated.
pid optional
The process ID of the Vector instance.
pod_name optional
The name of the pod from which the data originated.
uri optional
The sanitized URI from which the data originated.

component_received_events_total

counter
The number of events accepted by this component either from tagged origins like file and uri, or cumulatively from other origins.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
container_name optional
The name of the container from which the data originated.
file optional
The file from which the data originated.
host optional
The hostname of the system Vector is running on.
mode optional
The connection mode used by the component.
peer_addr optional
The IP from which the data originated.
peer_path optional
The pathname from which the data originated.
pid optional
The process ID of the Vector instance.
pod_name optional
The name of the pod from which the data originated.
uri optional
The sanitized URI from which the data originated.

component_sent_bytes_total

counter
The number of raw bytes sent by this component to destination sinks.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
endpoint optional
The endpoint to which the bytes were sent. For HTTP, this will be the host and path only, excluding the query string.
file optional
The absolute path of the destination file.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.
protocol
The protocol used to send the bytes.
region optional
The AWS region name to which the bytes were sent. In some configurations, this may be a literal hostname.

component_sent_event_bytes_total

counter
The total number of event bytes emitted by this component.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
output optional
The specific output of the component.
pid optional
The process ID of the Vector instance.

component_sent_events_total

counter
The total number of events emitted by this component.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
output optional
The specific output of the component.
pid optional
The process ID of the Vector instance.

utilization

gauge
A ratio from 0 to 1 of the load on a component. A value of 0 would indicate a completely idle component that is simply waiting for input. A value of 1 would indicate a that is never idle. This value is updated every 5 seconds.
component_id
The Vector component ID.
component_kind
The Vector component kind.
component_type
The Vector component type.
host optional
The hostname of the system Vector is running on.
pid optional
The process ID of the Vector instance.

Permissions

Platform: Amazon Web Services
Relevant policies
PolicyRequired forRequired when
s3:ListBucket
  • healthcheck
s3:PutObject
  • operation

How it works

AWS authentication

Vector checks for AWS credentials in the following order:

  1. The auth.access_key_id and auth.secret_access_key options.
  2. The AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment variables.
  3. In Web Identity Token credentials from the environment or container (including EKS). These credentials will automatically refresh when expired.
  4. ECS credentials (IAM roles for tasks). These credentials will automatically refresh when expired.
  5. As entries in the credentials file in the .aws directory in your home directory (~/.aws/credentials on Linux, OS X, and Unix; %userprofile%\.aws\credentials on Microsoft Windows).
  6. Using a named profile specified in the credentials file via the AWS_PROFILE environment variable.
  7. The IAM instance profile (only works if running on an EC2 instance with an instance profile/role). Requires IMDSv2 to be enabled. For EKS, you may need to increase the metadata token response hop limit to 2. These credentials will automatically refresh when expired.

If no credentials are found, Vector’s health check fails and an error is logged.

Obtaining an access key

In general, we recommend using instance profiles/roles whenever possible. In cases where this is not possible you can generate an AWS access key for any user within your AWS account. AWS provides a detailed guide on how to do this. Such created AWS access keys can be used via auth.access_key_id and auth.secret_access_key options.

Assuming roles

Vector can assume an AWS IAM role via the auth.assume_role option. This is an optional setting that is helpful for a variety of use cases, such as cross account access.

Buffers and batches

This component buffers & batches data as shown in the diagram above. You’ll notice that Vector treats these concepts differently, instead of treating them as global concepts, Vector treats them as sink specific concepts. This isolates sinks, ensuring services disruptions are contained and delivery guarantees are honored.

Batches are flushed when 1 of 2 conditions are met:

  1. The batch age meets or exceeds the configured timeout_secs.
  2. The batch size meets or exceeds the configured max_bytes or max_events.

Buffers are controlled via the buffer.* options.

Cross account object writing

If you’re using Vector to write objects across AWS accounts then you should consider setting the grant_full_control option to the bucket owner’s canonical user ID. AWS provides a full tutorial for this use case. If don’t know the bucket owner’s canonical ID you can find it by following this tutorial.

Health checks

Health checks ensure that the downstream service is accessible and ready to accept data. This check is performed upon sink initialization. If the health check fails an error will be logged and Vector will proceed to start.

Require health checks

If you’d like to exit immediately upon a health check failure, you can pass the --require-healthy flag:

vector --config /etc/vector/vector.yaml --require-healthy

Disable health checks

If you’d like to disable health checks for this sink you can set the healthcheck option to false.

Object Access Control List (ACL)

AWS S3 supports access control lists (ACL) for buckets and objects. In the context of Vector, only object ACLs are relevant (Vector does not create or modify buckets). You can set the object level ACL by using one of the acl, grant_full_control, grant_read, grant_read_acp, or grant_write_acp options.

acl.* vs grant_* options

The grant_* options name a specific entity to grant access to. The acl options is one of a set of specific canned ACLs that can only name the owner or world.

Object naming

Vector uses two different naming schemes for S3 objects. If you set the compression parameter to true (this is the default), Vector uses this scheme:

<key_prefix><timestamp>-<uuidv4>.log.gz

If compression isn’t enabled, Vector uses this scheme (only the file extension is different):

<key_prefix><timestamp>-<uuidv4>.log

Some sample S3 object names (with and without compression, respectively):

date=2019-06-18/1560886634-fddd7a0e-fad9-4f7e-9bce-00ae5debc563.log.gz
date=2019-06-18/1560886634-fddd7a0e-fad9-4f7e-9bce-00ae5debc563.log

Vector appends a UUIDV4 token to ensure there are no naming conflicts in the unlikely event that two Vector instances are writing data at the same time.

You can control the resulting name via the key_prefix, filename_time_format, and filename_append_uuid options.

For example, to store objects at the root S3 folder, without a timestamp or UUID use these configuration options:

key_prefix = "{{ my_file_name }}"
filename_time_format = ""
filename_append_uuid = false

Object Tags & metadata

Vector currently only supports AWS S3 object tags and does not support object metadata. If you require metadata support see issue #1694.

We believe tags are more flexible since they are separate from the actual S3 object. You can freely modify tags without modifying the object. Conversely, object metadata requires a full rewrite of the object to make changes.

Rate limits & adaptive concurrency

Adaptive Request Concurrency (ARC)

Adaptive Request Concurrency is a feature of Vector that does away with static concurrency limits and automatically optimizes HTTP concurrency based on downstream service responses. The underlying mechanism is a feedback loop inspired by TCP congestion control algorithms. Checkout the announcement blog post,

We highly recommend enabling this feature as it improves performance and reliability of Vector and the systems it communicates with. As such, we have made it the default, and no further configuration is required.

Static concurrency

If Adaptive Request Concurrency is not for you, you can manually set static concurrency limits by specifying an integer for request.concurrency:

sinks:
	my-sink:
		request:
			concurrency: 10

Rate limits

In addition to limiting request concurrency, you can also limit the overall request throughput via the request.rate_limit_duration_secs and request.rate_limit_num options.

sinks:
	my-sink:
		request:
			rate_limit_duration_secs: 1
			rate_limit_num: 10

These will apply to both adaptive and fixed request.concurrency values.

Retry policy

Vector will retry failed requests (status == 429, >= 500, and != 501). Other responses will not be retried. You can control the number of retry attempts and backoff rate with the request.retry_attempts and request.retry_backoff_secs options.

Server-Side Encryption (SSE)

AWS S3 offers server-side encryption. You can apply defaults at the bucket level or set the encryption at the object level. In the context, of Vector only the object level is relevant (Vector does not create or modify buckets). Although, we recommend setting defaults at the bucket level when possible. You can explicitly set the object level encryption via the server_side_encryption option.

State

This component is stateless, meaning its behavior is consistent across each input.

Storage class

AWS S3 offers storage classes. You can apply defaults, and rules, at the bucket level or set the storage class at the object level. In the context of Vector only the object level is relevant (Vector does not create or modify buckets). You can set the storage class via the storage_class option.

Transport Layer Security (TLS)

Vector uses OpenSSL for TLS protocols due to OpenSSL’s maturity. You can enable and adjust TLS behavior via the tls.* options and/or via an OpenSSL configuration file. The file location defaults to /usr/local/ssl/openssl.cnf or can be specified with the OPENSSL_CONF environment variable.