Support for the Prometheus remote write protocol
Interoperability with the Prometheus ecosystem.
We’re big fans of Prometheus at Timber, and as an extension of our Kubernetes integration we wanted to better understand how Vector could assist Prometheus operators. As noted in the Kubernetes highlight, it is our intent to be the only tool needed to collect and process all Kubernetes observability data, and working with Prometheus is core to our metrics strategy. As a result, 0.11.0 includes two new components that assist with a variety of Prometheus use cases:
Using the new
Prometheus operators can route a stream of Prometheus data to the archiving
solution of their choice. For this use case we recommend object stores for their
cheap and durable qualities:
[sources.prometheus] type = "prometheus_remote_write" [transforms.convert] type = "metric_to_log" inputs = ["prometheus"] [sinks.backup] type = "aws_s3" inputs = ["convert"]
gcp_cloud_storage or other object stores.
For large setups it’s common to couple Prometheus with another solution designed for long term storage and querying. This separates the concerns of fast short term querying and long-term low cost archiving.
Not using Prometheus? It’s very common to use Prometheus as a central proxy for exporting all metrics data. This is especially relevant in ecosystems like Kubernetes where Prometheus is tightly integrated.
[sources.prometheus] type = "prometheus_remote_write" [sinks.datadog] type = "datadog_metrics" inputs = ["prometheus"]