Logs are an essential part of observing any service; without them you'll have significant blind spots. But collecting and analyzing them can be a real challenge -- especially at scale. Not only do you need to solve the basic task of collecting your logs, but you must do it in a reliable, performant, and robust manner. Nothing is more frustrating than having your logs pipeline fall on it's face during an outage, or even worse, cause the outage!
Fear not! In this guide we'll build an observability pipeline that will send logs from Docker to anywhere.
What is Docker?
Docker is an open platform for developing, shipping, and running applications and services. Docker enables you to separate your services from your infrastructure so you can ship quickly. With Docker, you can manage your infrastructure in the same ways you manage your services. By taking advantage of Docker's methodologies for shipping, testing, and deploying code quickly, you can significantly reduce the delay between writing code and running it in production.
How This Guide Works
We'll be using [Vector][urls.vector_website] to accomplish this task. Vector is a popular open-source observability data pipeline. It's written in Rust, making it lightweight, ultra-fast and highly reliable. And we'll be deploying Vector as a agent.
What We'll Accomplish
We'll build an observability data pipeline that:
All in just a few minutes!
Start Vectordocker run \-d \-v ~/vector.toml:/etc/vector/vector.toml:ro \-p 8383:8383 \timberio/vector:0.13.X-debian
Observe Vectordocker logs -f $(docker ps -aqf "name=vector")explain this command
Vector is powerful tool and we're just scratching the surface in this guide. Here are a few pages we recommend that demonstrate the power and flexibility of Vector: