현재 이 페이지는 영어로만 제공됩니다.

Ashley Jeffs is a guest contributor from Timber.io. Julian Giuca, a Senior Director and Product GM at New Relic, also contributed to this post.

The success of your observability practice depends on the quality of the data you collect from your  systems—and today, your systems are producing all kinds of metrics and logs. You could gather this data in a handful of different ways by managing a handful of different tools, but these tools are complex, and reducing complexity is a huge win for your overall architecture.

This is exactly why the New Relic Logs team recently partnered with Timber.io to add a New Relic Logs integration to Vector—a monitoring and observability data router that collects and transforms metrics and log data from a wide range of sources, including as StatsD, Syslog, Docker, and Kafka. Now all this critical data can be collected and analyzed on the New Relic One platform.

Read on to learn more about how this integration works.

Capabilities
New relic virtuo customer story
Discover the power of New Relic log management
Manage your logs Manage your logs

What is Vector?

Vector is a lightweight and ultra-fast open source tool for building observability pipelines. With Vector, you can add transforms that automatically enrich your logs and metrics with useful environment metadata. For example, the Vector EC2 metadata transform enriches logs from EC2 with instance_id, hostname, region, vpc_id,  and more.

Here are the key components of a Vector observability pipeline:

  • Sources: Define where Vector should pull data from, or how it should receive data pushed to it. A pipeline can have any number of sources, and as they ingest data they proceed to normalize it into “events.”
    • Events: All logs and metrics that pass through a pipeline are described as events. (Check out Vector's data model docs for more about events.)
  • Transforms: Mutate events as they’re transported by Vector, via parsing, filtering, sampling, or aggregating. You can have any number of transforms in your pipeline, and compose them however you want.
  • Sinks: Create a destination for events. Each sink's design and transmission method is dictated by the downstream service it interacts with. For example, the New Relic logs sink, batches Vector log events to New Relic's log service via the logging API.

Getting Started with Vector’s New Relic Logs integration

The New Relic Logs integration is included in Vector. To get started pushing metrics and logs to New Relic, you simply need to install Vector and configure the New Relic Logs sink.

You can install Vector on a number of Linux operating systems, as well as MacOS, Windows, and Raspbian.

This example uses the Linux-based install script:

curl --proto '=https' --tlsv1.2 -sSf https://sh.vector.dev | sh

The installation creates a configuration file at /etc/vector/vector.toml. From here you need to specify a source (e.g., Vector’s stdin source) and add the new_relic_logs sink.

Note: To set up the new_relic_logs sink, you'll need your New Relic license key.

Add the following in /etc/vector/vector.toml:

[sources.stdin]

  type = "stdin"



[sinks.new_relic]

  type = "new_relic_logs"

  inputs = ["stdin"]

  license_key = "<YOUR LICENSE KEY>"

Finally, run Vector. Since we configured this example to use the stdin source, any lines we write will be sent to the new_relic_logs sink, which will populate in New Relic Logs. For example:

$ vector -c ./vector.toml

hello world!

New Relic Logs captures an event synced from a Vector pipeline

 

 

 

 

 

 

 

 

 

That's it—we've sent logs from an stdin source to New Relic using a Vector pipeline.

To learn more, check out the New Relic Logs Sink documentation. Join the Nerdlog discussion every Thursday at 12 p.m. PT on Twitch or keep up to date with our Nerdlog on your own like our Heroku integration article.